-
Notifications
You must be signed in to change notification settings - Fork 10
Description
Hi,
Thank you for your work.
I wanted to ask: are you only using the first question in each QAs.json entry as the instruction during training, as indicated here:
| question = QAs["questions"][0] |
If so, wouldn’t that underutilize the dataset, since most samples contain 5–6 questions? Have you tried incorporating all questions during training?
Additionally, I noticed that the evaluation metric you used is IoU as mentioned in the paper, which only considers the segmentation masks. In that case, the answers in QAs.json are not directly involved in training. Why, then, are they loaded into the data_dict here:
SegEarth-R1/segearth_r1/train/train_dataset.py
Lines 327 to 328 in f0c382c
| # answer_idx = random.randint(0, answer_num - 1) | |
| answer = QAs["answer"][0] |
Or am I missing something and the answers are also used during training? If so, I noticed you tried both random sampling and using answer[0] as the line above, could you clarify which approach you ultimately used, and whether using different strategies affected performance?
Regards,
Vicky