|
250 | 250 | "--deepspeed", "scripts/zero3.json", |
251 | 251 | "--model_name_or_path", "lmms-lab/llava-onevision-qwen2-0.5b-ov", |
252 | 252 | "--version", "qwen_1_5", |
253 | | - "--data_path", "scripts/train/vis_supervision_strong_baseline.yaml", |
| 253 | + "--data_path", "scripts/train/simple_tim_top5_cut.yaml", |
254 | 254 | "--video_folder", "/data/shaokai/", |
255 | 255 | "--mm_tunable_parts", "mm_vision_tower,mm_mlp_adapter,mm_language_model", |
256 | 256 | "--mm_vision_tower_lr", "2e-6", |
|
294 | 294 | "--val_metadata", "/data/shaokai/epic-kitchens-100-annotations/EPIC_100_validation.csv", |
295 | 295 | "--llava_num_frames", "16", |
296 | 296 | "--clip_length", "16", |
297 | | - "--action_representation", "GT_random_narration", |
| 297 | + "--action_representation", "topk_narration_cut_key", |
298 | 298 | "--topk_predictions", "5", |
299 | | - "--eval_steps", "10", |
| 299 | + "--eval_steps", "1", |
300 | 300 | "--vision_supervision", "all_newlines", |
301 | | - "--action_types", "97,300,3806" |
| 301 | + "--action_types", "97,300,3806", |
| 302 | + "--n_narration", "5" |
302 | 303 | ], |
303 | 304 | "console": "integratedTerminal", |
304 | 305 | "justMyCode": false, |
|
0 commit comments