Skip to content

Commit eee2422

Browse files
author
Haozhe Qi
committed
bug fix
1 parent a296605 commit eee2422

File tree

4 files changed

+98
-99
lines changed

4 files changed

+98
-99
lines changed

llava/train/train.py

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1206,8 +1206,7 @@ def _get_item(self, i) -> Dict[str, torch.Tensor]:
12061206
if not os.path.exists(video_file):
12071207
print("File {} not exist!".format(video_file))
12081208

1209-
# try:
1210-
if True:
1209+
try:
12111210
if "sharegpt4video" in video_folder:
12121211
frame_files = [os.path.join(video_file, f) for f in os.listdir(video_file) if os.path.isfile(os.path.join(video_file, f))]
12131212
frame_files.sort() # Ensure the frames are sorted if they are named sequentially
@@ -1312,12 +1311,12 @@ def _get_item(self, i) -> Dict[str, torch.Tensor]:
13121311
image = [(image, video[0].size, "video", action)]
13131312
sources = preprocess_multimodal(copy.deepcopy([e["conversations"] for e in sources]), self.data_args)
13141313
# print(sources)
1315-
# except Exception as e:
1316-
# import traceback
1317-
# traceback.print_exc()
1318-
# print(f"Error: {e}")
1319-
# print(f"Failed to read video file: {video_file}")
1320-
# return self._get_item(i + 1)
1314+
except Exception as e:
1315+
import traceback
1316+
traceback.print_exc()
1317+
print(f"Error: {e}")
1318+
print(f"Failed to read video file: {video_file}")
1319+
return self._get_item(i + 1)
13211320
else:
13221321
sources = copy.deepcopy([e["conversations"] for e in sources])
13231322

run_clariden.sbatch

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,12 +4,12 @@
44
#SBATCH --hint nomultithread
55
#SBATCH --cpus-per-task 288
66
#SBATCH --no-requeue
7-
#SBATCH --nodes 8 # number of Nodes
7+
#SBATCH --nodes 16 # number of Nodes
88
#SBATCH --ntasks-per-node 1 # number of MP tasks. IMPORTANT: torchrun represents just 1 Slurm task
99
#SBATCH --gres gpu:4 # Number of GPUs
1010
#SBATCH --time 23:00:00 # maximum execution time (DD-HH:MM:SS). Mandatory field in MN5
11-
#SBATCH --output logs/R-%x.%j-dev_7b_64f_10llavavideo_EK100_haozhe.out
12-
#SBATCH --error logs/R-%x.%j-dev_7b_64f_10llavavideo_EK100_haozhe.err
11+
#SBATCH --output logs/R-%x.%j-dev_7b_64f_EK100_haozhe.out
12+
#SBATCH --error logs/R-%x.%j-dev_7b_64f_EK100_haozhe.err
1313

1414
mkdir -p logs
1515

@@ -71,8 +71,8 @@ PYTHON_ARGS=" \
7171
--image_grid_pinpoints \"(1x1),...,(6x6)\" \
7272
--mm_patch_merge_type spatial_unpad \
7373
--bf16 True \
74-
--run_name dev_7b_64f_10llavavideo_EK100_haozhe \
75-
--output_dir experiments/dev_7b_64f_10llavavideo_EK100_haozhe \
74+
--run_name dev_7b_64f_EK100_haozhe \
75+
--output_dir experiments/dev_7b_64f_EK100_haozhe \
7676
--num_train_epochs 1 \
7777
--per_device_train_batch_size 1 \
7878
--per_device_eval_batch_size 4 \
@@ -99,7 +99,7 @@ PYTHON_ARGS=" \
9999
--mm_newline_position grid \
100100
--add_time_instruction True \
101101
--force_sample True \
102-
--mm_spatial_pool_stride 2
102+
--mm_spatial_pool_stride 2 \
103103
--root /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/EK100 \
104104
--action_predictions /iopsstor/scratch/cscs/hqi/VFM/llava_data/TIM_PREDS/tim_pred_ids_val.json \
105105
--val_metadata /iopsstor/scratch/cscs/hqi/VFM/EK100/epic-kitchens-100-annotations/EPIC_100_validation.csv \

run_llmseval_clariden.sbatch

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,8 @@
88
#SBATCH --ntasks-per-node 1 # number of MP tasks. IMPORTANT: torchrun represents just 1 Slurm task
99
#SBATCH --gres gpu:4 # Number of GPUs
1010
#SBATCH --time 23:00:00 # maximum execution time (DD-HH:MM:SS). Mandatory field in MN5
11-
#SBATCH --output logs/R-%x.%j-lmmseval-dev_7b_4f_llavavideo_test_haozhe.out
12-
#SBATCH --error logs/R-%x.%j-lmmseval-dev_7b_4f_llavavideo_test_haozhe.err
11+
#SBATCH --output logs/R-%x.%j-lmmseval-dev_LLaVA-Video-7B-Qwen2_64f_10llavavideo_EK100_haozhe.out
12+
#SBATCH --error logs/R-%x.%j-lmmseval-dev_LLaVA-Video-7B-Qwen2_64f_10llavavideo_EK100_haozhe.err
1313

1414
mkdir -p logs
1515

@@ -62,12 +62,12 @@ PYTHON_FILE="-m lmms_eval"
6262

6363
PYTHON_ARGS=" \
6464
--model llava_vid \
65-
--model_args pretrained=lmms-lab/LLaVA-Video-7B-Qwen2,conv_template=qwen_1_5,max_frames_num=64,mm_spatial_pool_mode=average \
66-
--tasks activitynetqa,videochatgpt,nextqa_mc_test,egoschema,video_dc499,videomme,videomme_w_subtitle,perceptiontest_val_mc \
65+
--model_args pretrained=experiments/dev_LLaVA-Video-7B-Qwen2_64f_10llavavideo_EK100_haozhe,conv_template=qwen_1_5,max_frames_num=64,mm_spatial_pool_mode=average \
66+
--tasks youcook2_val \
6767
--batch_size 1 \
6868
--log_samples \
6969
--log_samples_suffix llava_vid \
70-
--output_path ./logs/
70+
--output_path ./logs/ \
7171
--verbosity=DEBUG \
7272
"
7373

scripts/train/llava_video.yaml

Lines changed: 80 additions & 80 deletions
Original file line numberDiff line numberDiff line change
@@ -1,81 +1,81 @@
11
datasets:
2-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_academic_v0_1/0_30_s_academic_v0_1_cap_processed.json
3-
sampling_strategy: "first:1%"
4-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_youtube_v0_1/0_30_s_youtube_v0_1_cap_processed.json
5-
sampling_strategy: "first:10%"
6-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_academic_v0_1/30_60_s_academic_v0_1_cap_processed.json
7-
sampling_strategy: "first:10%"
8-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_youtube_v0_1/30_60_s_youtube_v0_1_cap_processed.json
9-
sampling_strategy: "first:10%"
10-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_academic_v0_1/1_2_m_academic_v0_1_cap_processed.json
11-
sampling_strategy: "first:10%"
12-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_youtube_v0_1/1_2_m_youtube_v0_1_cap_processed.json
13-
sampling_strategy: "first:10%"
14-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_academic_v0_1/0_30_s_academic_oe_v0_1_qa_processed.json
15-
sampling_strategy: "first:10%"
16-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_academic_v0_1/0_30_s_academic_oe_v0_1_qa_processed.json
17-
sampling_strategy: "first:10%"
18-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_youtube_v0_1/0_30_s_youtube_oe_v0_1_qa_processed.json
19-
sampling_strategy: "first:10%"
20-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_youtube_v0_1/0_30_s_youtube_mc_v0_1_qa_processed.json
21-
sampling_strategy: "first:10%"
22-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_activitynetqa/0_30_s_activitynetqa_oe_qa_processed.json
23-
sampling_strategy: "first:10%"
24-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_nextqa/0_30_s_nextqa_oe_qa_processed.json
25-
sampling_strategy: "first:10%"
26-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_nextqa/0_30_s_nextqa_mc_qa_processed.json
27-
sampling_strategy: "first:10%"
28-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_perceptiontest/0_30_s_perceptiontest_mc_qa_processed.json
29-
sampling_strategy: "first:10%"
30-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_academic_v0_1/30_60_s_academic_oe_v0_1_qa_processed.json
31-
sampling_strategy: "first:10%"
32-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_academic_v0_1/30_60_s_academic_mc_v0_1_qa_processed.json
33-
sampling_strategy: "first:10%"
34-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_youtube_v0_1/30_60_s_youtube_oe_v0_1_qa_processed.json
35-
sampling_strategy: "first:10%"
36-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_youtube_v0_1/30_60_s_youtube_mc_v0_1_qa_processed.json
37-
sampling_strategy: "first:10%"
38-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_activitynetqa/30_60_s_activitynetqa_oe_qa_processed.json
39-
sampling_strategy: "first:10%"
40-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_nextqa/30_60_s_nextqa_oe_qa_processed.json
41-
sampling_strategy: "first:10%"
42-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_nextqa/30_60_s_nextqa_mc_qa_processed.json
43-
sampling_strategy: "first:10%"
44-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_perceptiontest/30_60_s_perceptiontest_mc_qa_processed.json
45-
sampling_strategy: "first:10%"
46-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_academic_v0_1/1_2_m_academic_oe_v0_1_qa_processed.json
47-
sampling_strategy: "first:10%"
48-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_academic_v0_1/1_2_m_academic_mc_v0_1_qa_processed.json
49-
sampling_strategy: "first:10%"
50-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_youtube_v0_1/1_2_m_youtube_oe_v0_1_qa_processed.json
51-
sampling_strategy: "first:10%"
52-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_youtube_v0_1/1_2_m_youtube_mc_v0_1_qa_processed.json
53-
sampling_strategy: "first:10%"
54-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_activitynetqa/1_2_m_activitynetqa_oe_qa_processed.json
55-
sampling_strategy: "first:10%"
56-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_nextqa/1_2_m_nextqa_oe_qa_processed.json
57-
sampling_strategy: "first:10%"
58-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_nextqa/1_2_m_nextqa_mc_qa_processed.json
59-
sampling_strategy: "first:10%"
60-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_academic_v0_1/2_3_m_academic_v0_1_cap_processed.json
61-
sampling_strategy: "first:10%"
62-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_youtube_v0_1/2_3_m_youtube_v0_1_cap_processed.json
63-
sampling_strategy: "first:10%"
64-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_academic_v0_1/2_3_m_academic_oe_v0_1_qa_processed.json
65-
sampling_strategy: "first:10%"
66-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_academic_v0_1/2_3_m_academic_mc_v0_1_qa_processed.json
67-
sampling_strategy: "first:10%"
68-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_youtube_v0_1/2_3_m_youtube_oe_v0_1_qa_processed.json
69-
sampling_strategy: "first:10%"
70-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_youtube_v0_1/2_3_m_youtube_mc_v0_1_qa_processed.json
71-
sampling_strategy: "first:10%"
72-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_nextqa/2_3_m_nextqa_oe_qa_processed.json
73-
sampling_strategy: "first:10%"
74-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_nextqa/2_3_m_nextqa_mc_qa_processed.json
75-
sampling_strategy: "first:10%"
76-
- json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_activitynetqa/2_3_m_activitynetqa_oe_qa_processed.json
77-
sampling_strategy: "first:10%"
78-
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/llava_hound/sharegptvideo_qa_255k_processed.json
79-
# sampling_strategy: "first:10%"
80-
- json_path: /media/data/haozhe/VFM/EK100/EK100_in_LLAVA/TIM/tim_mc_top5_official_key/train_convs_narration_actionids.jsonl
81-
sampling_strategy: "first:1%"
2+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_academic_v0_1/0_30_s_academic_v0_1_cap_processed.json
3+
# sampling_strategy: "first:10%"
4+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_youtube_v0_1/0_30_s_youtube_v0_1_cap_processed.json
5+
# sampling_strategy: "first:10%"
6+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_academic_v0_1/30_60_s_academic_v0_1_cap_processed.json
7+
# sampling_strategy: "first:10%"
8+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_youtube_v0_1/30_60_s_youtube_v0_1_cap_processed.json
9+
# sampling_strategy: "first:10%"
10+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_academic_v0_1/1_2_m_academic_v0_1_cap_processed.json
11+
# sampling_strategy: "first:10%"
12+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_youtube_v0_1/1_2_m_youtube_v0_1_cap_processed.json
13+
# sampling_strategy: "first:10%"
14+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_academic_v0_1/0_30_s_academic_oe_v0_1_qa_processed.json
15+
# sampling_strategy: "first:10%"
16+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_academic_v0_1/0_30_s_academic_oe_v0_1_qa_processed.json
17+
# sampling_strategy: "first:10%"
18+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_youtube_v0_1/0_30_s_youtube_oe_v0_1_qa_processed.json
19+
# sampling_strategy: "first:10%"
20+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_youtube_v0_1/0_30_s_youtube_mc_v0_1_qa_processed.json
21+
# sampling_strategy: "first:10%"
22+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_activitynetqa/0_30_s_activitynetqa_oe_qa_processed.json
23+
# sampling_strategy: "first:10%"
24+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_nextqa/0_30_s_nextqa_oe_qa_processed.json
25+
# sampling_strategy: "first:10%"
26+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_nextqa/0_30_s_nextqa_mc_qa_processed.json
27+
# sampling_strategy: "first:10%"
28+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/0_30_s_perceptiontest/0_30_s_perceptiontest_mc_qa_processed.json
29+
# sampling_strategy: "first:10%"
30+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_academic_v0_1/30_60_s_academic_oe_v0_1_qa_processed.json
31+
# sampling_strategy: "first:10%"
32+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_academic_v0_1/30_60_s_academic_mc_v0_1_qa_processed.json
33+
# sampling_strategy: "first:10%"
34+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_youtube_v0_1/30_60_s_youtube_oe_v0_1_qa_processed.json
35+
# sampling_strategy: "first:10%"
36+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_youtube_v0_1/30_60_s_youtube_mc_v0_1_qa_processed.json
37+
# sampling_strategy: "first:10%"
38+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_activitynetqa/30_60_s_activitynetqa_oe_qa_processed.json
39+
# sampling_strategy: "first:10%"
40+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_nextqa/30_60_s_nextqa_oe_qa_processed.json
41+
# sampling_strategy: "first:10%"
42+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_nextqa/30_60_s_nextqa_mc_qa_processed.json
43+
# sampling_strategy: "first:10%"
44+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/30_60_s_perceptiontest/30_60_s_perceptiontest_mc_qa_processed.json
45+
# sampling_strategy: "first:10%"
46+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_academic_v0_1/1_2_m_academic_oe_v0_1_qa_processed.json
47+
# sampling_strategy: "first:10%"
48+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_academic_v0_1/1_2_m_academic_mc_v0_1_qa_processed.json
49+
# sampling_strategy: "first:10%"
50+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_youtube_v0_1/1_2_m_youtube_oe_v0_1_qa_processed.json
51+
# sampling_strategy: "first:10%"
52+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_youtube_v0_1/1_2_m_youtube_mc_v0_1_qa_processed.json
53+
# sampling_strategy: "first:10%"
54+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_activitynetqa/1_2_m_activitynetqa_oe_qa_processed.json
55+
# sampling_strategy: "first:10%"
56+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_nextqa/1_2_m_nextqa_oe_qa_processed.json
57+
# sampling_strategy: "first:10%"
58+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/1_2_m_nextqa/1_2_m_nextqa_mc_qa_processed.json
59+
# sampling_strategy: "first:10%"
60+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_academic_v0_1/2_3_m_academic_v0_1_cap_processed.json
61+
# sampling_strategy: "first:10%"
62+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_youtube_v0_1/2_3_m_youtube_v0_1_cap_processed.json
63+
# sampling_strategy: "first:10%"
64+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_academic_v0_1/2_3_m_academic_oe_v0_1_qa_processed.json
65+
# sampling_strategy: "first:10%"
66+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_academic_v0_1/2_3_m_academic_mc_v0_1_qa_processed.json
67+
# sampling_strategy: "first:10%"
68+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_youtube_v0_1/2_3_m_youtube_oe_v0_1_qa_processed.json
69+
# sampling_strategy: "first:10%"
70+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_youtube_v0_1/2_3_m_youtube_mc_v0_1_qa_processed.json
71+
# sampling_strategy: "first:10%"
72+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_nextqa/2_3_m_nextqa_oe_qa_processed.json
73+
# sampling_strategy: "first:10%"
74+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_nextqa/2_3_m_nextqa_mc_qa_processed.json
75+
# sampling_strategy: "first:10%"
76+
# - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/2_3_m_activitynetqa/2_3_m_activitynetqa_oe_qa_processed.json
77+
# sampling_strategy: "first:10%"
78+
# # - json_path: /iopsstor/scratch/cscs/hqi/VFM/onevision/llava_video/LLaVA-Video-178K/llava_hound/sharegptvideo_qa_255k_processed.json
79+
# # sampling_strategy: "first:10%"
80+
- json_path: /iopsstor/scratch/cscs/hqi/VFM/llava_data/EK100_inst_train/tim_mc_top5_GT_random_narration/train_convs_narration_actionids.jsonl
81+
sampling_strategy: "all"

0 commit comments

Comments
 (0)