1- {
2- "version" : " 0.2.0" ,
3- "configurations" : [
4- {
5- "name" : " Run LLAVA Training with torchrun" ,
6- "type" : " debugpy" ,
7- "request" : " launch" ,
8- "module" : " torch.distributed.run" ,
9- "env" : {
10- "CUDA_VISIBLE_DEVICES" : " 1,2" ,
11- "OMP_NUM_THREADS" : " 8" ,
12- "NCCL_IB_DISABLE" : " 0" ,
13- "NCCL_IB_GID_INDEX" : " 3" ,
14- "NCCL_SOCKET_IFNAME" : " eth0" ,
15- "NCCL_DEBUG" : " INFO" ,
16- "ACCELERATE_CPU_AFFINITY" : " 1" ,
17- "LD_PRELOAD" : " /usr/lib/x86_64-linux-gnu/libffi.so.7" ,
18- },
19- "args" : [
20- " --nproc_per_node=2" ,
21- " --nnodes=1" ,
22- " --node_rank=0" ,
23- " --master_addr=127.0.0.1" ,
24- " --master_port=29500" ,
25- " llava/train/train_mem.py" ,
26- " --deepspeed" , " scripts/zero3.json" ,
27- " --model_name_or_path" , " lmms-lab/llava-onevision-qwen2-0.5b-ov" ,
28- " --version" , " qwen_1_5" ,
29- " --data_path" , " scripts/train/onevision.yaml" ,
30- // "--image_folder", "/mediaPFM/data/haozhe/onevision/llava_data",
31- " --image_folder" , " /mediaPFM/data/haozhe/onevision/llava_data/geo3k/" ,
32- " --video_folder" , " /mediaPFM/data/haozhe/onevision/llava_video" ,
33- " --mm_tunable_parts" , " mm_vision_tower,mm_mlp_adapter,mm_language_model" ,
34- " --mm_vision_tower_lr" , " 2e-6" ,
35- " --vision_tower" , " google/siglip-so400m-patch14-384" ,
36- " --mm_projector_type" , " mlp2x_gelu" ,
37- " --mm_vision_select_layer" , " -2" ,
38- " --mm_use_im_start_end" , " False" ,
39- " --mm_use_im_patch_token" , " False" ,
40- " --group_by_modality_length" , " True" ,
41- " --image_aspect_ratio" , " anyres_max_9" ,
42- " --image_grid_pinpoints" , " (1x1),...,(6x6)" ,
43- " --mm_patch_merge_type" , " spatial_unpad" ,
44- " --bf16" , " True" ,
45- " --run_name" , " test" ,
46- " --output_dir" , " experiments/test" ,
47- " --num_train_epochs" , " 1" ,
48- " --per_device_train_batch_size" , " 1" ,
49- " --per_device_eval_batch_size" , " 4" ,
50- " --gradient_accumulation_steps" , " 2" ,
51- " --evaluation_strategy" , " no" ,
52- " --save_strategy" , " steps" ,
53- " --save_steps" , " 1000" ,
54- " --save_total_limit" , " 1" ,
55- " --learning_rate" , " 1e-5" ,
56- " --weight_decay" , " 0." ,
57- " --warmup_ratio" , " 0.03" ,
58- " --lr_scheduler_type" , " cosine" ,
59- " --logging_steps" , " 1" ,
60- " --tf32" , " True" ,
61- " --model_max_length" , " 32768" ,
62- " --gradient_checkpointing" , " True" ,
63- " --dataloader_num_workers" , " 4" ,
64- " --lazy_preprocess" , " True" ,
65- " --report_to" , " wandb" ,
66- " --torch_compile" , " True" ,
67- " --torch_compile_backend" , " inductor" ,
68- " --dataloader_drop_last" , " True" ,
69- " --frames_upbound" , " 32" ,
70- ],
71- "console" : " integratedTerminal" ,
72- "justMyCode" : false ,
73- "cwd" : " ${workspaceFolder}"
74- }
75- ]
76- }
77-
78-
791// {
80- // // Use IntelliSense to learn about possible attributes.
81- // // Hover to view descriptions of existing attributes.
82- // // For more information, visit: https://go.microsoft.com/fwlink/?linkid=830387
832// "version": "0.2.0",
843// "configurations": [
854// {
86- // "name": "Python: Current File ",
5+ // "name": "Run LLAVA Training with torchrun ",
876// "type": "debugpy",
887// "request": "launch",
89- // "program": "docs/LLaVA_OneVision_Tutorials.py",
8+ // "module": "torch.distributed.run",
9+ // "env": {
10+ // "CUDA_VISIBLE_DEVICES": "1,2,3",
11+ // "OMP_NUM_THREADS": "8",
12+ // "NCCL_IB_DISABLE": "0",
13+ // "NCCL_IB_GID_INDEX": "3",
14+ // "NCCL_SOCKET_IFNAME": "eth0",
15+ // "NCCL_DEBUG": "INFO",
16+ // "ACCELERATE_CPU_AFFINITY": "1",
17+ // "LD_PRELOAD": "/usr/lib/x86_64-linux-gnu/libffi.so.7",
18+ // "WANDB_API_KEY": "65aeda82a75f1eed29c8e9250b175fcc73dca0d7",
19+ // },
20+ // "args": [
21+ // "--nproc_per_node=3",
22+ // "--nnodes=1",
23+ // "--node_rank=0",
24+ // "--master_addr=127.0.0.1",
25+ // "--master_port=29500",
26+ // "llava/train/train_mem.py",
27+ // "--deepspeed", "scripts/zero3.json",
28+ // "--model_name_or_path", "lmms-lab/llava-onevision-qwen2-0.5b-ov",
29+ // "--version", "qwen_1_5",
30+ // "--data_path", "scripts/train/onevision.yaml",
31+ // // "--image_folder", "/mediaPFM/data/haozhe/onevision/llava_data",
32+ // "--image_folder", "/mediaPFM/data/haozhe/onevision/llava_data/geo3k/",
33+ // "--video_folder", "/mediaPFM/data/haozhe/onevision/llava_video",
34+ // "--mm_tunable_parts", "mm_vision_tower,mm_mlp_adapter,mm_language_model",
35+ // "--mm_vision_tower_lr", "2e-6",
36+ // "--vision_tower", "google/siglip-so400m-patch14-384",
37+ // "--mm_projector_type", "mlp2x_gelu",
38+ // "--mm_vision_select_layer", "-2",
39+ // "--mm_use_im_start_end", "False",
40+ // "--mm_use_im_patch_token", "False",
41+ // "--group_by_modality_length", "True",
42+ // "--image_aspect_ratio", "anyres_max_9",
43+ // "--image_grid_pinpoints", "(1x1),...,(6x6)",
44+ // "--mm_patch_merge_type", "spatial_unpad",
45+ // "--bf16", "True",
46+ // "--run_name", "test",
47+ // "--output_dir", "experiments/test",
48+ // "--num_train_epochs", "1",
49+ // "--per_device_train_batch_size", "1",
50+ // "--per_device_eval_batch_size", "4",
51+ // "--gradient_accumulation_steps", "2",
52+ // "--evaluation_strategy", "no",
53+ // "--save_strategy", "steps",
54+ // "--save_steps", "1000",
55+ // "--save_total_limit", "1",
56+ // "--learning_rate", "1e-5",
57+ // "--weight_decay", "0.",
58+ // "--warmup_ratio", "0.03",
59+ // "--lr_scheduler_type", "cosine",
60+ // "--logging_steps", "1",
61+ // "--tf32", "True",
62+ // "--model_max_length", "32768",
63+ // "--gradient_checkpointing", "True",
64+ // "--dataloader_num_workers", "4",
65+ // "--lazy_preprocess", "True",
66+ // "--report_to", "wandb",
67+ // "--torch_compile", "True",
68+ // "--torch_compile_backend", "inductor",
69+ // "--dataloader_drop_last", "True",
70+ // "--frames_upbound", "32",
71+ // ],
9072// "console": "integratedTerminal",
91- // "env":{"CUDA_VISIBLE_DEVICES":"0",
92- // "LD_PRELOAD": "/usr/lib/x86_64-linux-gnu/libffi.so.7",
93- // "LD_LIBRARY_PATH": "/home/haozhe/miniconda3/envs/llava/lib"},
9473// "justMyCode": false,
95- // // "args": [
96- // // "--run_dir_name", "test",
97- // // // "--use_big_decoder"
98- // // ]
74+ // "cwd": "${workspaceFolder}"
9975// }
10076// ]
101- // }
77+ // }
78+
79+
80+ {
81+ // Use IntelliSense to learn about possible attributes.
82+ // Hover to view descriptions of existing attributes.
83+ // For more information, visit: https://go.microsoft.com/fwlink/?linkid=830387
84+ "version" : " 0.2.0" ,
85+ "configurations" : [
86+ {
87+ "name" : " Python: Current File" ,
88+ "type" : " debugpy" ,
89+ "request" : " launch" ,
90+ "program" : " docs/LLaVA_OneVision_Tutorials.py" ,
91+ "console" : " integratedTerminal" ,
92+ "env" :{"CUDA_VISIBLE_DEVICES" :" 0" ,
93+ "LD_PRELOAD" : " /usr/lib/x86_64-linux-gnu/libffi.so.7" },
94+ "justMyCode" : false ,
95+ // "args": [
96+ // "--run_dir_name", "test",
97+ // // "--use_big_decoder"
98+ // ]
99+ }
100+ ]
101+ }
0 commit comments