File tree Expand file tree Collapse file tree 1 file changed +8
-7
lines changed
Expand file tree Collapse file tree 1 file changed +8
-7
lines changed Original file line number Diff line number Diff line change 1- MODEL=EVA02-CLIP-L-14
1+ MODEL=EVA02-CLIP-L-14-336
22PRETRAINED=eva_clip
3- python -m torch.distributed.launch --nproc_per_node=2 \
3+ python -m torch.distributed.launch --nproc_per_node=8 \
44 --use_env training/main.py \
55 --enable-deepspeed \
66 --grad-checkpointing \
77 --name=" T_vitl336_Rcc12mR_Rcc3m_4ep" \
88 --save-frequency 1 \
99 --zeroshot-frequency 1 \
10- --report-to=" " \
10+ --report-to=" tensorboard, wandb " \
1111 --wandb-project-name=" LLM2CLIP" \
1212 --wandb-notes=" EVA02-CLIP-L-14-336" \
13- --train-data-list " /home/aiscuser/LLM2CLIP/llm2clip/data/cc3m/{00..00004}.tar" \
14- --train-num-samples-list 2873538 \
13+ --train-data-list " data/cc3m/cc3m-train-{00..0287}.tar;data/cc12m/cc12m-train-{00..1001}.tar" \
14+ --train-num-samples-list 2873538 10000225 \
15+ --eval-data-file=training/eval_datasets.yaml \
1516 --pretrained=${PRETRAINED} \
1617 --dataset-resampled \
1718 --precision " fp16" \
1819 --warmup 0 \
19- --batch-size=16 \
20+ --batch-size=512 \
2021 --eval-batch-size=1024 \
2122 --log-every-n-steps 50 \
2223 --epochs=20 \
@@ -31,7 +32,7 @@ python -m torch.distributed.launch --nproc_per_node=2 \
3132 --visual-ld=0.85 \
3233 --grad-clip-norm=5.0 \
3334 --smoothing=0. \
34- --workers=1 \
35+ --workers=8 \
3536 --model=${MODEL} \
3637 --seed 4096 \
3738 --gather-with-grad \
You can’t perform that action at this time.
0 commit comments