File tree Expand file tree Collapse file tree 1 file changed +3
-9
lines changed
examples/gpu/llm/fine-tuning/Llama2 Expand file tree Collapse file tree 1 file changed +3
-9
lines changed Original file line number Diff line number Diff line change @@ -26,9 +26,7 @@ Run_llama2-7b_fsdp_alpaca_converge() {
2626 --warmup_ratio 0.03 \
2727 --lr_scheduler_type " cosine" \
2828 --logging_steps 1 \
29- --optim " adamw_torch_fused" \
30- --fsdp " full_shard auto_wrap" \
31- --fsdp_transformer_layer_cls_to_wrap ' LlamaDecoderLayer' 2>&1 | tee llama2_fsdp_alpaca_adamfuse_bs4_3epoch_converge.log
29+ --optim " adamw_torch_fused" 2>&1 | tee llama2_fsdp_alpaca_adamfuse_bs4_3epoch_converge.log
3230
3331}
3432
@@ -57,9 +55,7 @@ Run_llama2-7b_fsdp_alpaca_peft_converge() {
5755 --warmup_ratio 0.03 \
5856 --lr_scheduler_type " cosine" \
5957 --logging_steps 1 \
60- --optim " adamw_torch_fused" \
61- --fsdp " full_shard auto_wrap" \
62- --fsdp_transformer_layer_cls_to_wrap ' LlamaDecoderLayer' 2>&1 | tee llama2_fsdp_alpaca_peft_adamfuse_bs4_3epoch_converge.log
58+ --optim " adamw_torch_fused" 2>&1 | tee llama2_fsdp_alpaca_peft_adamfuse_bs4_3epoch_converge.log
6359}
6460
6561
@@ -87,9 +83,7 @@ Run_llama2-70b_fsdp_alpaca_peft_converge() {
8783 --weight_decay 0. \
8884 --warmup_ratio 0.03 \
8985 --lr_scheduler_type " cosine" \
90- --logging_steps 1 \
91- --fsdp " full_shard auto_wrap" \
92- --fsdp_transformer_layer_cls_to_wrap ' LlamaDecoderLayer' 2>&1 | tee llama2_70b_fsdp_alpaca_peft_bs1_3epoch_converge.log
86+ --logging_steps 1 2>&1 | tee llama2_70b_fsdp_alpaca_peft_bs1_3epoch_converge.log
9387
9488}
9589
You can’t perform that action at this time.
0 commit comments