We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 5b02483 commit b751e16Copy full SHA for b751e16
examples/llm_sparsity/launch_finetune.sh
@@ -91,8 +91,8 @@ CMD="accelerate launch --multi_gpu --mixed_precision bf16 finetune.py \
91
--warmup_ratio 0.0 \
92
--lr_scheduler_type cosine \
93
--logging_steps 1 \
94
- --fsdp 'full_shard auto_wrap' \
95
- --fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' \
+ --fsdp full_shard auto_wrap \
+ --fsdp_transformer_layer_cls_to_wrap LlamaDecoderLayer \
96
--tf32 True \
97
--modelopt_restore_path $MODELOPT_RESTORE_PATH \
98
--report_to tensorboard \
examples/llm_sparsity/requirements.txt
@@ -1,3 +1,4 @@
1
flash-attn
2
sentencepiece>=0.2.0
3
tensorboardX
4
+transformers>=4.57.0
0 commit comments