@@ -59,14 +59,14 @@ Provide the pretrained checkpoint path through variable `${HF_MODEL_CKPT}`:
5959 TP=1 \
6060 HF_MODEL_CKPT=< pretrained_model_name_or_path> \
6161 MLM_MODEL_SAVE=/tmp/Llama-3.2-1B-Instruct-FP8 \
62- bash megatron-lm /examples/post_training/modelopt/quantize.sh meta-llama/Llama-3.2-1B-Instruct fp8
62+ bash Megatron-LM /examples/post_training/modelopt/quantize.sh meta-llama/Llama-3.2-1B-Instruct fp8
6363
6464\
6565 PP=1 \
6666 HF_MODEL_CKPT=< pretrained_model_name_or_path> \
6767 MLM_MODEL_LOAD=/tmp/Llama-3.2-1B-Instruct-FP8 \
6868 EXPORT_DIR=/tmp/Llama-3.2-1B-Instruct-Export \
69- bash megatron-lm /examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
69+ bash Megatron-LM /examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
7070
7171```
7272
@@ -86,14 +86,14 @@ required for training is generated on the fly.
8686 TP=1 \
8787 HF_MODEL_CKPT=< pretrained_model_name_or_path> \
8888 MLM_MODEL_SAVE=/tmp/Llama-3.2-1B-Eagle3 \
89- bash megatron-lm /examples/post_training/modelopt/eagle3.sh meta-llama/Llama-3.2-1B-Instruct
89+ bash Megatron-LM /examples/post_training/modelopt/eagle3.sh meta-llama/Llama-3.2-1B-Instruct
9090
9191\
9292 PP=1 \
9393 HF_MODEL_CKPT=< pretrained_model_name_or_path> \
9494 MLM_MODEL_LOAD=/tmp/Llama-3.2-1B-Eagle3 \
9595 EXPORT_DIR=/tmp/Llama-3.2-1B-Eagle3-Export \
96- bash megatron-lm /examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
96+ bash Megatron-LM /examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
9797```
9898
9999Periodically, ** acceptance length (AL)** is evaluated on MT-Bench prompts. You can find resumable
@@ -130,7 +130,7 @@ PP=1 \
130130TARGET_NUM_LAYERS=24 \
131131HF_MODEL_CKPT=< pretrained_model_name_or_path> \
132132MLM_MODEL_SAVE=Qwen3-8B-Pruned \
133- bash megatron-lm /examples/post_training/modelopt/prune.sh qwen/Qwen3-8B
133+ bash Megatron-LM /examples/post_training/modelopt/prune.sh qwen/Qwen3-8B
134134```
135135
136136> [ !TIP]
@@ -146,7 +146,7 @@ quantization.
146146``` sh
147147\
148148 HF_MODEL_CKPT=< pretrained_model_name_or_path> \
149- bash megatron-lm /examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
149+ bash Megatron-LM /examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
150150```
151151
152152> ** ❗ IMPORTANT:** ` pretrained_model_card ` ** CANNOT** be a path to a local pretrained checkpoint.
@@ -163,7 +163,7 @@ to pass all variables instead. If you have your own script, use `${SANDBOX_ENV_S
163163``` sh
164164\
165165 SANDBOX_ENV_SETUP=< path_to_your_script> \
166- bash megatron-lm /examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
166+ bash Megatron-LM /examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
167167```
168168
169169If you use our ` slurm ` script, then you ** MUST USE** ` ${SANDBOX_ENV_SETUP} ` (default: ` ./env_setup_template.sh ` ).
0 commit comments