Skip to content

Commit 51a45dc

Browse files
Rename examples/megatron-lm to examples/Megatron-LM (#481)
Signed-off-by: Keval Morabia <[email protected]>
1 parent f2eb794 commit 51a45dc

File tree

2 files changed

+8
-8
lines changed

2 files changed

+8
-8
lines changed

examples/megatron-lm/README.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -59,14 +59,14 @@ Provide the pretrained checkpoint path through variable `${HF_MODEL_CKPT}`:
5959
TP=1 \
6060
HF_MODEL_CKPT=<pretrained_model_name_or_path> \
6161
MLM_MODEL_SAVE=/tmp/Llama-3.2-1B-Instruct-FP8 \
62-
bash megatron-lm/examples/post_training/modelopt/quantize.sh meta-llama/Llama-3.2-1B-Instruct fp8
62+
bash Megatron-LM/examples/post_training/modelopt/quantize.sh meta-llama/Llama-3.2-1B-Instruct fp8
6363

6464
\
6565
PP=1 \
6666
HF_MODEL_CKPT=<pretrained_model_name_or_path> \
6767
MLM_MODEL_LOAD=/tmp/Llama-3.2-1B-Instruct-FP8 \
6868
EXPORT_DIR=/tmp/Llama-3.2-1B-Instruct-Export \
69-
bash megatron-lm/examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
69+
bash Megatron-LM/examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
7070

7171
```
7272

@@ -86,14 +86,14 @@ required for training is generated on the fly.
8686
TP=1 \
8787
HF_MODEL_CKPT=<pretrained_model_name_or_path> \
8888
MLM_MODEL_SAVE=/tmp/Llama-3.2-1B-Eagle3 \
89-
bash megatron-lm/examples/post_training/modelopt/eagle3.sh meta-llama/Llama-3.2-1B-Instruct
89+
bash Megatron-LM/examples/post_training/modelopt/eagle3.sh meta-llama/Llama-3.2-1B-Instruct
9090

9191
\
9292
PP=1 \
9393
HF_MODEL_CKPT=<pretrained_model_name_or_path> \
9494
MLM_MODEL_LOAD=/tmp/Llama-3.2-1B-Eagle3 \
9595
EXPORT_DIR=/tmp/Llama-3.2-1B-Eagle3-Export \
96-
bash megatron-lm/examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
96+
bash Megatron-LM/examples/post_training/modelopt/export.sh meta-llama/Llama-3.2-1B-Instruct
9797
```
9898

9999
Periodically, **acceptance length (AL)** is evaluated on MT-Bench prompts. You can find resumable
@@ -130,7 +130,7 @@ PP=1 \
130130
TARGET_NUM_LAYERS=24 \
131131
HF_MODEL_CKPT=<pretrained_model_name_or_path> \
132132
MLM_MODEL_SAVE=Qwen3-8B-Pruned \
133-
bash megatron-lm/examples/post_training/modelopt/prune.sh qwen/Qwen3-8B
133+
bash Megatron-LM/examples/post_training/modelopt/prune.sh qwen/Qwen3-8B
134134
```
135135

136136
> [!TIP]
@@ -146,7 +146,7 @@ quantization.
146146
```sh
147147
\
148148
HF_MODEL_CKPT=<pretrained_model_name_or_path> \
149-
bash megatron-lm/examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
149+
bash Megatron-LM/examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
150150
```
151151

152152
> **❗ IMPORTANT:** `pretrained_model_card` **CANNOT** be a path to a local pretrained checkpoint.
@@ -163,7 +163,7 @@ to pass all variables instead. If you have your own script, use `${SANDBOX_ENV_S
163163
```sh
164164
\
165165
SANDBOX_ENV_SETUP=<path_to_your_script> \
166-
bash megatron-lm/examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
166+
bash Megatron-LM/examples/post_training/modelopt/quantize.sh [pretrained_model_card] [qformat]
167167
```
168168

169169
If you use our `slurm` script, then you **MUST USE** `${SANDBOX_ENV_SETUP}` (default: `./env_setup_template.sh`).

examples/megatron-lm/interactive.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,12 +18,12 @@
1818
if [ -n "${USER_FSW}" ]; then
1919
echo "USER_FSW is set to ${USER_FSW}"
2020
else
21+
echo "USER_FSW is not set, using default /tmp"
2122
USER_FSW=/tmp
2223
fi
2324

2425
docker run --gpus all --init -it --rm --network host --ipc=host \
2526
--user $(id -u):$(id -g) \
2627
-v $PWD:/workspace/nmm-sandbox \
2728
-v ${USER_FSW}:/workspace/scratch \
28-
-v /home/chenhany/projects/nmm-sandbox/modelopt:/workspace/TensorRT-Model-Optimizer \
2929
nvidia-modelopt-megatron:latest bash

0 commit comments

Comments
 (0)