@@ -132,7 +132,7 @@ run_and_eval() {
132132}
133133
134134# W4A16
135- # run_and_eval "llama3_ddp_example.py" 4 "Meta-Llama-3-8B-Instruct-W4A16-G128-DDP4" "gsm8k" 2048 # .7111 .7127
135+ run_and_eval " llama3_ddp_example.py" 4 " Meta-Llama-3-8B-Instruct-W4A16-G128-DDP4" " gsm8k" 2048 # .7111 .7127
136136# run_and_eval "llama3_ddp_example.py" 1 "Meta-Llama-3-8B-Instruct-W4A16-G128-DDP1" "gsm8k" 2048 # .702 .702
137137
138138# run_and_eval "qwen3_vl_8b_gptq_int4_ddp_example.py" 4 "Qwen3-VL-8B-Instruct-GPTQ-W4A16-G128-DDP4" "gsm8k" 2048 # .8514 .8476
@@ -150,17 +150,17 @@ run_and_eval() {
150150
151151
152152# NVFP4
153- run_and_eval " llama3_ddp_nvfp4.py" 4 " Meta-Llama-3-8B-Instruct-GPTQ-NVFP4A16-DDP4" " gsm8k" 2048
154- run_and_eval " llama3_ddp_nvfp4.py" 1 " Meta-Llama-3-8B-Instruct-GPTQ-NVFP4A16-DDP1" " gsm8k" 2048
153+ # run_and_eval "llama3_ddp_nvfp4.py" 4 "Meta-Llama-3-8B-Instruct-GPTQ-NVFP4A16-DDP4" "gsm8k" 2048
154+ # run_and_eval "llama3_ddp_nvfp4.py" 1 "Meta-Llama-3-8B-Instruct-GPTQ-NVFP4A16-DDP1" "gsm8k" 2048
155155
156- run_and_eval " qwen3_vl_8b_gptq_nvfp4_ddp_example.py" 4 " Qwen3-VL-8B-Instruct-GPTQ-NVFP4A16-DDP4" " gsm8k" 2048
157- run_and_eval " qwen3_vl_8b_gptq_nvfp4_ddp_example.py" 1 " Qwen3-VL-8B-Instruct-GPTQ-NVFP4A16-DDP1" " gsm8k" 2048
156+ # run_and_eval "qwen3_vl_8b_gptq_nvfp4_ddp_example.py" 4 "Qwen3-VL-8B-Instruct-GPTQ-NVFP4A16-DDP4" "gsm8k" 2048
157+ # run_and_eval "qwen3_vl_8b_gptq_nvfp4_ddp_example.py" 1 "Qwen3-VL-8B-Instruct-GPTQ-NVFP4A16-DDP1" "gsm8k" 2048
158158
159- run_and_eval " qwen3_30b_moe_gptq_nvfp4_ddp_example.py" 4 " Qwen3-30B-A3B-GPTQ-NVFP4A16-DDP4" " gsm8k" 2048
160- run_and_eval " qwen3_30b_moe_gptq_nvfp4_ddp_example.py" 1 " Qwen3-30B-A3B-GPTQ-NVFP4A16-DDP1" " gsm8k" 2048
159+ # run_and_eval "qwen3_30b_moe_gptq_nvfp4_ddp_example.py" 4 "Qwen3-30B-A3B-GPTQ-NVFP4A16-DDP4" "gsm8k" 2048
160+ # run_and_eval "qwen3_30b_moe_gptq_nvfp4_ddp_example.py" 1 "Qwen3-30B-A3B-GPTQ-NVFP4A16-DDP1" "gsm8k" 2048
161161
162- run_and_eval " llama4_gptq_nvfp4_ddp_example.py" 4 " Llama-4-Scout-17B-16E-Instruct-GPTQ-NVFP4A16-DDP4" " gsm8k" 8192
163- run_and_eval " llama4_gptq_nvfp4_ddp_example.py" 1 " Llama-4-Scout-17B-16E-Instruct-GPTQ-NVFP4A16-DDP1" " gsm8k" 8192
162+ # run_and_eval "llama4_gptq_nvfp4_ddp_example.py" 4 "Llama-4-Scout-17B-16E-Instruct-GPTQ-NVFP4A16-DDP4" "gsm8k" 8192
163+ # run_and_eval "llama4_gptq_nvfp4_ddp_example.py" 1 "Llama-4-Scout-17B-16E-Instruct-GPTQ-NVFP4A16-DDP1" "gsm8k" 8192
164164
165165# run_and_eval "qwen3_vl_235b_moe_nvfp4_ddp_example.py" 8 "Qwen3-VL-235B-A22B-Instruct-GPTQ-NVFP4A16-DDP8" "gsm8k" 2048
166166# run_and_eval "qwen3_vl_235b_moe_nvfp4_ddp_example.py" 1 "Qwen3-VL-235B-A22B-Instruct-GPTQ-NVFP4A16-DDP1" "gsm8k" 2048
@@ -171,10 +171,10 @@ echo "============================================"
171171echo " Starting base model evaluations"
172172echo " ============================================"
173173
174- eval_base_model " meta-llama/Meta-Llama-3-8B-Instruct" " gsm8k" 2048
175- eval_base_model " Qwen/Qwen3-VL-8B-Instruct" " gsm8k" 2048
176- eval_base_model " Qwen/Qwen3-30B-A3B" " gsm8k" 2048
177- eval_base_model " meta-llama/Llama-4-Scout-17B-16E-Instruct" " gsm8k" 8192
174+ # eval_base_model "meta-llama/Meta-Llama-3-8B-Instruct" "gsm8k" 2048
175+ # eval_base_model "Qwen/Qwen3-VL-8B-Instruct" "gsm8k" 2048
176+ # eval_base_model "Qwen/Qwen3-30B-A3B" "gsm8k" 2048
177+ # eval_base_model "meta-llama/Llama-4-Scout-17B-16E-Instruct" "gsm8k" 8192
178178
179179echo " ============================================"
180180echo " All runs complete!"
0 commit comments