Skip to content

Commit 8303cfa

Browse files
authored
[None][fix] Fix import issues in layer-wise benchmarks (#8827)
Signed-off-by: Tailing Yuan <[email protected]>
1 parent 4873ca0 commit 8303cfa

File tree

3 files changed

+14
-31
lines changed

3 files changed

+14
-31
lines changed

tests/integration/defs/conftest.py

Lines changed: 0 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -626,25 +626,11 @@ def deepseek_v3_model_root(request):
626626
elif request.param == "DeepSeek-V3-Lite-nvfp4_moe_only":
627627
deepseek_v3_model_root = os.path.join(models_root, "DeepSeek-V3-Lite",
628628
"nvfp4_moe_only")
629-
elif request.param == "DeepSeek-V3.2-Exp":
630-
deepseek_v3_model_root = os.path.join(models_root,
631-
"DeepSeek-V3.2-Exp-hf")
632629
assert exists(
633630
deepseek_v3_model_root), f"{deepseek_v3_model_root} does not exist!"
634631
return deepseek_v3_model_root
635632

636633

637-
@pytest.fixture(scope="function")
638-
def deepseek_r1_model_root(request):
639-
models_root = llm_models_root()
640-
if request.param == "DeepSeek-R1-0528-FP4-v2":
641-
deepseek_r1_model_root = os.path.join(models_root, "DeepSeek-R1",
642-
"DeepSeek-R1-0528-FP4-v2")
643-
assert exists(
644-
deepseek_r1_model_root), f"{deepseek_r1_model_root} does not exist!"
645-
return deepseek_r1_model_root
646-
647-
648634
@pytest.fixture(scope="session")
649635
def trt_performance_cache_name():
650636
return "performance.cache"

tests/integration/test_lists/waives.txt

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -294,8 +294,6 @@ full:L20/accuracy/test_llm_api_pytorch.py::TestMinistral8BInstruct::test_fp8 SKI
294294
full:L40S/accuracy/test_llm_api_pytorch.py::TestLlama3_2_1B::test_fp8_prequantized SKIP (https://nvbugs/5542862)
295295
full:L40S/accuracy/test_llm_api_pytorch.py::TestMinistral8BInstruct::test_fp8 SKIP (https://nvbugs/5542862)
296296
unittest/_torch/multi_gpu_modeling/test_llama3.py::test_llama_3_3 SKIP (https://nvbugs/5536131)
297-
unittest/tools/test_layer_wise_benchmarks.py::test_deepseek_r1_ctx_tep SKIP (never tested)
298-
unittest/tools/test_layer_wise_benchmarks.py::test_deepseek_r1_gen_scaled_from_16_dep SKIP (never tested)
299297
examples/test_eagle.py::test_llm_eagle_1gpu_modelopt_ckpt[llama3.1-eagle-8b-hf_v0.5-float16-bs8] SKIP (https://nvbugs/5546507)
300298
examples/test_eagle.py::test_llm_eagle_1gpu[EAGLE-Vicuna-7B-v1.3-float16-bs1-eagle1] SKIP (https://nvbugs/5546507)
301299
examples/test_eagle.py::test_llm_eagle_1gpu[EAGLE-Vicuna-7B-v1.3-float16-bs1-eagle2] SKIP (https://nvbugs/5546507)

tests/unittest/tools/test_layer_wise_benchmarks.py

Lines changed: 14 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -1,23 +1,22 @@
11
import os
2+
from subprocess import check_call
23

34
import pytest
45
import torch
5-
from defs.conftest import deepseek_r1_model_root # noqa: F401
6-
from defs.conftest import deepseek_v3_model_root # noqa: F401
7-
from defs.trt_test_alternative import check_call
86
from utils.cpp_paths import llm_root # noqa: F401
7+
from utils.llm_data import llm_models_root
98

109

1110
@pytest.mark.skipif(torch.cuda.device_count() < 4,
1211
reason="needs 4 GPUs to run this test")
13-
@pytest.mark.parametrize("deepseek_r1_model_root", ["DeepSeek-R1-0528-FP4-v2"],
14-
indirect=True)
15-
def test_deepseek_r1_ctx_tep(llm_root, deepseek_r1_model_root):
12+
def test_deepseek_r1_ctx_tep(llm_root):
13+
model_root = llm_models_root(check=True)
1614
check_call([
1715
"./mpi_launch.sh",
1816
"./run_single.sh",
1917
"config_ctx.yaml",
20-
"--model=" + deepseek_r1_model_root,
18+
"--model",
19+
model_root / "DeepSeek-R1" / "DeepSeek-R1-0528-FP4-v2",
2120
"--no-enable-attention-dp",
2221
"--moe-backend=TRTLLM",
2322
],
@@ -31,14 +30,14 @@ def test_deepseek_r1_ctx_tep(llm_root, deepseek_r1_model_root):
3130

3231
@pytest.mark.skipif(torch.cuda.device_count() < 4,
3332
reason="needs 4 GPUs to run this test")
34-
@pytest.mark.parametrize("deepseek_v3_model_root", ["DeepSeek-V3.2-Exp"],
35-
indirect=True)
36-
def test_deepseek_v32_ctx_dep(llm_root, deepseek_v3_model_root):
33+
def test_deepseek_v32_ctx_dep(llm_root):
34+
model_root = llm_models_root(check=True)
3735
check_call([
3836
"./mpi_launch.sh",
3937
"./run_single.sh",
4038
"config_ctx.yaml",
41-
"--model=" + deepseek_v3_model_root,
39+
"--model",
40+
model_root / "DeepSeek-V3.2-Exp-hf",
4241
"--tokens-per-block=64",
4342
"--moe-backend=DEEPGEMM",
4443
],
@@ -51,14 +50,14 @@ def test_deepseek_v32_ctx_dep(llm_root, deepseek_v3_model_root):
5150

5251
@pytest.mark.skipif(torch.cuda.device_count() < 4,
5352
reason="needs 4 GPUs to run this test")
54-
@pytest.mark.parametrize("deepseek_r1_model_root", ["DeepSeek-R1-0528-FP4-v2"],
55-
indirect=True)
56-
def test_deepseek_r1_gen_scaled_from_16_dep(llm_root, deepseek_r1_model_root):
53+
def test_deepseek_r1_gen_scaled_from_16_dep(llm_root):
54+
model_root = llm_models_root(check=True)
5755
check_call([
5856
"./mpi_launch.sh",
5957
"./run_single.sh",
6058
"config_gen.yaml",
61-
"--model=" + deepseek_r1_model_root,
59+
"--model",
60+
model_root / "DeepSeek-R1" / "DeepSeek-R1-0528-FP4-v2",
6261
"--layer-indices=5,6",
6362
"--scaled-from=16",
6463
"--moe-backend=WIDEEP",

0 commit comments

Comments
 (0)