Skip to content

Commit d5a8079

Browse files
authored
Revert "[infra] Unwaive unittests/_torch" (NVIDIA#4950)
1 parent 743fb0a commit d5a8079

File tree

2 files changed

+1
-7
lines changed

2 files changed

+1
-7
lines changed

tests/integration/test_lists/waives.txt

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -380,14 +380,9 @@ full:B200/examples/test_qwen.py::test_llm_qwen_7b_multi_gpus_summary[qwen2.5_7b_
380380
accuracy/test_cli_flow.py::TestMixtral8x22B::test_int8_plugin_tp8[renormalize-tensor_parallel] SKIP (https://nvbugs/5273695)
381381
examples/test_gpt.py::test_starcoder_fp8_quantization_2gpu[starcoder] SKIP (https://nvbugs/5144931)
382382
examples/test_gpt.py::test_starcoder_fp8_quantization_2gpu[starcoderplus] SKIP (https://nvbugs/5144931)
383+
unittest/_torch -k "not (modeling or multi_gpu or auto_deploy)" SKIP (https://nvbugs/5280806)
383384
examples/test_whisper.py::test_llm_whisper_general[large-v3-disable_gemm_plugin-disable_attention_plugin-disable_weight_only-float16-nb:1-use_python_runtime] SKIP (https://nvbugs/5244570)
384385
unittest/_torch/speculative/test_eagle3.py SKIP (https://nvbugs/5280806)
385-
unittest/_torch/modules/test_fused_moe.py SKIP (https://nvbugspro.nvidia.com/bug/5324229)
386-
unittest/_torch/modules/test_moe_load_balancer.py SKIP (https://nvbugspro.nvidia.com/bug/5324229)
387-
unittest/_torch/speculative/test_ngram.py SKIP (https://nvbugspro.nvidia.com/bug/5324239)
388-
unittest/_torch/test_pytorch_model_engine.py SKIP (https://nvbugspro.nvidia.com/bug/5324248)
389-
unittest/_torch/test_resource_manager.py SKIP (https://nvbugspro.nvidia.com/bug/5324252)
390-
unittest/_torch/thop/test_selective_scan_op.py SKIP (https://nvbugspro.nvidia.com/bug/5324258)
391386
triton_server/test_triton_rcca.py::test_mistral_beam_search[rcca_4714407-True-10---False-True-False-0-128-disableDecoupleMode-inflight_fused_batching-disableTrtOverlap--guaranteed_no_evict---1-1-1-False-ensemble] SKIP (https://nvbugs/5240060)
392387
triton_server/test_triton.py::test_triton_extensive[triton-extensive] SKIP
393388
triton_server/test_triton.py::test_gpt_speculative_decoding[gpt-speculative-decoding] SKIP

tests/unittest/_torch/speculative/test_eagle3.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,6 @@
1717
[[True, "TRTLLM"], [False, "TRTLLM"],
1818
[True, "FLASHINFER"], [False, "FLASHINFER"]])
1919
def test_llama_eagle3(use_cuda_graph: bool, attn_backend: str):
20-
pytest.skip("Test is hanging")
2120
total_mem_gb = torch.cuda.get_device_properties(0).total_memory / 1e9
2221
if total_mem_gb < 35:
2322
pytest.skip("Not enough memory to load target + draft model")

0 commit comments

Comments
 (0)