File tree Expand file tree Collapse file tree 2 files changed +0
-4
lines changed
tensorrt_llm/_torch/pyexecutor
tests/integration/test_lists Expand file tree Collapse file tree 2 files changed +0
-4
lines changed Original file line number Diff line number Diff line change @@ -888,9 +888,6 @@ def _create_warmup_request(
888888 available_tokens = kv_cache_manager .get_num_available_tokens (
889889 self .runtime_draft_len )
890890 available_blocks = kv_cache_manager .get_num_free_blocks ()
891- print (
892- f"available_tokens: { available_tokens } , num_tokens: { num_tokens } , num_gen_requests: { num_gen_requests } "
893- )
894891 if num_tokens > self .max_num_tokens or num_tokens > available_tokens :
895892 return None
896893
Original file line number Diff line number Diff line change @@ -375,7 +375,6 @@ accuracy/test_llm_api_pytorch_multimodal.py::TestQwen2_5_VL_7B::test_auto_dtype
375375accuracy/test_llm_api_pytorch_multimodal.py::TestLlava_V1_6_Mistral_7B::test_auto_dtype SKIP (https://nvbugs/5707087)
376376accuracy/test_llm_api_pytorch_multimodal.py::TestPhi4MMFusedVisionLora::test_auto_dtype SKIP (https://nvbugs/5707087)
377377disaggregated/test_disaggregated.py::test_disaggregated_ctxtp2pp2_gentp2pp2[TinyLlama-1.1B-Chat-v1.0] SKIP (https://nvbugs/5705199)
378- accuracy/test_llm_api_pytorch.py::TestLlama3_1_8BInstruct::test_fp8_4gpus[tp4-fp8kv=False-attn_backend=FLASHINFER-torch_compile=True] SKIP (https://nvbugs/5707359)
379378unittest/_torch/modules/test_fused_moe.py::test_fused_moe_fp8_blockwise_wide_ep[NotEnabled] SKIP (https://nvbugs/5707392)
380379accuracy/test_llm_api_pytorch.py::TestLlama3_3NemotronSuper49Bv1::test_auto_dtype_tp2 SKIP (https://nvbugs/5707145)
381380accuracy/test_llm_api_pytorch.py::TestLlama3_3NemotronSuper49Bv1::test_fp8_prequantized_tp2 SKIP (https://nvbugs/5707145)
You can’t perform that action at this time.
0 commit comments