@@ -1674,13 +1674,15 @@ def test_nvfp4_4gpus_online_eplb(self, fp8kv):
16741674 (False , False , False , True ),
16751675 (True , False , True , True ), (True , True , True , True )])
16761676 @parametrize_with_ids ("mtp_nextn" , [0 , 2 ])
1677- @parametrize_with_ids ("moe_backend" , ["CUTLASS" , "TRTLLM" ])
1677+ @parametrize_with_ids ("moe_backend" , ["CUTLASS" , "TRTLLM" , "CUTEDSL" ])
16781678 def test_nvfp4 (self , fp8kv , attention_dp , cuda_graph , overlap_scheduler ,
16791679 torch_compile , mtp_nextn , moe_backend ):
16801680 if moe_backend == "TRTLLM" and (get_sm_version () == 120
16811681 or get_sm_version () == 121 ):
16821682 pytest .skip (
16831683 "MOE TRTLLM backend does not support SM version 120 or 121" )
1684+ if moe_backend == "CUTEDSL" and get_sm_version () != 100 :
1685+ pytest .skip (f"{ moe_backend } backend supports SM 100 only" )
16841686
16851687 kv_cache_config = KvCacheConfig (free_gpu_memory_fraction = 0.75 )
16861688 torch_compile_config = TorchCompileConfig (
@@ -1767,7 +1769,7 @@ def test_nvfp4_batch_waiting(self, torch_compile, fp8kv, cuda_graph,
17671769 (2 , 2 , 1 ), (1 , 4 , 1 )],
17681770 ids = ["tp4" , "ep4" , "tp2pp2" , "pp4" ])
17691771 @parametrize_with_ids ("mtp_nextn" , [0 , 2 ])
1770- @parametrize_with_ids ("moe_backend" , ["CUTLASS" , "TRTLLM" ])
1772+ @parametrize_with_ids ("moe_backend" , ["CUTLASS" , "TRTLLM" , "CUTEDSL" ])
17711773 def test_nvfp4_4gpus (self , fp8kv , attention_dp , cuda_graph ,
17721774 overlap_scheduler , tp_size , pp_size , ep_size ,
17731775 torch_compile , mtp_nextn , moe_backend ):
@@ -1777,6 +1779,9 @@ def test_nvfp4_4gpus(self, fp8kv, attention_dp, cuda_graph,
17771779 or get_sm_version () == 121 ):
17781780 pytest .skip (
17791781 "MOE TRTLLM backend does not support SM version 120 or 121" )
1782+ if moe_backend == "CUTEDSL" and get_sm_version () != 100 :
1783+ pytest .skip (f"{ moe_backend } backend supports SM 100 only" )
1784+
17801785 kv_cache_config = KvCacheConfig (free_gpu_memory_fraction = 0.75 )
17811786 # Picewise Cuda Graph cannot be enabled for nvfp4 attention dp.
17821787 torch_compile_config = TorchCompileConfig (
0 commit comments