Skip to content

Commit 4c1d9d0

Browse files
authored
[None][chore] Pass without_comm to cutlass and deepgemm (#11229)
Signed-off-by: xxi <xxi@nvidia.com>
1 parent 36cb5f8 commit 4c1d9d0

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

tensorrt_llm/_torch/modules/fused_moe/create_moe.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -196,6 +196,7 @@ def create_moe_backend(
196196
swiglu_beta=swiglu_beta,
197197
swiglu_limit=swiglu_limit,
198198
init_load_balancer=init_load_balancer,
199+
without_comm=without_comm,
199200
activation_type=activation_type,
200201
)
201202
elif moe_cls == WideEPMoE:
@@ -256,6 +257,7 @@ def create_moe_backend(
256257
weight_loading_mode=weight_loading_mode,
257258
apply_router_weight_on_input=apply_router_weight_on_input,
258259
layer_idx=layer_idx,
260+
without_comm=without_comm,
259261
)
260262
elif moe_cls == TritonFusedMoE:
261263
assert not apply_router_weight_on_input, "apply_router_weight_on_input is not supported in TritonFusedMoE."

0 commit comments

Comments
 (0)