Skip to content

Commit 489dd60

Browse files
authored
[#10513][fix] AutoDeploy: removed self.mlp_type leftovers from last moe refactor (#10512)
Signed-off-by: Eran Geva <[email protected]>
1 parent e033129 commit 489dd60

File tree

1 file changed

+1
-2
lines changed
  • tensorrt_llm/_torch/auto_deploy/transform/library

1 file changed

+1
-2
lines changed

tensorrt_llm/_torch/auto_deploy/transform/library/sharding.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -646,7 +646,6 @@ def apply(self, gm: GraphModule, node: Node) -> None:
646646
gm,
647647
node,
648648
self.config,
649-
self.mlp_type,
650649
scale_names=self.scale_names(),
651650
)
652651

@@ -664,7 +663,7 @@ def scale_names(self) -> List[str]:
664663
return ["input_scale", "weight_scale", "alpha"]
665664

666665
def apply(self, gm: GraphModule, node: Node) -> None:
667-
_insert_sharded_moe(gm, node, self.config, self.mlp_type, scale_names=self.scale_names())
666+
_insert_sharded_moe(gm, node, self.config, scale_names=self.scale_names())
668667

669668

670669
EP_SHARDING_RULES = [

0 commit comments

Comments
 (0)