Skip to content

Commit 0cd2bc6

Browse files
authored
add self.global_num_experts for moe layer plugin (vllm-project#1769)
@czhu15 @Wei-Lin-Intel
1 parent e9f568c commit 0cd2bc6

File tree

1 file changed

+3
-0
lines changed
  • vllm/model_executor/layers/fused_moe

1 file changed

+3
-0
lines changed

vllm/model_executor/layers/fused_moe/layer.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -956,6 +956,7 @@ def __init__(
956956
UnquantizedFusedMoEMethod):
957957
moe_op = VllmMixtureOfExpertsOp(
958958
num_experts,
959+
self.global_num_experts,
959960
experts_min,
960961
experts_max,
961962
)
@@ -964,12 +965,14 @@ def __init__(
964965
) and not envs.VLLM_HPU_FORCE_CHANNEL_FP8:
965966
moe_op = VllmMixtureOfExpertsOpFP8(
966967
num_experts,
968+
self.global_num_experts,
967969
experts_min,
968970
experts_max,
969971
)
970972
else:
971973
moe_op = VllmMixtureOfExpertsOpFP8PerChannel(
972974
num_experts,
975+
self.global_num_experts,
973976
experts_min,
974977
experts_max,
975978
)

0 commit comments

Comments
 (0)