We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 83f7bbb commit d3c18c9Copy full SHA for d3c18c9
vllm/model_executor/models/glm4_moe.py
@@ -125,9 +125,8 @@ def __init__(
125
quant_config=None,
126
prefix=f"{prefix}.gate")
127
128
- # noaux_tc is not set in transformers new config now
129
- self.gate.e_score_correction_bias = (nn.Parameter(
130
- torch.empty(config.n_routed_experts)))
+ self.gate.e_score_correction_bias = nn.Parameter(
+ torch.empty(config.n_routed_experts, dtype=torch.float32))
131
132
# Load balancing settings.
133
vllm_config = get_current_vllm_config()
0 commit comments