Skip to content

Commit 8735824

Browse files
zyongyediegocastanibm
authored andcommitted
[gpt-oss] fix model config with hf_config (vllm-project#22401)
Signed-off-by: Yongye Zhu <[email protected]> Signed-off-by: Diego-Castan <[email protected]>
1 parent 5178be0 commit 8735824

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

vllm/model_executor/models/gpt_oss.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -61,9 +61,9 @@ def __init__(
6161
"original_max_position_embeddings":
6262
config.rope_scaling["original_max_position_embeddings"],
6363
"beta_fast":
64-
config.rope_ntk_beta,
64+
config.rope_scaling["beta_fast"],
6565
"beta_slow":
66-
config.rope_ntk_alpha,
66+
config.rope_scaling["beta_slow"],
6767
},
6868
is_neox_style=True,
6969
)
@@ -154,7 +154,7 @@ def __init__(
154154
dtype=torch.bfloat16)
155155
assert config.intermediate_size % self.world_size == 0
156156
self.experts = FusedMoE(num_experts=config.num_local_experts,
157-
top_k=config.num_experts_per_token,
157+
top_k=config.num_experts_per_tok,
158158
hidden_size=config.hidden_size,
159159
intermediate_size=config.intermediate_size,
160160
reduce_results=True,

0 commit comments

Comments
 (0)