Skip to content

Commit 3920faa

Browse files
committed
correct rope type
1 parent cff16cc commit 3920faa

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

src/llama-model.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14326,7 +14326,7 @@ struct llm_build_hunyuan_moe : public llm_graph_context {
1432614326

1432714327
auto * inp_attn = build_attn_inp_kv_unified();
1432814328

14329-
const float kq_scale = hparams.f_attention_scale == 0.0f ? 1.0f/sqrtf(float(n_embd_head)) : hparams.f_attention_scale;
14329+
const float kq_scale = 1.0f / sqrtf(float(n_embd_head));
1433014330

1433114331
ggml_tensor * inp_out_ids = build_inp_out_ids();
1433214332

@@ -14999,7 +14999,6 @@ llama_rope_type llama_model_rope_type(const llama_model * model) {
1499914999
case LLM_ARCH_BAILINGMOE:
1500015000
case LLM_ARCH_NEO_BERT:
1500115001
case LLM_ARCH_ARCEE:
15002-
case LLM_ARCH_HUNYUAN_MOE:
1500315002
return LLAMA_ROPE_TYPE_NORM;
1500415003

1500515004
// the pairs of head values are offset by n_rot/2
@@ -15035,6 +15034,7 @@ llama_rope_type llama_model_rope_type(const llama_model * model) {
1503515034
case LLM_ARCH_EXAONE:
1503615035
case LLM_ARCH_MINICPM3:
1503715036
case LLM_ARCH_DOTS1:
15037+
case LLM_ARCH_HUNYUAN_MOE:
1503815038
return LLAMA_ROPE_TYPE_NEOX;
1503915039

1504015040
case LLM_ARCH_QWEN2VL:

0 commit comments

Comments
 (0)