Skip to content

Commit 34cc679

Browse files
committed
Merge remote-tracking branch 'other/xsn/hunyuan-moe' into hunyuan
2 parents 8fd547b + 3920faa commit 34cc679

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

src/llama-model.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14326,7 +14326,7 @@ struct llm_build_hunyuan_moe : public llm_graph_context {
1432614326

1432714327
auto * inp_attn = build_attn_inp_kv_unified();
1432814328

14329-
const float kq_scale = hparams.f_attention_scale == 0.0f ? 1.0f/sqrtf(float(n_embd_head)) : hparams.f_attention_scale;
14329+
const float kq_scale = 1.0f / sqrtf(float(n_embd_head));
1433014330

1433114331
ggml_tensor * inp_out_ids = build_inp_out_ids();
1433214332

@@ -15001,7 +15001,6 @@ llama_rope_type llama_model_rope_type(const llama_model * model) {
1500115001
case LLM_ARCH_BAILINGMOE:
1500215002
case LLM_ARCH_NEO_BERT:
1500315003
case LLM_ARCH_ARCEE:
15004-
case LLM_ARCH_HUNYUAN_MOE:
1500515004
return LLAMA_ROPE_TYPE_NORM;
1500615005

1500715006
// the pairs of head values are offset by n_rot/2
@@ -15037,6 +15036,7 @@ llama_rope_type llama_model_rope_type(const llama_model * model) {
1503715036
case LLM_ARCH_EXAONE:
1503815037
case LLM_ARCH_MINICPM3:
1503915038
case LLM_ARCH_DOTS1:
15039+
case LLM_ARCH_HUNYUAN_MOE:
1504015040
return LLAMA_ROPE_TYPE_NEOX;
1504115041

1504215042
case LLM_ARCH_QWEN2VL:

0 commit comments

Comments
 (0)