Skip to content

Commit bb23dd0

Browse files
committed
Pass and read n_ff_exp
1 parent 07a5c76 commit bb23dd0

File tree

2 files changed

+3
-0
lines changed

2 files changed

+3
-0
lines changed

convert_hf_to_gguf.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2828,6 +2828,8 @@ def set_gguf_parameters(self):
28282828
self.gguf_writer.add_expert_count(self.hparams["moe_num_experts"])
28292829
self.gguf_writer.add_expert_used_count(self.hparams["moe_k"])
28302830
self.gguf_writer.add_interleave_moe_layer_step(self.hparams["moe_layer_interval"])
2831+
if (moe_intermediate_size := self.hparams.get("moe_intermediate_size")) is not None:
2832+
self.gguf_writer.add_expert_feed_forward_length(moe_intermediate_size)
28312833

28322834
def tensor_force_quant(self, name: str, new_name: str, bid: int | None, n_dims: int) -> gguf.GGMLQuantizationType | bool:
28332835
if "exps" in new_name:

src/llama-model.cpp

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1611,6 +1611,7 @@ void llama_model::load_hparams(llama_model_loader & ml) {
16111611
{
16121612
ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
16131613
if (arch == LLM_ARCH_ERNIE4_5_MOE) {
1614+
ml.get_key(LLM_KV_EXPERT_FEED_FORWARD_LENGTH, hparams.n_ff_exp);
16141615
ml.get_key(LLM_KV_INTERLEAVE_MOE_LAYER_STEP, hparams.n_moe_layer_step);
16151616
}
16161617
switch (hparams.n_layer) {

0 commit comments

Comments
 (0)