Skip to content

Commit c1a38c6

Browse files
committed
model: add KORMoForCausalLM
1 parent cf968a7 commit c1a38c6

File tree

2 files changed

+5
-3
lines changed

2 files changed

+5
-3
lines changed

convert_hf_to_gguf.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2370,7 +2370,8 @@ def prepare_tensors(self):
23702370
"VLlama3ForCausalLM",
23712371
"LlavaForConditionalGeneration",
23722372
"VoxtralForConditionalGeneration",
2373-
"LlamaModel")
2373+
"LlamaModel",
2374+
"KORMoForCausalLM")
23742375
class LlamaModel(TextModel):
23752376
model_arch = gguf.MODEL_ARCH.LLAMA
23762377
undo_permute = True
@@ -8237,7 +8238,6 @@ def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iter
82378238

82388239
return [(self.map_tensor_name(name), data_torch)]
82398240

8240-
82418241
@ModelBase.register("ExaoneForCausalLM")
82428242
class ExaoneModel(TextModel):
82438243
model_arch = gguf.MODEL_ARCH.EXAONE

gguf-py/gguf/tensor_mapping.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -154,7 +154,8 @@ class TensorNameMap:
154154
"model.layers.{bid}.operator_norm", # lfm2
155155
"model.transformer.blocks.{bid}.attn_norm", # llada
156156
"layers.{bid}.input_layernorm", # qwen3-embedding
157-
"model.layers.{bid}.attention_layernorm" # apertus
157+
"model.layers.{bid}.attention_layernorm", # apertus
158+
"model.layers.{bid}.pre_attention_layernorm", # kormo
158159
),
159160

160161
# Attention norm 2
@@ -342,6 +343,7 @@ class TensorNameMap:
342343
"model.transformer.blocks.{bid}.ff_norm", # llada
343344
"layers.{bid}.post_attention_layernorm", # qwen3-embedding
344345
"model.layers.{bid}.feedforward_layernorm", # apertus
346+
"model.layers.{bid}.pre_mlp_layernorm", # kormo
345347
),
346348

347349
# Pre feed-forward norm

0 commit comments

Comments
 (0)