Skip to content

Commit e53632b

Browse files
committed
fix: Remove unnecessary duplicate keys in converter
Co-authored-by: Francis Couture-Harpin <[email protected]> (thanks for the sharp eyes and patience!) Branch: GraniteFour Signed-off-by: Gabe Goodhart <[email protected]>
1 parent 04883fc commit e53632b

File tree

1 file changed

+0
-18
lines changed

1 file changed

+0
-18
lines changed

convert_hf_to_gguf.py

Lines changed: 0 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -6552,23 +6552,11 @@ def set_gguf_parameters(self):
65526552
specific to this model. The result is some duplication of how the params
65536553
get set. The following warnings are expected during conversion:
65546554
6555-
WARNING:Duplicated key name 'granitehybrid.embedding_length'
6556-
WARNING:Duplicated key name 'granitehybrid.block_count'
6557-
WARNING:Duplicated key name 'granitehybrid.vocab_size'
6558-
WARNING:Duplicated key name 'granitehybrid.feed_forward_length'
6559-
WARNING:Duplicated key name 'granitehybrid.attention.head_count'
65606555
WARNING:Duplicated key name 'granitehybrid.attention.head_count_kv'
6561-
WARNING:Duplicated key name 'granitehybrid.attention.layer_norm_rms_epsilon'
65626556
WARNING:Duplicated key name 'granitehybrid.context_length'
65636557
"""
65646558
GraniteMoeModel.set_gguf_parameters(self)
65656559

6566-
## General Params ##
6567-
self.gguf_writer.add_embedding_length(self.d_model)
6568-
self.gguf_writer.add_block_count(self.block_count)
6569-
self.gguf_writer.add_vocab_size(self.hparams["vocab_size"])
6570-
self.gguf_writer.add_feed_forward_length(self.hparams["intermediate_size"])
6571-
65726560
## Mamba mixer params ##
65736561
self.gguf_writer.add_ssm_conv_kernel(self.find_hparam(["conv_kernel", "d_conv"]))
65746562
self.gguf_writer.add_ssm_state_size(self.find_hparam(["state_size", "d_state"]))
@@ -6585,14 +6573,8 @@ def set_gguf_parameters(self):
65856573
]
65866574
if rope_dim := self.hparams.get("attn_rotary_emb"):
65876575
self.gguf_writer.add_rope_dimension_count(rope_dim)
6588-
self.gguf_writer.add_head_count(self.hparams["num_attention_heads"])
65896576
self.gguf_writer.add_head_count_kv(head_count_kv_vec)
65906577

6591-
## Feed Forward Params ##
6592-
self.gguf_writer.add_layer_norm_rms_eps(
6593-
self.find_hparam(["layer_norm_epsilon", "rms_norm_eps"], optional=True) or 1e-5
6594-
)
6595-
65966578
## If Bamba, use rope, otherwise don't
65976579
use_rope = "BambaForCausalLM" in self.hparams["architectures"]
65986580
self.gguf_writer.add_rope_scaling_finetuned(use_rope)

0 commit comments

Comments
 (0)