@@ -6552,23 +6552,11 @@ def set_gguf_parameters(self):
65526552 specific to this model. The result is some duplication of how the params
65536553 get set. The following warnings are expected during conversion:
65546554
6555- WARNING:Duplicated key name 'granitehybrid.embedding_length'
6556- WARNING:Duplicated key name 'granitehybrid.block_count'
6557- WARNING:Duplicated key name 'granitehybrid.vocab_size'
6558- WARNING:Duplicated key name 'granitehybrid.feed_forward_length'
6559- WARNING:Duplicated key name 'granitehybrid.attention.head_count'
65606555 WARNING:Duplicated key name 'granitehybrid.attention.head_count_kv'
6561- WARNING:Duplicated key name 'granitehybrid.attention.layer_norm_rms_epsilon'
65626556 WARNING:Duplicated key name 'granitehybrid.context_length'
65636557 """
65646558 GraniteMoeModel .set_gguf_parameters (self )
65656559
6566- ## General Params ##
6567- self .gguf_writer .add_embedding_length (self .d_model )
6568- self .gguf_writer .add_block_count (self .block_count )
6569- self .gguf_writer .add_vocab_size (self .hparams ["vocab_size" ])
6570- self .gguf_writer .add_feed_forward_length (self .hparams ["intermediate_size" ])
6571-
65726560 ## Mamba mixer params ##
65736561 self .gguf_writer .add_ssm_conv_kernel (self .find_hparam (["conv_kernel" , "d_conv" ]))
65746562 self .gguf_writer .add_ssm_state_size (self .find_hparam (["state_size" , "d_state" ]))
@@ -6585,14 +6573,8 @@ def set_gguf_parameters(self):
65856573 ]
65866574 if rope_dim := self .hparams .get ("attn_rotary_emb" ):
65876575 self .gguf_writer .add_rope_dimension_count (rope_dim )
6588- self .gguf_writer .add_head_count (self .hparams ["num_attention_heads" ])
65896576 self .gguf_writer .add_head_count_kv (head_count_kv_vec )
65906577
6591- ## Feed Forward Params ##
6592- self .gguf_writer .add_layer_norm_rms_eps (
6593- self .find_hparam (["layer_norm_epsilon" , "rms_norm_eps" ], optional = True ) or 1e-5
6594- )
6595-
65966578 ## If Bamba, use rope, otherwise don't
65976579 use_rope = "BambaForCausalLM" in self .hparams ["architectures" ]
65986580 self .gguf_writer .add_rope_scaling_finetuned (use_rope )
0 commit comments