@@ -281,8 +281,10 @@ def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iter
281281 return [(self .map_tensor_name (name ), data_torch )]
282282
283283 def tensor_force_quant (self , name : str , new_name : str , bid : int | None , n_dims : int ) -> gguf .GGMLQuantizationType | bool :
284+ # 删除未使用的参数,避免警告或错误
284285 del name , new_name , bid , n_dims # unused
285286
287+ # 返回False,表示不进行量化
286288 return False
287289
288290 # some models need extra generated tensors (like rope_freqs)
@@ -5053,14 +5055,18 @@ def set_gguf_parameters(self):
50535055 n_embed = self .hparams .get ("hidden_size" , self .hparams .get ("n_embed" ))
50545056 n_head = self .hparams .get ("n_head" , self .hparams .get ("num_attention_heads" ))
50555057 n_head_kv = self .hparams .get ("multi_query_group_num" , self .hparams .get ("num_key_value_heads" , n_head ))
5056- self .gguf_writer .add_context_length (self .hparams .get ("seq_length" , n_embed ))
5058+ if (n_ctx := self .find_hparam (["max_position_embeddings" , "n_ctx" , "seq_length" ], optional = True )) is not None :
5059+ self .gguf_writer .add_context_length (n_ctx )
5060+ else :
5061+ self .gguf_writer .add_context_length (n_embed )
50575062 self .gguf_writer .add_embedding_length (n_embed )
50585063 self .gguf_writer .add_feed_forward_length (self .hparams .get ("ffn_hidden_size" , self .hparams .get ("intermediate_size" , 4 * n_embed )))
50595064 self .gguf_writer .add_block_count (self .hparams .get ("num_layers" , self .hparams ["num_hidden_layers" ]))
50605065 self .gguf_writer .add_head_count (n_head )
50615066 self .gguf_writer .add_head_count_kv (n_head_kv )
50625067 self .gguf_writer .add_layer_norm_rms_eps (self .hparams .get ("layernorm_epsilon" ,1e-5 ))
50635068 self .gguf_writer .add_file_type (self .ftype )
5069+ self .gguf_writer .set_gguf_parameters (self )
50645070 if "attention_dim" in self .hparams :
50655071 rope_dim = self .hparams ["attention_dim" ]
50665072 else :
0 commit comments