@@ -1148,45 +1148,6 @@ def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iter
11481148 return tensors
11491149
11501150
1151- @Model .register ("PersimmonForCausalLM" )
1152- class PersimmonModel (Model ):
1153- model_arch = gguf .MODEL_ARCH .PERSIMMON
1154-
1155- def set_gguf_parameters (self ):
1156- block_count = self .hparams .get ("num_layers" , self .hparams .get ("num_hidden_layers" ))
1157- head_count = self .hparams ["num_attention_heads" ]
1158- head_count_kv = head_count
1159- hidden_size = self .hparams ["hidden_size" ]
1160-
1161- self .gguf_writer .add_name ('persimmon-8b-chat' )
1162- self .gguf_writer .add_context_length (self .hparams ["max_position_embeddings" ])
1163- self .gguf_writer .add_embedding_length (hidden_size )
1164- self .gguf_writer .add_block_count (block_count )
1165- self .gguf_writer .add_feed_forward_length (self .hparams ["intermediate_size" ])
1166-
1167- # NOTE: not sure about this change - why does the model not have a rope dimension count when it is smaller
1168- # than the head size?
1169- # ref: https://github.com/ggerganov/llama.cpp/pull/4889
1170- # self.gguf_writer.add_rope_dimension_count(hidden_size // head_count)
1171- self .gguf_writer .add_rope_dimension_count (hidden_size // head_count // 2 )
1172-
1173- self .gguf_writer .add_head_count (head_count )
1174- self .gguf_writer .add_head_count_kv (head_count_kv )
1175- self .gguf_writer .add_rope_freq_base (self .hparams ["rope_theta" ])
1176- self .gguf_writer .add_layer_norm_eps (self .hparams ["layer_norm_eps" ])
1177-
1178- def set_vocab (self ):
1179- self ._set_vocab_sentencepiece ()
1180- # self.gguf_writer.add_bos_token_id(71013)
1181- # self.gguf_writer.add_eos_token_id(71013)
1182-
1183- def extra_f32_tensors (self , name : str , new_name : str , bid : int | None , n_dims : int ) -> bool :
1184- del name , new_name , bid , n_dims # unused
1185-
1186- # TODO: FP16 conversion produces garbage outputs. (Q8_0 does not, so..?)
1187- return True
1188-
1189-
11901151@Model .register ("StableLmForCausalLM" , "StableLMEpochForCausalLM" , "LlavaStableLMEpochForCausalLM" )
11911152class StableLMModel (Model ):
11921153 model_arch = gguf .MODEL_ARCH .STABLELM
0 commit comments