@@ -2718,7 +2718,7 @@ def set_gguf_parameters(self):
27182718 self .gguf_writer .add_add_bos_token (False )
27192719
27202720
2721- @Model .register ("Phi3ForCausalLM" , "Phi4MMForCausalLM" )
2721+ @Model .register ("Phi3ForCausalLM" )
27222722class Phi3MiniModel (Model ):
27232723 model_arch = gguf .MODEL_ARCH .PHI3
27242724
@@ -2729,7 +2729,7 @@ def set_vocab(self):
27292729 with open (tokenizer_config_file , "r" , encoding = "utf-8" ) as f :
27302730 tokenizer_config_json = json .load (f )
27312731 tokenizer_class = tokenizer_config_json ['tokenizer_class' ]
2732- if tokenizer_class == 'GPT2Tokenizer' or tokenizer_class == 'GPT2TokenizerFast' :
2732+ if tokenizer_class == 'GPT2Tokenizer' :
27332733 return self ._set_vocab_gpt2 ()
27342734
27352735 from sentencepiece import SentencePieceProcessor
@@ -2895,16 +2895,6 @@ def generate_extra_tensors(self) -> Iterable[tuple[str, Tensor]]:
28952895 yield (self .format_tensor_name (gguf .MODEL_TENSOR .ROPE_FACTORS_LONG ), torch .tensor (long_factors , dtype = torch .float32 ))
28962896 yield (self .format_tensor_name (gguf .MODEL_TENSOR .ROPE_FACTORS_SHORT ), torch .tensor (short_factors , dtype = torch .float32 ))
28972897
2898- def modify_tensors (self , data_torch : Tensor , name : str , bid : int | None ) -> Iterable [tuple [str , Tensor ]]:
2899- if "base_layer" in name :
2900- name = name .replace ("base_layer." , "" )
2901- # TODO: a big TODO, for simplification, we are skipping multimodal tensors for now
2902- if name .startswith ("model.embed_tokens_extend" ) or "lora_" in name :
2903- logger .error (f"Skipping multimodal tensor: { name !r} " )
2904- return []
2905- else :
2906- return super ().modify_tensors (data_torch , name , bid )
2907-
29082898
29092899@Model .register ("PhiMoEForCausalLM" )
29102900class PhiMoeModel (Phi3MiniModel ):
0 commit comments