@@ -1476,6 +1476,16 @@ def _try_set_pooling_type(self) -> None:
14761476 raise NotImplementedError ("Only MEAN, CLS, and LAST pooling types supported" )
14771477 self .gguf_writer .add_pooling_type (pooling_type )
14781478
1479+ def _set_vocab_glmedge (self ):
1480+ from transformers import AutoTokenizer
1481+ tokenizer = AutoTokenizer .from_pretrained (self .dir_model , trust_remote_code = True )
1482+ special_vocab = gguf .SpecialVocab (self .dir_model , load_merges = True )
1483+ special_vocab ._set_special_token ("eos" , tokenizer .get_added_vocab ()["<|endoftext|>" ])
1484+ special_vocab ._set_special_token ("eot" , tokenizer .get_added_vocab ()["<|user|>" ])
1485+ special_vocab ._set_special_token ("unk" , tokenizer .get_added_vocab ()["<|endoftext|>" ])
1486+ special_vocab ._set_special_token ("bos" , tokenizer .get_added_vocab ()["<|endoftext|>" ])
1487+ special_vocab .add_to_gguf (self .gguf_writer )
1488+
14791489 def _set_vocab_interns1 (self ):
14801490 tokens : list [str ] = []
14811491 toktypes : list [int ] = []
@@ -2359,6 +2369,7 @@ def prepare_tensors(self):
23592369 "VLlama3ForCausalLM" ,
23602370 "LlavaForConditionalGeneration" ,
23612371 "VoxtralForConditionalGeneration" ,
2372+ "GlmasrModel" ,
23622373 "LlamaModel" )
23632374class LlamaModel (TextModel ):
23642375 model_arch = gguf .MODEL_ARCH .LLAMA
@@ -2410,6 +2421,8 @@ def set_vocab(self):
24102421 # Apply to granite small models only
24112422 if self .hparams .get ("vocab_size" , 32000 ) == 49152 :
24122423 self .gguf_writer .add_add_bos_token (False )
2424+ if self .hf_arch == "GlmasrModel" :
2425+ self ._set_vocab_glmedge ()
24132426
24142427 def set_gguf_parameters (self ):
24152428 super ().set_gguf_parameters ()
@@ -2564,22 +2577,6 @@ def set_gguf_parameters(self):
25642577 self .gguf_writer .add_rope_scaling_orig_ctx_len (rope_scaling ["original_max_position_embeddings" ])
25652578
25662579
2567- @ModelBase .register ("GlmasrModel" )
2568- class GlmasrModel (LlamaModel ):
2569- model_arch = gguf .MODEL_ARCH .LLAMA
2570-
2571- def set_vocab (self ):
2572- super ().set_vocab ()
2573- from transformers import AutoTokenizer
2574- tokenizer = AutoTokenizer .from_pretrained (self .dir_model , trust_remote_code = True )
2575- special_vocab = gguf .SpecialVocab (self .dir_model , load_merges = True )
2576- special_vocab ._set_special_token ("eos" , tokenizer .get_added_vocab ()["<|endoftext|>" ])
2577- special_vocab ._set_special_token ("eot" , tokenizer .get_added_vocab ()["<|user|>" ])
2578- special_vocab ._set_special_token ("unk" , tokenizer .get_added_vocab ()["<|endoftext|>" ])
2579- special_vocab ._set_special_token ("bos" , tokenizer .get_added_vocab ()["<|endoftext|>" ])
2580- special_vocab .add_to_gguf (self .gguf_writer )
2581-
2582-
25832580@ModelBase .register ("AfmoeForCausalLM" )
25842581class AfmoeModel (LlamaModel ):
25852582 model_arch = gguf .MODEL_ARCH .AFMOE
0 commit comments