Skip to content

Commit 98cf99f

Browse files
committed
check root architecture for convert hf script
1 parent 103e894 commit 98cf99f

File tree

1 file changed

+13
-16
lines changed

1 file changed

+13
-16
lines changed

convert_hf_to_gguf.py

Lines changed: 13 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -1476,6 +1476,16 @@ def _try_set_pooling_type(self) -> None:
14761476
raise NotImplementedError("Only MEAN, CLS, and LAST pooling types supported")
14771477
self.gguf_writer.add_pooling_type(pooling_type)
14781478

1479+
def _set_vocab_glmedge(self):
1480+
from transformers import AutoTokenizer
1481+
tokenizer = AutoTokenizer.from_pretrained(self.dir_model, trust_remote_code=True)
1482+
special_vocab = gguf.SpecialVocab(self.dir_model, load_merges=True)
1483+
special_vocab._set_special_token("eos", tokenizer.get_added_vocab()["<|endoftext|>"])
1484+
special_vocab._set_special_token("eot", tokenizer.get_added_vocab()["<|user|>"])
1485+
special_vocab._set_special_token("unk", tokenizer.get_added_vocab()["<|endoftext|>"])
1486+
special_vocab._set_special_token("bos", tokenizer.get_added_vocab()["<|endoftext|>"])
1487+
special_vocab.add_to_gguf(self.gguf_writer)
1488+
14791489
def _set_vocab_interns1(self):
14801490
tokens: list[str] = []
14811491
toktypes: list[int] = []
@@ -2359,6 +2369,7 @@ def prepare_tensors(self):
23592369
"VLlama3ForCausalLM",
23602370
"LlavaForConditionalGeneration",
23612371
"VoxtralForConditionalGeneration",
2372+
"GlmasrModel",
23622373
"LlamaModel")
23632374
class LlamaModel(TextModel):
23642375
model_arch = gguf.MODEL_ARCH.LLAMA
@@ -2410,6 +2421,8 @@ def set_vocab(self):
24102421
# Apply to granite small models only
24112422
if self.hparams.get("vocab_size", 32000) == 49152:
24122423
self.gguf_writer.add_add_bos_token(False)
2424+
if self.hf_arch == "GlmasrModel":
2425+
self._set_vocab_glmedge()
24132426

24142427
def set_gguf_parameters(self):
24152428
super().set_gguf_parameters()
@@ -2564,22 +2577,6 @@ def set_gguf_parameters(self):
25642577
self.gguf_writer.add_rope_scaling_orig_ctx_len(rope_scaling["original_max_position_embeddings"])
25652578

25662579

2567-
@ModelBase.register("GlmasrModel")
2568-
class GlmasrModel(LlamaModel):
2569-
model_arch = gguf.MODEL_ARCH.LLAMA
2570-
2571-
def set_vocab(self):
2572-
super().set_vocab()
2573-
from transformers import AutoTokenizer
2574-
tokenizer = AutoTokenizer.from_pretrained(self.dir_model, trust_remote_code=True)
2575-
special_vocab = gguf.SpecialVocab(self.dir_model, load_merges=True)
2576-
special_vocab._set_special_token("eos", tokenizer.get_added_vocab()["<|endoftext|>"])
2577-
special_vocab._set_special_token("eot", tokenizer.get_added_vocab()["<|user|>"])
2578-
special_vocab._set_special_token("unk", tokenizer.get_added_vocab()["<|endoftext|>"])
2579-
special_vocab._set_special_token("bos", tokenizer.get_added_vocab()["<|endoftext|>"])
2580-
special_vocab.add_to_gguf(self.gguf_writer)
2581-
2582-
25832580
@ModelBase.register("AfmoeForCausalLM")
25842581
class AfmoeModel(LlamaModel):
25852582
model_arch = gguf.MODEL_ARCH.AFMOE

0 commit comments

Comments
 (0)