@@ -2145,7 +2145,6 @@ def __init__(self, *args, **kwargs):
21452145
21462146 def set_vocab (self ):
21472147 self ._set_vocab_gpt2 ()
2148- self .gguf_writer .add_add_bos_token (True )
21492148
21502149 def set_gguf_parameters (self ):
21512150 super ().set_gguf_parameters ()
@@ -3918,9 +3917,6 @@ def _xlmroberta_set_vocab(self) -> None:
39183917 special_vocab = gguf .SpecialVocab (self .dir_model , n_vocab = len (tokens ))
39193918 special_vocab .add_to_gguf (self .gguf_writer )
39203919
3921- self .gguf_writer .add_add_bos_token (True )
3922- self .gguf_writer .add_add_eos_token (True )
3923-
39243920
39253921@ModelBase .register ("DistilBertModel" , "DistilBertForMaskedLM" , "DistilBertForSequenceClassification" )
39263922class DistilBertModel (BertModel ):
@@ -3962,8 +3958,6 @@ def set_vocab(self):
39623958 bpe_tok_path = self .dir_model / "tokenizer.json"
39633959 if bpe_tok_path .exists ():
39643960 self ._set_vocab_gpt2 ()
3965- self .gguf_writer .add_add_bos_token (True )
3966- self .gguf_writer .add_add_eos_token (True )
39673961
39683962 # we need this to validate the size of the token_type embeddings
39693963 # though currently we are passing all zeros to the token_type embeddings
@@ -4848,8 +4842,6 @@ def set_vocab(self):
48484842 self .gguf_writer .add_token_type_count (2 )
48494843 else :
48504844 raise NotImplementedError (f'Tokenizer { tokenizer_class } is not supported for JinaBertModel' )
4851- self .gguf_writer .add_add_bos_token (True )
4852- self .gguf_writer .add_add_eos_token (True )
48534845
48544846
48554847@ModelBase .register ("OpenELMForCausalLM" )
@@ -5451,9 +5443,6 @@ def set_vocab(self):
54515443 special_vocab = gguf .SpecialVocab (self .dir_model , n_vocab = len (tokens ))
54525444 special_vocab .add_to_gguf (self .gguf_writer )
54535445
5454- self .gguf_writer .add_add_bos_token (False )
5455- self .gguf_writer .add_add_eos_token (True )
5456-
54575446 def set_gguf_parameters (self ):
54585447 if (n_ctx := self .find_hparam (["n_positions" ], optional = True )) is None :
54595448 logger .warning ("Couldn't find context length in config.json, assuming default value of 512" )
@@ -5591,9 +5580,6 @@ def set_vocab(self):
55915580 special_vocab = gguf .SpecialVocab (self .dir_model , n_vocab = len (tokens ))
55925581 special_vocab .add_to_gguf (self .gguf_writer )
55935582
5594- self .gguf_writer .add_add_bos_token (False )
5595- self .gguf_writer .add_add_eos_token (True )
5596-
55975583 def set_gguf_parameters (self ):
55985584 if (n_ctx := self .find_hparam (["n_positions" ], optional = True )) is None :
55995585 logger .warning ("Couldn't find context length in config.json, assuming default value of 512" )
@@ -6389,8 +6375,8 @@ def parse_args() -> argparse.Namespace:
63896375 help = "model is executed on big endian machine" ,
63906376 )
63916377 parser .add_argument (
6392- "model" , type = Path ,
6393- help = "directory containing model file" ,
6378+ "model" , type = str ,
6379+ help = "directory containing model file or huggingface repository ID (if --remote) " ,
63946380 nargs = "?" ,
63956381 )
63966382 parser .add_argument (
@@ -6493,18 +6479,20 @@ def main() -> None:
64936479 else :
64946480 logging .basicConfig (level = logging .INFO )
64956481
6496- dir_model = args .model
6497-
64986482 if args .remote :
6483+ hf_repo_id = args .model
64996484 from huggingface_hub import snapshot_download
65006485 local_dir = snapshot_download (
6501- repo_id = str ( dir_model ) ,
6486+ repo_id = hf_repo_id ,
65026487 allow_patterns = ["LICENSE" , "*.json" , "*.md" , "*.txt" , "tokenizer.model" ])
65036488 dir_model = Path (local_dir )
65046489 logger .info (f"Downloaded config and tokenizer to { local_dir } " )
6490+ else :
6491+ hf_repo_id = None
6492+ dir_model = Path (args .model )
65056493
65066494 if not dir_model .is_dir ():
6507- logger .error (f'Error: { args . model } is not a directory' )
6495+ logger .error (f'Error: { dir_model } is not a directory' )
65086496 sys .exit (1 )
65096497
65106498 ftype_map : dict [str , gguf .LlamaFileType ] = {
@@ -6524,9 +6512,9 @@ def main() -> None:
65246512
65256513 if args .outfile is not None :
65266514 fname_out = args .outfile
6527- elif args . remote :
6515+ elif hf_repo_id :
65286516 # if remote, use the model ID as the output file name
6529- fname_out = Path ("./" + str ( args . model ) .replace ("/" , "-" ) + "-{ftype}.gguf" )
6517+ fname_out = Path ("./" + hf_repo_id .replace ("/" , "-" ) + "-{ftype}.gguf" )
65306518 else :
65316519 fname_out = dir_model
65326520
@@ -6555,7 +6543,7 @@ def main() -> None:
65556543 split_max_tensors = args .split_max_tensors ,
65566544 split_max_size = split_str_to_n_bytes (args .split_max_size ), dry_run = args .dry_run ,
65576545 small_first_shard = args .no_tensor_first_split ,
6558- remote_hf_model_id = str ( args . model ) if args . remote else None )
6546+ remote_hf_model_id = hf_repo_id )
65596547
65606548 if args .vocab_only :
65616549 logger .info ("Exporting model vocab..." )
0 commit comments