We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 6eb541d commit 24f7e42Copy full SHA for 24f7e42
tuning/sft_trainer.py
@@ -290,8 +290,10 @@ def train(
290
)
291
if tokenizer.eos_token != configs.DEFAULT_PAD_TOKEN:
292
tokenizer.pad_token = configs.DEFAULT_PAD_TOKEN
293
+ special_tokens_dict["pad_token"] = configs.DEFAULT_PAD_TOKEN
294
else:
295
tokenizer.eos_token = configs.DEFAULT_EOS_TOKEN
296
+ special_tokens_dict["eos_token"] = configs.DEFAULT_EOS_TOKEN
297
298
# TODO: lower priority but understand if resizing impacts inference quality and why its needed.
299
# It makes sense if we manipulate tokenizer that we also save it and provide it to inference.
0 commit comments