Skip to content
Discussion options

You must be logged in to vote

I haven't tested this, but in general I think the right place to do this would be in the after_init callback.

I haven't done much training/testing with gpt2 models, but I can see that in our tests we just use "<|endoftext|>" as the pad token, which is already in the vocab.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@noobistz
Comment options

Answer selected by noobistz
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feat / transformer Feature: Transformer
2 participants