Skip to content
Discussion options

You must be logged in to vote

My best guess is that one of the evaluation corpora has longer texts than the others, but the details would depend on your span_getters setting.

You probably need to add a model_max_length to your transformer model as described here: #7393 (comment)

You can probably achieve this by adding tokenizer_config.json with the right settings to the saved transformers/model directory without retraining. I haven't tested this, but it will probably look something like this:

{"model_max_length": 512}

Replies: 1 comment 3 replies

Comment options

You must be logged in to vote
3 replies
@nleguillarme
Comment options

@adrianeboyd
Comment options

@nleguillarme
Comment options

Answer selected by adrianeboyd
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feat / transformer Feature: Transformer
2 participants