Skip to content
Discussion options

You must be logged in to vote

Hi!

I understand that you want to train a tagger, parser, ner and trainable_lemmatizer all using the same transformer model, correct?
As one alternative to merging the datasets and running the training simultaneously, you could also train the NER separately, then source its transformer and ner, freeze both and train the other components on top of it. Or vice versa, train the other components with a transformer, then source all of it, freeze everything and train an NER model. Even if this is not what you set out to do initially, this might actually obtain better performance than working with partially annotated datasets for all of the components.

If you do want to stick to the original set…

Replies: 2 comments 9 replies

Comment options

You must be logged in to vote
9 replies
@emiltj
Comment options

@adrianeboyd
Comment options

@emiltj
Comment options

@adrianeboyd
Comment options

@emiltj
Comment options

Answer selected by emiltj
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
training Training and updating models feat / ner Feature: Named Entity Recognizer feat / tagger Feature: Part-of-speech tagger feat / lemmatizer Feature: Rule-based and lookup lemmatization feat / transformer Feature: Transformer
3 participants