Variety in lemmatizer output #10632
-
I am training a NER model that also needs to have a lemmatizer. This lemmatizer I initialized with the source model en_core_web_sm as that is fine and does not need to be trained in my use case. The NER part works well, however, I have some questions about some of the behaviour of the lemmatizer.
So I was wondering, can this behaviour be explained by the lemmatizer being context dependent or are these artifacts of some mistake during training? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Hello, |
Beta Was this translation helpful? Give feedback.
Hello,
you also need to freeze the
tok2vec
component, since all other components sourced fromen_core_web_sm
are listening to it.The lemmatizer depends on the
tagger
, so you can also verify whether the POS tags are also correct.