Training both NER and Neural Edit-Tree Lemmatization with a transformer #10554
Replies: 1 comment 4 replies
-
We are currently preparing for the edit-tree lemmatizer to be included in mainline spaCy, so it is likely that there will be models available in the future (hopefully spaCy 3.3). The models that Adriane Boyd trained for the UD benchmarks also use the edit tree lemmatizer: https://explosion.ai/blog/ud-benchmarks-v3-2
You can train them at the same time when you do have NER and lemmas as part of the same data set. Otherwise, you will have to train on both data sets separately, but you can still make them part of one configuration. For example, you could train the lemmatizer first and then create a configuration for NER, where you source the lemmatizer and set the lemmatizer as frozen to ensure that it doesn't get updated while training NER.
Which language/model are you referring to? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
I am trying to use the "Experimental" Neural Edit-Tree Lemmatizer from the Experimentation repo using a transformer (xlm-roberta-base). I did not find any model on GitHub along with the lemmatizer so I assume it needs to be trained. Ideally we would like to train it along with a NER training to use a single pipline for both lemmatization and NER
Is it possible to train both at the same time? If so, how would you set-up your training since the data set we have for NER does not have lemmatization and vice versa?
Finally which data set was used to train Neural Edit-Tree Lemmatizer with roberta-xml ?
Thanks !
Beta Was this translation helpful? Give feedback.
All reactions