Best practices for NER training with new entity types #8288
Replies: 1 comment 1 reply
-
As a note, when you ask "is it better to X or Y?" type questions, sometimes we can give you an answer based on our experience, but often the best answer is "try both and see what works". Modern machine learning is highly empirical, which means people try stuff until it works, and without getting really into your requirements and data it's hard to make general recommendations. That said I will try to give some advice.
If you have enough data it's usually better to train the tok2vec/transformer as well. Where the borderline for "enough data" is is unclear, but for transformers you need more data. If you have a lot of data there's generally not much benefit to fine tuning.
Do you care about the entity types that are not in your data? If not then I would skip fine tuning. Fine tuning is possible in the scenario you describe but will be complicated if you want to avoid forgetting the entities you aren't using. This depends a lot on the quantity of your data and how much it is like the data for the pretrained models.
I wouldn't try training a model with fewer than several hundred examples, though how much is enough depends a lot on the specifics of the entities. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
I have some questions about training a NER model. Although these questions are not totally new, I couldn’t find satisfactory answers in the previous discussions, and I think they could be useful also for other people.
Any suggestion will be appreciated.
Thank you!
Beta Was this translation helpful? Give feedback.
All reactions