Batch Size for NER #8301
Replies: 1 comment 7 replies
-
Great to hear you're seeing such a performance boost when switching to spaCy 3! 20 percent point is quite an unexpectedly huge leap though, but perhaps your dataset isn't very large? Because in that case, larger variations are expected. With respect to the batch training size, these defaults have simply kind of worked for us when benchmarking our pretrained pipelines on standard evaluation sets. It's always difficult to recommend good default settings in general, as every use-case and dataset is different. But you can try out some different runs, varying the batch size in the config, and see whether that makes any difference in your specific project. Feel free to report the results here, as I'm sure others may be interested as well! |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
First off thank you! After changing architecture from following Spacy 2.0 to 3.0 saw a 20 point increase in F1-Score!
My question is about the training batch size for NER, 100 start and 1000 stop seems large, why is the default that high and have you seen success with lesser numbers?
Beta Was this translation helpful? Give feedback.
All reactions