'en_core_web_trf' optimal optimizer's learning rate and number of training epochs? #7066
traceymai
started this conversation in
Language Support
Replies: 1 comment 4 replies
-
Also, the reason I had to change the default learning rate for this model is that if I just get my optimizer from |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi guys, I've been experimenting with spaCy v3.0's pretrained models
en_core_web_trf
anden_core_web_lg
(in particular I was using thetextcat
pipeline for sentiment analysis). From experimenting with learning rate settings (my optimizer is Adam v1) of 0.01 and 0.001, my transformers model's accuracy increased from 76% to 80% with 10 training epochs, which in my opinion is pretty good. However I tried to switch up the number of training epochs abit to 15 (loss has been decreasing throughout all 15 training epochs, which I'd think is good?), and the accuracy surprisingly dropped to 78% again. I was just wondering if in your experience there's any way to determine (or if there exists) an optimal learning rate and an optimal number of training epochs I should train this trf model for? Any idea is greatly welcome!Beta Was this translation helpful? Give feedback.
All reactions