Trained model returning ==Weblinks== #156
Unanswered
robinlaegler
asked this question in
Q&A
Replies: 1 comment 5 replies
-
|
I don't think batch size is likely the issue. How much data do you have? Increasing would probably help. Also how many epochs did you train for? I have gotten good results around 10k. |
Beta Was this translation helpful? Give feedback.
5 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, i trained a en-de model using the OPUS-Wikipedia Dataset and ~10.000 sentences of my own training data. I dropped the batch size to 800. I trained it on GoogleColab , because I don't have a GPU laying around. The training finished in about 5 hours. The model is pretty bad and it sometimes returns "==Weblinks==" or "==Einzelnachweise==".
Does anybody dealt with the same problem or knows where these tags are coming from?
And are there any tips to train a better model? Do I need more data or a higher batch size? Which is more important?
Beta Was this translation helpful? Give feedback.
All reactions