Error while Training NER using transformers #10510
-
Hi. I have been trying to train my custom NER component with the Transformer embeddings (albert-base-v2) configuration. however, when I am trying to run the training using config files via the terminal on my GPU, it throws a Runtime OOM error:
The error pops up at the start itself, even before executing for 1 epoch. I found some answers online which spoke about clearing cache via code, but since I am training via config files, how do I resolve this?I even tried decreasing the batch size but it didn't have any effect. Any help would be much appreciated. Thank you! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Hello, |
Beta Was this translation helpful? Give feedback.
Hello,
If I see it correctly, the machine that you're currently using has only 2GB of memory available for training. This is not suited for training with transformers embeddings. You might get it working by using very small batch sizes and small transformer models, but I wouldn't recommend that.