Skip to content
Discussion options

You must be logged in to vote

The base model influences the type of network architecture that is used for training, so it can certainly influence memory usage. A transformer-based model for instance will have a different memory usage than a small model. So if you're using en_core_web_sm as base model, the config will source components from that and the architecture will be less complex.

To verify - can you share the exact commands you used to train with Prodigy, and to export from Prodigy and then train?

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by svlandeg
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
more-info-needed This issue needs more information ✨ prodigy Issues related to using spaCy with the Prodigy annotation tool perf / memory Performance: memory use feat / spancat Feature: Span Categorizer
2 participants