Skip to content
Discussion options

You must be logged in to vote

Have you tried passing a batch_size to nlp.pipe?

Not sure what environment you're developing in, but note that Jupyter kernels often set a memory limit lower than the system memory, so you might want to look into adjusting that.

upper hundreds of documents, total ~1.5M words.

Does that mean each document is around 10,000 words? That's pretty long - you might find it easier to work with documents if you slice them into paragraphs or other sub-units.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@Uzay-G
Comment options

Answer selected by Uzay-G
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
perf / memory Performance: memory use
2 participants