Performance with nlp.pipe
#9598
-
Hello, I've been using However, it seems now that when I process a large batch with When I process the documents individually with Is there a way I could fix this? Maybe instead of calling |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Have you tried passing a Not sure what environment you're developing in, but note that Jupyter kernels often set a memory limit lower than the system memory, so you might want to look into adjusting that.
Does that mean each document is around 10,000 words? That's pretty long - you might find it easier to work with documents if you slice them into paragraphs or other sub-units. |
Beta Was this translation helpful? Give feedback.
Have you tried passing a
batch_size
tonlp.pipe
?Not sure what environment you're developing in, but note that Jupyter kernels often set a memory limit lower than the system memory, so you might want to look into adjusting that.
Does that mean each document is around 10,000 words? That's pretty long - you might find it easier to work with documents if you slice them into paragraphs or other sub-units.