Would it be faster to process entire corpus? #7326
the1gofer
started this conversation in
Help: Best practices
Replies: 1 comment 6 replies
-
If you're running into speed issues with lots of documents we usually recommend using nlp.pipe. From the docs:
How large are your individual documents? |
Beta Was this translation helpful? Give feedback.
6 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I have about 1600 documents that I'm wanting to process with Spacy, but it takes ages to generate the doc for that many strings. Would it be faster to combine them all into one large text then slice them up somehow down the road?
Beta Was this translation helpful? Give feedback.
All reactions