Large SpaCy models is (consistently?) faster than medium? #8611
Replies: 1 comment 2 replies
-
Thanks for reporting this. That said, it is the "large" model, not the "slow" model 😄 As far as I'm aware the architecture of the medium and large models is basically the same, the main difference being the number of word vectors. The main criteria for deciding between the medium and large models should be memory or disk constraints, not speed. Assuming your machine has plenty of power to handle the large model, I wouldn't expect a significant speed difference between medium and large models. It's a little surprising the large one is faster, but as expected the difference is very small. In general you need to be very cautious of artificial benchmarks. One problem with yours, for example, is that you're processing the same sentence repeatedly, which is not a realistic load. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
How to reproduce the behavior
As seen in the code below, we see that the larger SpaCy models perform faster than the medium model.
The code to reproduce is quite simply:
Which prints:
Applying the same approach to English we get.
increasing x to 4 we still see that the large model is faster
increasing x to 4 we still see that the large model is still faster
and with x = 30
Looking into their config.cfg files I didn't find anything to warrant this, am I missing something obvious? Any reason why this might be?
Additional info
I looked into this due to the performance on a server (Ubuntu w. GPU) and the results are similar.
Your Environment / Info about spaCy
Beta Was this translation helpful? Give feedback.
All reactions