This repository was archived by the owner on Jul 22, 2025. It is now read-only.
Request for small sonar model #118
Unanswered
liam-twist
asked this question in
Q&A
Replies: 2 comments
-
I also ask for the mini model when I test the answers in the current Sonar, for my language (Polish) the better answers were in the older model than in Sonar and it actually works much faster, when I have a csv file with 150 queries the earlier model worked faster with the current one it takes a little too long to do so |
Beta Was this translation helpful? Give feedback.
0 replies
-
+1 for this feature request, real need for a 7/8B parameter model for low latency responses, or the existing sonar model inference significantly increased. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
We're currently using the
llama-3.1-sonar-small-128k-online
model, which is set to be deprecated on 2025-02-22. We're moving over to the newsonar
model but noticing it is significantly slower than the legacy small model - it seems to have similar speed to the legacy large model.For our use case speed it rather critical... are there any plans to release a smaller, faster variation of the
sonar
model?Beta Was this translation helpful? Give feedback.
All reactions