This repository was archived by the owner on Jul 22, 2025. It is now read-only.
API very slow / returning 5xx errors #53
Closed
wisnewskirobert
started this conversation in
General
Replies: 2 comments
-
Also here are my parameters: max_tokens: 1000, // Adjust as needed |
Beta Was this translation helpful? Give feedback.
0 replies
-
Hey, I recommend using sonar-large instead; it is much more stable. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
hi there,
I am testing the perplexity API to see if I can use it in a new app.
It is taking around 47 seconds to return a response. Out of my last 6 tests I got 3 5xx errors and 3 200s.
Not sure I will be able to use it with this kind of performance. Maybe its just a 1 off today?
Here are the details of my call:
Company Information: { "id": "e0b0e4d3-1ffb-402e-906a-3d6540216208", "model": "llama-3.1-sonar-huge-128k-online", "created": 1728991890, "usage": { "prompt_tokens": 128, "completion_tokens": 741, "total_tokens": 869 }, "object": "chat.completion",
thanks,
Beta Was this translation helpful? Give feedback.
All reactions