Config setting for request timeout and retries in custom endpoints #5567
jameslamine
started this conversation in
Feature Requests & Suggestions
Replies: 1 comment
-
Implemented here: #5568 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Summary
Add two new settings to custom endpoints:
Under the hood these will set
maxRetries
andtimeout
in the Open AI SDK.Example
Details
Sometimes the downstream LLM API hangs and we have to wait for the API call to time-out before seeing an error. Users can manually cancel the request, but generally when a request is hanging it's never going to respond. Better to cancel the request and re-try.
In general, it's best practice to set relatively low timeouts, for example 5 seconds. When using the streaming API you can use relatively low timeouts, but in-general this should be user configurable.
The openai nodejs SDK supports a configurable timeout parameter: https://github.com/openai/openai-node?tab=readme-ov-file#timeouts
Let's add support for that as a yaml config option in
endpoints
Beta Was this translation helpful? Give feedback.
All reactions