[Bug]: Agents not passing max_tokens to LiteLLM endpoint #6884
Replies: 2 comments
-
Thanks, I confirmed this is happening and will have a fix soon. Due to a discrepancy in how the fields are interpreted in the SDK we are using. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Closed by #6886 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
What happened?
I have an agent configured to use a model from my LiteLLM endpoint.
When setting the max_tokens in the agent config, this property is not included in the request to LiteLLM.
If I use OpenAI or Anthropic endpoints directly, max_tokens is correctly sent.
Version Information
ghcr.io/danny-avila/librechat-dev:latest@sha256:71f4135a646e0318a5f0b7079a4d253ba2673c64bac67caef883c4035e2f9888
Steps to Reproduce
Create a LiteLLM endpoint:
Create an agent that uses this endpoint.
Config from DB, with a few identifying fields removed:
Use the agent, note that it does not cap the output at 512 tokens.
View request details in LiteLLM, note that temperature is included correctly but max_tokens is not:
What browsers are you seeing the problem on?
No response
Relevant log output
Screenshots
Code of Conduct
Beta Was this translation helpful? Give feedback.
All reactions