You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
fix: ensure user's max tokens setting overrides model defaults across all providers
- Updated BaseOpenAiCompatibleProvider to use getModelMaxOutputTokens()
- Fixed ChutesHandler to respect user's custom max tokens
- Fixed LiteLLM createMessage and completePrompt methods
- Fixed Glama createMessage and completePrompt methods
- Fixed Unbound createMessage and completePrompt methods
- Fixed Mistral getModel method to use getModelMaxOutputTokens()
- Fixed XAI to use getModelMaxOutputTokens()
- Fixed OpenAI addMaxTokensIfNeeded to use getModelMaxOutputTokens()
- Fixed Gemini to use maxTokens from getModel() which already applies user settings
This ensures that when users set a custom max output tokens value in their provider settings, it will be respected across all providers (capped to the model's actual maximum).
0 commit comments