Replies: 4 comments 3 replies
-
Not at the moment but I'm actually working on this, this week |
Beta Was this translation helpful? Give feedback.
-
@danny-avila Any news on this front? A way to disable all local context window management and let the remote chat API do it would be great, especially now that we're getting longer and longer contexts and fitting limited windows is not so much of an issue anymore. |
Beta Was this translation helpful? Give feedback.
-
I submitted a PR for together.ai models in #2458. Perhaps something like that would work for groq as well. There doesn't seem to be a naming convention across providers but maintaining a list doesn't seem to be too onerous. |
Beta Was this translation helpful? Give feedback.
-
Is there a way to set this from the custom model in librechat.yaml? I can work around this issue by increasing the max context length in the prompt settings as a user, but I'd like the default setting to not be so low. It can be confusing for my end-users. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I'm using Docker and already setup librechat.yaml to access my home endpoint running miqu 70b 32k ctx. It is works but very limiting max tokens. It also wrong max tokens on Groq mixtral 32k, only max at 8k.
How to make it accepting 32k tokens? Is there any settings to change that limiting 4095 max tokens? Thank's.
Update:
Found my temporary fix. This error throwing in
LibreChat/api/app/clients/BaseClient.js
and I manually fixed this with:I hope there's official fix for this.
Beta Was this translation helpful? Give feedback.
All reactions