Add Kimi K2 Context Window #8597
Replies: 1 comment
-
Closed by #8604 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
What happened?
UPDATE: The issue is that
Max Context Tokens
setting is being set far too low by default for the Kimi K2 model through OpenRouter. Manually setting it to the model's real max context size of131000
fixed the issue. However, there is probably still a bug here related to how the maximum context size is determined for models through OpenRouter.I've noticed issues attempting to use the Kimi K2 model through OpenRouter (moonshotai/kimi-k2
) where some past details in the conversation appear to be very quickly forgotten by the model.After some investigation I've found this is caused by LibreChat, not OpenRouter or the Kimi K2 model itself. By intercepting requests from LibreChat to OpenRouter usingmitmproxy
, I can see that past messages are sometimes dropped from the conversation history in the request sent to the OpenRouter completions endpoint.I'm unsure of the exact parameters of this bug, but I can consistently reproduce it with the procedure described below. Having longer response in the conversation history appears to be necessary to trigger it. Kimi K2 is the only model I've found to be affected.Let me know if you have difficulty reproducing this and I can investigate further.Version Information
0.7.9-rc1
Steps to Reproduce
moonshotai/kimi-k2
via OpenRouterRemember this fact: the apple is purple. Don't repeat it to me, just say you'll remember it.
)Write a long essay on the Frutiger Aero aesthetic
)What color was the apple?
)mitmproxy
and repeat the request to Kimi K2.For example, see the following intercepted request:
What browsers are you seeing the problem on?
No response
Relevant log output
Screenshots
No response
Code of Conduct
Beta Was this translation helpful? Give feedback.
All reactions