-
Notifications
You must be signed in to change notification settings - Fork 13.4k
model : add text-only support for Kimi-VL #15051
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
I don't think this is the correct fix, Have you tried |
This gguf was made with just the "basic" commit without the fix, you can see that
|
And now that I'm looking at the "fixed" gguf again, this one still has the comma as the BOS print_info: vocab type = BPE |
Ok I understand where the problem is coming from now @CISC This is the kimi-vl config.json structure, it puts the x_token_id in the "text_config":
in gguf-py\gguf\vocab.py, it looks only at the config["x_token_id"], not config["text_config"]["x_token_id"] :
So I added some extra logic to look at the text_config. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice catch!
… text_config) (ggml-org#15051) * basic kimi-vl textmodel conversion * check config["text_config"] for special tokens
The text model portion of moonshotai/Kimi-VL-A3B-Instruct is functionally identical to moonshotai/Moonlight-16B-A3B-Instruct, but there is an error in the model's config files. The Kimi-VL models should be using token "<|im_end|>" as their EOS token, not "[EOS]". Without this fix, generation was stopping after any comma "," and I'm not really sure why.
Just wanted to get this merged before I really start working on getting the vision portion working.