- 
                Notifications
    You must be signed in to change notification settings 
- Fork 155
Port of Qwen3-VL support from mainline #883
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
- convert_hf_to_gguf.py - Not touched, use llama.cpp to convert model instead - sysl and metal support for imrope not added - Vulkan support for imrope not tested - Code not tested
| I tried to build this branch with cpu-only and CUDA but it kept failing: CUDA 
 
 | 
| @ranilongxi, thank you; should be fixed now. | 
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Apart from the comments, LGTM.
But we need to get this tested by someone before merging.
| Thank you @ikawrakow, I'll do my best to resolve your comments. There are still some compilation issues which I'll resolve as well. | 
| @ikawrakow please let me know if you are happy with how I've addressed your comments: | 
Source: ggml-org/llama.cpp#16780