v0.3.16-cu128-AVX2-win-20251108
·
56 commits
to main
since this release
feat: Update Submodule vendor/llama.cpp 48bd265..299f5d7
feat: Update llama.cpp API and supplementing the State/sessions API
feat: Better Qwen3VL chat template. (Thank to @alcoftTAO)
Note: llama_chat_template now allows for more flexible input of parameters required by the model and the application of more complex Jinja formats.
The initial input parameters for Qwen3VLChatHandler have changed: "use_think_prompt" has been changed to "force_reasoning".