Ollama下运行Qwen3如何控制思维链是否生成? #1556
Unanswered
BoogonClothman
asked this question in
Q&A
Replies: 1 comment 2 replies
-
|
Ollama官方博客(https://ollama.com/blog/thinking)有给API例子,试试看,是它们自己的格式。 如果使用的是OpenAI API,得换个框架,比如llama.cpp里的llama-server,使用方式见相关PR:ggml-org/llama.cpp#13196。 |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment

Uh oh!
There was an error while loading. Please reload this page.
-
目前因为硬件限制,暂时使用Qwen3:4B,由于我需要完成的任务对实时性要求很高,我打算取消思维链让模型直接输出结果,但是无论是在命令行还是在Python 里调API,加上后缀/no_think都没有让模型的思维链关闭,仍然继续进行thinking,这是什么问题?
ollama版本也更新到0.9了,通过/set指令是正常关闭的,我想知道在API调用的情况下,该怎么关闭思维链
Beta Was this translation helpful? Give feedback.
All reactions