Replies: 1 comment
-
Since llama-server is supposed to provide OpenAI compatible endpoints it makes a lot of sense to support We could add “openai” as an option for —reasoning-format. The operators choose to |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Several clients do not support the
reasoning_content
field, and it seems like both clients and inference servers have converged to areasoning
field.One such example is OpenAI's own compatibility test for GPT-OSS.
llama.cpp fails 30/30 tests simply because it places the reasoning in
reasoning_content
and notreasoning
.The agent-js library from OpenAI now includes this functionality (openai/openai-agents-js@7b437d9). I've verified that they return the reasoning within the message, which meets the
gpt-oss
spec requiring CoT output from the final tool call message.Analysis (with reasoning_content field)
When modified to use
reasoning
, llama.cpp passes 30/30 tests withreasoning_effort = low
.Analysis (with reasoning field)
I'm willing to submit a pull request for this issue, but since it appears to be a feature request, I'm posting it in discussions first to gather feedback.
From what I can recall, agentic coding tools like codex and crush support
reasoning
but lackreasoning_content
support. There are likely other tools that behave similarly.Thoughts?
Beta Was this translation helpful? Give feedback.
All reactions