gpt-oss-20b - tool calling issues? #15141
Unanswered
DigitalRudeness
asked this question in
Q&A
Replies: 1 comment
-
Yes, it's currently a bit broken as of right now. Check out the latest work in #15158 + my comment at the end that contains another patch. In my testing, it works fine with those two changes, but more testing would always be welcome :) |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi @ALL,
I'm currently struggling on executing tool calls with the new gpt-oss-20b model.
I'm using a self programmed python solution for creating chat requests and tool calls against llama-server.
The generated tool calls look like the example below (standard, nothing special):
As you can see I'm trying to let the model choose from different functions either with params or without depending on the request.
When I'm e.g. requesting the weather for a city in a region (ignore the system prompt, it will be outsourced from regular requests)....
... I would either expect the model to execute the fetched function directly or at least return a tool_call structure containing the fetched function (and parameters if needed) to be manually executed at parsing the response, like e.g. command-r does. But what I get is just the reasoning content just mentioning what has to be used in a floating text form (output is assembled also by a self coded HTTP class. So please don't be confused about the structure):
This is a bit unhandy to parse and I even think I'm missing sth. to complete the whole request correctly because a resoning content can't be the end of the "chain"....
Would be nice if anybody could tell me what I'm doing wrong.. :-)
THX A LOT IN ADVANCE! :-)
Beta Was this translation helpful? Give feedback.
All reactions