Skip to content

Commit 2de6550

Browse files
authored
Change default InferenceClient model to Qwen/Qwen3-Next-80B-A3B-Thinking (#1813)
1 parent 246e873 commit 2de6550

File tree

22 files changed

+39
-39
lines changed

22 files changed

+39
-39
lines changed

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -137,7 +137,7 @@ model = OpenAIModel(
137137
from smolagents import TransformersModel
138138

139139
model = TransformersModel(
140-
model_id="Qwen/Qwen3-4B-Instruct-2507",
140+
model_id="Qwen/Qwen3-Next-80B-A3B-Thinking",
141141
max_new_tokens=4096,
142142
device_map="auto"
143143
)
@@ -178,7 +178,7 @@ You can run agents from CLI using two commands: `smolagent` and `webagent`.
178178
`smolagent` is a generalist command to run a multi-step `CodeAgent` that can be equipped with various tools.
179179

180180
```bash
181-
smolagent "Plan a trip to Tokyo, Kyoto and Osaka between Mar 28 and Apr 7." --model-type "InferenceClientModel" --model-id "Qwen/Qwen3-Next-80B-A3B-Instruct" --imports pandas numpy --tools web_search
181+
smolagent "Plan a trip to Tokyo, Kyoto and Osaka between Mar 28 and Apr 7." --model-type "InferenceClientModel" --model-id "Qwen/Qwen3-Next-80B-A3B-Thinking" --imports pandas numpy --tools web_search
182182
```
183183

184184
Meanwhile `webagent` is a specific web-browsing agent using [helium](https://github.com/mherrmann/helium) (read more [here](https://github.com/huggingface/smolagents/blob/main/src/smolagents/vision_web_browser.py)).

docs/source/en/examples/async_agent.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ from starlette.routing import Route
3939
from smolagents import CodeAgent, InferenceClientModel
4040

4141
agent = CodeAgent(
42-
model=InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct"),
42+
model=InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking"),
4343
tools=[],
4444
)
4545

docs/source/en/examples/multiagents.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,13 +36,13 @@ from huggingface_hub import login
3636
login()
3737
```
3838

39-
⚡️ Our agent will be powered by [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) using `InferenceClientModel` class that uses HF's Inference API: the Inference API allows to quickly and easily run any OS model.
39+
⚡️ Our agent will be powered by [Qwen/Qwen3-Next-80B-A3B-Thinking](https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking) using `InferenceClientModel` class that uses HF's Inference API: the Inference API allows to quickly and easily run any OS model.
4040

4141
> [!TIP]
4242
> Inference Providers give access to hundreds of models, powered by serverless inference partners. A list of supported providers can be found [here](https://huggingface.co/docs/inference-providers/index).
4343
4444
```py
45-
model_id = "Qwen/Qwen2.5-Coder-32B-Instruct"
45+
model_id = "Qwen/Qwen3-Next-80B-A3B-Thinking"
4646
```
4747

4848
## 🔍 Create a web search tool

docs/source/en/examples/rag.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -156,7 +156,7 @@ from smolagents import InferenceClientModel, CodeAgent
156156
# Initialize the agent with our retriever tool
157157
agent = CodeAgent(
158158
tools=[retriever_tool], # List of tools available to the agent
159-
model=InferenceClientModel(), # Default model "Qwen/Qwen2.5-Coder-32B-Instruct"
159+
model=InferenceClientModel(), # Default model "Qwen/Qwen3-Next-80B-A3B-Thinking"
160160
max_steps=4, # Limit the number of reasoning steps
161161
verbosity_level=2, # Show detailed agent reasoning
162162
)

docs/source/en/examples/text_to_sql.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -175,14 +175,14 @@ for table in ["receipts", "waiters"]:
175175

176176
print(updated_description)
177177
```
178-
Since this request is a bit harder than the previous one, we’ll switch the LLM engine to use the more powerful [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct)!
178+
Since this request is a bit harder than the previous one, we’ll switch the LLM engine to use the more powerful [Qwen/Qwen3-Next-80B-A3B-Thinking](https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking)!
179179

180180
```py
181181
sql_engine.description = updated_description
182182

183183
agent = CodeAgent(
184184
tools=[sql_engine],
185-
model=InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct"),
185+
model=InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking"),
186186
)
187187

188188
agent.run("Which waiter got more total money from tips?")

docs/source/en/tutorials/tools.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -275,7 +275,7 @@ Then you can use this tool just like any other tool. For example, let's improve
275275
```python
276276
from smolagents import CodeAgent, InferenceClientModel
277277

278-
model = InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct")
278+
model = InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking")
279279
agent = CodeAgent(tools=[image_generation_tool], model=model)
280280

281281
agent.run(
@@ -323,7 +323,7 @@ Let's add the `model_download_tool` to an existing agent initialized with only t
323323
```python
324324
from smolagents import InferenceClientModel
325325

326-
model = InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct")
326+
model = InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking")
327327

328328
agent = CodeAgent(tools=[], model=model, add_base_tools=True)
329329
agent.tools[model_download_tool.name] = model_download_tool

docs/source/hi/examples/multiagents.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,12 +39,12 @@ from huggingface_hub import login
3939
login()
4040
```
4141

42-
⚡️ हमारा एजेंट [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) द्वारा संचालित होगा जो `InferenceClientModel` क्लास का उपयोग करता है जो HF के Inference API का उपयोग करता है: Inference API किसी भी OS मॉडल को जल्दी और आसानी से चलाने की अनुमति देता है।
42+
⚡️ हमारा एजेंट [Qwen/Qwen3-Next-80B-A3B-Thinking](https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking) द्वारा संचालित होगा जो `InferenceClientModel` क्लास का उपयोग करता है जो HF के Inference API का उपयोग करता है: Inference API किसी भी OS मॉडल को जल्दी और आसानी से चलाने की अनुमति देता है।
4343

4444
_नोट:_ The Inference API विभिन्न मानदंडों के आधार पर मॉडल होस्ट करता है, और डिप्लॉय किए गए मॉडल बिना पूर्व सूचना के अपडेट या बदले जा सकते हैं। इसके बारे में अधिक जानें [यहां](https://huggingface.co/docs/api-inference/supported-models)
4545

4646
```py
47-
model_id = "Qwen/Qwen2.5-Coder-32B-Instruct"
47+
model_id = "Qwen/Qwen3-Next-80B-A3B-Thinking"
4848
```
4949

5050
## 🔍 एक वेब सर्च टूल बनाएं

docs/source/hi/examples/text_to_sql.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -166,14 +166,14 @@ for table in ["receipts", "waiters"]:
166166

167167
print(updated_description)
168168
```
169-
चूंकि यह रिक्वेस्ट पिछले वाले से थोड़ी कठिन है, हम LLM इंजन को अधिक शक्तिशाली [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) का उपयोग करने के लिए स्विच करेंगे!
169+
चूंकि यह रिक्वेस्ट पिछले वाले से थोड़ी कठिन है, हम LLM इंजन को अधिक शक्तिशाली [Qwen/Qwen3-Next-80B-A3B-Thinking](https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Thinking) का उपयोग करने के लिए स्विच करेंगे!
170170

171171
```py
172172
sql_engine.description = updated_description
173173

174174
agent = CodeAgent(
175175
tools=[sql_engine],
176-
model=InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct"),
176+
model=InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking"),
177177
)
178178

179179
agent.run("Which waiter got more total money from tips?")

docs/source/hi/tutorials/tools.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -121,7 +121,7 @@ image_generation_tool("A sunny beach")
121121
```python
122122
from smolagents import CodeAgent, InferenceClientModel
123123

124-
model = InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct")
124+
model = InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking")
125125
agent = CodeAgent(tools=[image_generation_tool], model=model)
126126

127127
agent.run(
@@ -169,7 +169,7 @@ agent.run("How many more blocks (also denoted as layers) are in BERT base encode
169169
```python
170170
from smolagents import InferenceClientModel
171171

172-
model = InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct")
172+
model = InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking")
173173

174174
agent = CodeAgent(tools=[], model=model, add_base_tools=True)
175175
agent.tools[model_download_tool.name] = model_download_tool

docs/source/ko/examples/async_agent.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ from starlette.routing import Route
3939
from smolagents import CodeAgent, InferenceClientModel
4040

4141
agent = CodeAgent(
42-
model=InferenceClientModel(model_id="Qwen/Qwen2.5-Coder-32B-Instruct"),
42+
model=InferenceClientModel(model_id="Qwen/Qwen3-Next-80B-A3B-Thinking"),
4343
tools=[],
4444
)
4545

0 commit comments

Comments
 (0)