Skip to content

Commit abaede6

Browse files
authored
feat(genapi): update Continue suggested models (#5645)
* feat(genapi): update Continue suggested models * feat(genapi): update IntelliJ Continue suggested models * feat(genapi): update Zed integration with Qwen3 Coder * feat(genapi): update troubleshooting Remove duplicated content, and add links to IDEs configuration for maximum context window reached.
1 parent fd5432a commit abaede6

File tree

4 files changed

+28
-41
lines changed

4 files changed

+28
-41
lines changed

pages/generative-apis/reference-content/adding-ai-to-intellij-using-continue.mdx

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -56,21 +56,21 @@ To link Continue with Scaleway’s Generative APIs, you need to configure the se
5656
- If you have already configured a **Local Assistant**, click **Local Assistant** then click the **wheel icon** to open your existing `config.yaml`
5757
- Otherwise, create a `config.yaml` file inside your `.continue` directory.
5858
- Add the following configuration to enable Scaleway's Generative API. This configuration uses three different models for each task:
59-
- `devstral-small-2505` for agentic workflows through a chat interface
60-
- `qwen2.5-coder-32b-instruct` for autocompletion when editing a file
59+
- `qwen3-235b-a22b-instruct-2507` for agentic workflows through a chat interface (`qwen3-coder-30b-a3b-instruct` can also be used)
60+
- `qwen3-coder-30b-a3b-instruct` for autocompletion when editing a file
6161
- `bge-multilingual-gemma2` for embedding and retrieving code context
6262
```yaml
6363
name: Continue Config
6464
version: 0.0.1
6565
models:
66-
- name: Devstral - Scaleway
66+
- name: Qwen3 235B - Scaleway
6767
provider: openai
68-
model: devstral-small-2505
68+
model: qwen3-235b-a22b-instruct-2507
6969
apiBase: https://api.scaleway.ai/v1/
7070
apiKey: ###SCW_SECRET_KEY###
7171
defaultCompletionOptions:
72-
maxTokens: 8000
73-
contextLength: 50000
72+
maxTokens: 4000
73+
contextLength: 40000
7474
roles:
7575
- chat
7676
- apply
@@ -80,12 +80,12 @@ To link Continue with Scaleway’s Generative APIs, you need to configure the se
8080
- tool_use
8181
- name: Autocomplete - Scaleway
8282
provider: openai
83-
model: qwen2.5-coder-32b-instruct
83+
model: qwen3-coder-30b-a3b-instruct
8484
apiBase: https://api.scaleway.ai/v1/
8585
apiKey: ###SCW_SECRET_KEY###
8686
defaultCompletionOptions:
8787
maxTokens: 8000
88-
contextLength: 50000
88+
contextLength: 100000
8989
roles:
9090
- autocomplete
9191
- name: Embeddings Model - Scaleway
@@ -127,8 +127,8 @@ Alternatively, a `config.json` file can be used with the following format. Note
127127
{
128128
"models": [
129129
{
130-
"model": "devstral-small-2505",
131-
"title": "Devstral - Scaleway",
130+
"model": "qwen3-235b-a22b-instruct-2507",
131+
"title": "Qwen3 235B - Scaleway",
132132
"provider": "openai",
133133
"apiKey": "###SCW_SECRET_KEY###"
134134
}
@@ -139,7 +139,7 @@ Alternatively, a `config.json` file can be used with the following format. Note
139139
"apiKey": "###SCW_SECRET_KEY###"
140140
},
141141
"tabAutocompleteModel": {
142-
"model": "qwen2.5-coder-32b-instruct",
142+
"model": "qwen3-coder-30b-a3b-instruct",
143143
"title": "Autocomplete - Scaleway",
144144
"provider": "openai",
145145
"apiKey": "###SCW_SECRET_KEY###"

pages/generative-apis/reference-content/adding-ai-to-vscode-using-continue.mdx

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -56,21 +56,21 @@ To link Continue with Scaleway's Generative APIs, you can configure a settings f
5656
- If you have already configured a **Local Assistant**, click **Local Assistant**, then click the **wheel icon** to open your existing `config.yaml`
5757
- Otherwise, create a `config.yaml` file inside your `.continue` directory.
5858
- Add the following configuration to enable Scaleway's Generative API. This configuration uses three different models for each tasks:
59-
- `devstral-small-2505` for agentic workflows through a chat interface
60-
- `qwen2.5-coder-32b-instruct` for autocompletion when editing a file
59+
- `qwen3-235b-a22b-instruct-2507` for agentic workflows through a chat interface (`qwen3-coder-30b-a3b-instruct` can also be used)
60+
- `qwen3-coder-30b-a3b-instruct` for autocompletion when editing a file
6161
- `bge-multilingual-gemma2` for embedding and retrieving code context
6262
```yaml
6363
name: Continue Config
6464
version: 0.0.1
6565
models:
66-
- name: Devstral - Scaleway
66+
- name: Qwen3 235B - Scaleway
6767
provider: openai
68-
model: devstral-small-2505
68+
model: qwen3-235b-a22b-instruct-2507
6969
apiBase: https://api.scaleway.ai/v1/
7070
apiKey: ###SCW_SECRET_KEY###
7171
defaultCompletionOptions:
72-
maxTokens: 8000
73-
contextLength: 50000
72+
maxTokens: 4000
73+
contextLength: 40000
7474
roles:
7575
- chat
7676
- apply
@@ -80,12 +80,12 @@ To link Continue with Scaleway's Generative APIs, you can configure a settings f
8080
- tool_use
8181
- name: Autocomplete - Scaleway
8282
provider: openai
83-
model: qwen2.5-coder-32b-instruct
83+
model: qwen3-coder-30b-a3b-instruct
8484
apiBase: https://api.scaleway.ai/v1/
8585
apiKey: ###SCW_SECRET_KEY###
8686
defaultCompletionOptions:
8787
maxTokens: 8000
88-
contextLength: 50000
88+
contextLength: 100000
8989
roles:
9090
- autocomplete
9191
- name: Embeddings Model - Scaleway
@@ -127,8 +127,8 @@ Alternatively, a `config.json` file can be used with the following format. Note
127127
{
128128
"models": [
129129
{
130-
"model": "devstral-small-2505",
131-
"title": "Devstral - Scaleway",
130+
"model": "qwen3-235b-a22b-instruct-2507",
131+
"title": "Qwen3 235B - Scaleway",
132132
"provider": "openai",
133133
"apiKey": "###SCW_SECRET_KEY###"
134134
}
@@ -139,7 +139,7 @@ Alternatively, a `config.json` file can be used with the following format. Note
139139
"apiKey": "###SCW_SECRET_KEY###"
140140
},
141141
"tabAutocompleteModel": {
142-
"model": "qwen2.5-coder-32b-instruct",
142+
"model": "qwen3-coder-30b-a3b-instruct",
143143
"title": "Autocomplete - Scaleway",
144144
"provider": "openai",
145145
"apiKey": "###SCW_SECRET_KEY###"

pages/generative-apis/reference-content/adding-ai-to-zed-ide.mdx

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -28,8 +28,8 @@ Zed is an IDE (Integrated Development Environment) including AI coding assistanc
2828
"api_url": "https://api.scaleway.ai/v1",
2929
"available_models": [
3030
{
31-
"name": "devstral-small-2505",
32-
"display_name": "Devstral Small - Scaleway",
31+
"name": "qwen3-coder-30b-a3b-instruct",
32+
"display_name": "Qwen3 Coder - Scaleway",
3333
"max_tokens": 128000
3434
}
3535
],
@@ -40,13 +40,13 @@ Zed is an IDE (Integrated Development Environment) including AI coding assistanc
4040
"default_profile": "write",
4141
"default_model": {
4242
"provider": "openai",
43-
"model": "devstral-small-2505"
43+
"model": "qwen3-coder-30b-a3b-instruct"
4444
}
4545
}
4646
}
4747
```
4848

49-
This configuration will add a `devstral-small-2505` Scaleway hosted model available with the Zed `openai` provider, and use it as default model.
49+
This configuration will add a `qwen3-coder-30b-a3b-instruct` Scaleway hosted model available with the Zed `openai` provider, and use it as default model.
5050

5151
2. Open AI Assistant configuration by either using the command palette and typing `assistant: show configuration` or clicking on the bottom right **Assistant Panel** button and then **Assistant menu** in top right and finally **Configure**.
5252

@@ -55,5 +55,5 @@ Zed is an IDE (Integrated Development Environment) including AI coding assistanc
5555
Your key will be deleted if you restart Zed. To store it permanently, you can set up your Scaleway secret key as `OPENAI_API_KEY` environment variable and restart Zed.
5656
</Message>
5757

58-
4. Your setup is complete. If you open a new chat and select the `Devstral Small - Scaleway` model, you can send text and see the model using local tools to perform actions such as file reading or editing. You can review changes before they are applied and also add custom remote tools using [Custom MCP Servers](https://zed.dev/docs/ai/mcp).
58+
4. Your setup is complete. If you open a new chat and select the `Qwen3 Coder - Scaleway` model, you can send text and see the model using local tools to perform actions such as file reading or editing. You can review changes before they are applied and also add custom remote tools using [Custom MCP Servers](https://zed.dev/docs/ai/mcp).
5959
Additionally, you can use the [Inline Assist](https://zed.dev/docs/ai/inline-assistant) feature when editing your code.

pages/generative-apis/troubleshooting/fixing-common-issues.mdx

Lines changed: 1 addition & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,7 @@ Below are common issues that you may encounter when using Generative APIs, their
1717

1818
### Solution
1919
- Reduce your input size below what is [supported by the model](/generative-apis/reference-content/supported-models/).
20+
- If you are using a third party tool such as IDEs, you should edit their configuration to set an appropriate maximum context window for the model. More information for [VS Code (Continue)](/generative-apis/reference-content/adding-ai-to-vscode-using-continue/#configure-continue-through-a-configuration-file), [IntelliJ (Continue)](/generative-apis/reference-content/adding-ai-to-intellij-using-continue/#configure-continue-through-configuration-file) and [Zed](/generative-apis/reference-content/adding-ai-to-zed-ide/).
2021
- Use a model supporting longer context window values.
2122
- Use [Managed Inference](/managed-inference/), where the context window can be increased for [several configurations with additional GPU vRAM](/managed-inference/reference-content/supported-models/). For instance, `llama-3.3-70b-instruct` model in `fp8` quantization can be served with:
2223
- `15k` tokens context window on `H100` Instances
@@ -51,20 +52,6 @@ Below are common issues that you may encounter when using Generative APIs, their
5152

5253
## 416: Range Not Satisfiable - max_completion_tokens is limited for this model
5354

54-
### Cause
55-
- You provided a value for `max_completion_tokens` that is too high and not supported by the model you are using.
56-
57-
### Solution
58-
- Remove `max_completion_tokens` field from your request or client library, or reduce its value below what is [supported by the model](https://www.scaleway.com/en/docs/generative-apis/reference-content/supported-models/).
59-
- As an example, when using the [init_chat_model from Langchain](https://python.langchain.com/api_reference/_modules/langchain/chat_models/base.html#init_chat_model), you should edit the `max_tokens` value in the following configuration:
60-
```python
61-
llm = init_chat_model("llama-3.3-70b-instruct", max_tokens="8000", model_provider="openai", base_url="https://api.scaleway.ai/v1", temperature=0.7)
62-
```
63-
- Use a model supporting higher `max_completion_tokens` value.
64-
- Use [Managed Inference](/managed-inference/), where these limits on completion tokens do not apply (your completion tokens amount will still be limited by the maximum context window supported by the model).
65-
66-
## 416: Range Not Satisfiable - max_completion_tokens is limited for this model
67-
6855
### Cause
6956
- You provided a value for `max_completion_tokens` which is too high, and not supported by the model you are using.
7057

0 commit comments

Comments
 (0)