You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: pages/account/how-to/open-a-support-ticket.mdx
+2-3Lines changed: 2 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -52,7 +52,7 @@ Providing a clear subject and description will help us resolve your issue faster
52
52
Example: “The issue occurs when attempting to start an Instance after applying a configuration update in the Scaleway console.”
53
53
54
54
-**Expected behavior:** explain what you expected to happen.
55
-
Example: “The instance should start within 2 minutes without errors.”
55
+
Example: “The Instance should start within 2 minutes without errors.”
56
56
57
57
-**Actual behavior:** describe what is happening instead.
58
58
Example: “The Instance remains in "Starting" status for over 10 minutes and then switches to "Error".
@@ -71,7 +71,6 @@ Examples:
71
71
- Screenshot of the network tab of your browser’s Developer Tools (right-click anywhere on the page and select **Inspect**. Go to the **Network tab** in the Developer Tools panel.)
72
72
- Logs
73
73
74
-
75
74
<Messagetype="important">
76
75
If you have lost access to the Scaleway console and want to create a ticket, you must first [follow this procedure](/account/how-to/use-2fa/#how-to-regain-access-to-your-account) to regain access to your account.
Copy file name to clipboardExpand all lines: pages/audit-trail/quickstart.mdx
+1-4Lines changed: 1 addition & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -39,7 +39,4 @@ Refer to the [dedicated documentation page](/audit-trail/how-to/configure-audit-
39
39
40
40
<Messagetype="tip">
41
41
If no events display after you use the filter, try switching the region from the **Region** drop-down, or adjusting your search. Find out how to troubleshoot event issues in our [dedicated documentation](/audit-trail/troubleshooting/cannot-see-events/).
Copy file name to clipboardExpand all lines: pages/cockpit/how-to/configure-alerts-for-scw-resources.mdx
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -52,5 +52,5 @@ This page shows you how to configure alerts for Scaleway resources in Grafana us
52
52
</Message>
53
53
54
54
<Messagetype="tip">
55
-
Find out how to send Cockpit's alert notifications to Slack using a webkook URL in our [dedicated documentation](/tutorials/configure-slack-alerting/).
55
+
Find out how to send Cockpit's alert notifications to Slack using a webhook URL in our [dedicated documentation](/tutorials/configure-slack-alerting/).
Copy file name to clipboardExpand all lines: pages/environmental-footprint/how-to/track-monthly-footprint.mdx
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -42,7 +42,7 @@ categories:
42
42
For a detailed description of how the water consumption is calculated, refer to the [Water Consumption section](/environmental-footprint/additional-content/environmental-footprint-calculator/#water-consumption) of the Environmental Footprint calculation breakdown documentation page.
43
43
</Message>
44
44
-**5.** The total water consumption and carbon footprint of each of your Projects.
45
-
-**6.** The total water consumption and carbon footprint per geographical location (Region and Availability Zone)
45
+
-**6.** The total water consumption and carbon footprint per geographical location (region and Availability Zone)
46
46
-**7.** The total water consumption and carbon footprint of each of your products.
47
47
48
48
For both the carbon emissions, and the water consumption, the power consumption of your active resources is used in the calculation. The way you use your resources has a direct impact on power consumption. Therefore, results may vary greatly from one month to another.
Copy file name to clipboardExpand all lines: pages/generative-apis/troubleshooting/fixing-common-issues.mdx
+17-17Lines changed: 17 additions & 17 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,15 +16,15 @@ Below are common issues that you may encounter when using Generative APIs, their
16
16
## 400: Bad Request - You exceeded maximum context window for this model
17
17
18
18
### Cause
19
-
- You provided an input exceeding the maximum context window (also known as context length) for the model you are using.
20
-
- You provided a long input and requested a long input (in `max_completion_tokens` field), which added together, exceed the maximum context window of the model you are using.
19
+
- You provided an input exceeding the maximum context window (also known as context length) for the model you are using.
20
+
- You provided a long input and requested a long input (in `max_completion_tokens` field), which added together, exceeds the maximum context window of the model you are using.
21
21
22
22
### Solution
23
-
- Reduce your input size below what is [supported by the model](/generative-apis/reference-content/supported-models/).
23
+
- Reduce your input size below what is [supported by the model](/generative-apis/reference-content/supported-models/).
24
24
- Use a model supporting longer context window values.
25
25
- Use [Managed Inference](/managed-inference/), where the context window can be increased for [several configurations with additional GPU vRAM](/managed-inference/reference-content/supported-models/). For instance, `llama-3.3-70b-instruct` model in `fp8` quantization can be served with:
26
-
-`15k` tokens context window on `H100`instances
27
-
-`128k` tokens context window on `H100-2`instances.
26
+
-`15k` tokens context window on `H100`Instances
27
+
-`128k` tokens context window on `H100-2`Instances
28
28
29
29
## 403: Forbidden - Insufficient permissions to access the resource
30
30
@@ -46,7 +46,7 @@ Below are common issues that you may encounter when using Generative APIs, their
46
46
- You provided a value for `max_completion_tokens` that is too high and not supported by the model you are using.
47
47
48
48
### Solution
49
-
- Remove `max_completion_tokens` field from your request or client library, or reduce its value below what is [supported by the model](https://www.scaleway.com/en/docs/generative-apis/reference-content/supported-models/).
49
+
- Remove `max_completion_tokens` field from your request or client library, or reduce its value below what is [supported by the model](https://www.scaleway.com/en/docs/generative-apis/reference-content/supported-models/).
50
50
- As an example, when using the [init_chat_model from Langchain](https://python.langchain.com/api_reference/_modules/langchain/chat_models/base.html#init_chat_model), you should edit the `max_tokens` value in the following configuration:
@@ -57,16 +57,16 @@ Below are common issues that you may encounter when using Generative APIs, their
57
57
## 416: Range Not Satisfiable - max_completion_tokens is limited for this model
58
58
59
59
### Cause
60
-
- You provided `max_completion_tokens` value too high, that is not supported by the model you are using.
60
+
- You provided `max_completion_tokens` value too high, which is not supported by the model you are using.
61
61
62
62
### Solution
63
-
- Remove the `max_completion_tokens` field from your request or client library, or reduce its value below what is [supported by the model](https://www.scaleway.com/en/docs/generative-apis/reference-content/supported-models/).
63
+
- Remove the `max_completion_tokens` field from your request or client library, or reduce its value below what is [supported by the model](https://www.scaleway.com/en/docs/generative-apis/reference-content/supported-models/).
64
64
- As an example, when using the [init_chat_model from Langchain](https://python.langchain.com/api_reference/_modules/langchain/chat_models/base.html#init_chat_model), you should edit the `max_tokens` value in the following configuration:
- Use a model supporting a higher `max_completion_tokens` value.
69
-
- Use [Managed Inference](/managed-inference/), where these limits on completion tokens do not apply (your completion tokens amount will still be limited by the maximum context window supported by the model).
69
+
- Use [Managed Inference](/managed-inference/), where these limits on completion tokens do not apply (your completion tokens amount will still be limited by the maximum context window supported by the model).
70
70
71
71
## 429: Too Many Requests - You exceeded your current quota of requests/tokens per minute
72
72
@@ -79,15 +79,15 @@ Below are common issues that you may encounter when using Generative APIs, their
79
79
-[Add a payment method](/billing/how-to/add-payment-method/#how-to-add-a-credit-card) and [validate your identity](/account/how-to/verify-identity/) to increase automatically your quotas [based on standard limits](/organizations-and-projects/additional-content/organization-quotas/#generative-apis).
80
80
-[Ask our support](https://console.scaleway.com/support/tickets/create) to raise your quota.
81
81
- Reduce the size of the input or output tokens processed by your API requests.
82
-
- Use [Managed Inference](/managed-inference/), where these quota do not apply (your throughput will be only limited by the amount of Inference Deployment your provision)
82
+
- Use [Managed Inference](/managed-inference/), where these quotas do not apply (your throughput will be only limited by the amount of Inference Deployment your provision)
83
83
84
84
## 429: Too Many Requests - You exceeded your current threshold of concurrent requests
85
85
86
86
### Cause
87
87
- You kept too many API requests opened at the same time (number of HTTP sessions opened in parallel)
88
88
89
89
### Solution
90
-
- Smooth out your API requests rate by limiting the number of API requests you perform at the same time (eg. requests which did not receive a complete response and are still opened) so that you remain below your [organization quotas for Generative APIs](/organizations-and-projects/additional-content/organization-quotas/#generative-apis).
90
+
- Smooth out your API requests rate by limiting the number of API requests you perform at the same time (eg. requests which did not receive a complete response and are still opened) so that you remain below your [Organization quotas for Generative APIs](/organizations-and-projects/additional-content/organization-quotas/#generative-apis).
91
91
- Use [Managed Inference](/managed-inference/), where concurrent request limit do not apply. Note that exceeding the number of concurrent requests your Inference Deployment can handle may impact performance metrics.
92
92
93
93
@@ -162,15 +162,15 @@ Below are common issues that you may encounter when using Generative APIs, their
162
162
- Counter for **Tokens Processed** or **API Requests** should display a correct value (different from 0)
163
163
- Graph across time should be empty
164
164
165
-
## Embeddings vectors cannot be stored in database or used with a third-party library
165
+
## Embeddings vectors cannot be stored in a database or used with a third-party library
166
166
167
167
### Cause
168
168
The embedding model you are using generates vector representations with a fixed dimension number, which is too high for your database or third-party library.
169
169
- For example, the embedding model `bge-multilingual-gemma2` generates vector representations with `3584` dimensions. However, when storing vectors using PostgreSQL `pgvector` extensions, indexes (in `hnsw` or `ivvflat` formats) only support up to `2000` dimensions.
170
170
171
171
### Solution
172
-
- Use a vector store supporting higher dimensions number, such as [Qdrant](https://www.scaleway.com/en/docs/tutorials/deploying-qdrant-vectordb-kubernetes/).
173
-
- Do not use indexes for vectors or disable them from your third-party library. This may limit performance in vector similarity search for significant volumes.
172
+
- Use a vector store supporting higher dimensions numbers, such as [Qdrant](https://www.scaleway.com/en/docs/tutorials/deploying-qdrant-vectordb-kubernetes/).
173
+
- Do not use indexes for vectors or disable them from your third-party library. This may limit performance in vector similarity search for significant volumes.
174
174
- When using [Langchain PGVector method](https://python.langchain.com/docs/integrations/vectorstores/pgvector/), this method does not create an index by default and should not raise errors.
175
175
- When using the [Mastra](https://mastra.ai/) library with `vectorStoreName: "pgvector"`, specify indexConfig type as `flat` to avoid creating any index on vector dimensions.
176
176
```typescript
@@ -180,7 +180,7 @@ The embedding model you are using generates vector representations with a fixed
180
180
indexConfig: {"type":"flat"},
181
181
});
182
182
```
183
-
- Use a model with a lower number of dimensions. Using [Managed Inference](https://console.scaleway.com/inference/deployments), you can deploy for instance the`sentence-t5-xxl` model, which represents vectors with `768` dimensions.
183
+
- Use a model with a lower number of dimensions. Using [Managed Inference](https://console.scaleway.com/inference/deployments), you can deploy for instance the`sentence-t5-xxl` model, which represents vectors with `768` dimensions.
184
184
185
185
## Previous messages are not taken into account by the model
This snippet will output the model response, which is `4`.
222
-
- When exceeding maximum context window, you should receive a `400 - BadRequestError` detailing context length value you exceeded. In this case, you should reduce the size of the content you send to the API.
222
+
- When exceeding the maximum context window, you should receive a `400 - BadRequestError` detailing the context length value you exceeded. In this case, you should reduce the size of the content you send to the API.
223
223
224
224
## Best practices for optimizing model performance
225
225
@@ -234,4 +234,4 @@ This snippet will output the model response, which is `4`.
234
234
### Debugging silent errors
235
235
- For cases where no explicit error is returned:
236
236
- Verify all fields in the API request are correctly named and formatted.
237
-
- Test the request with smaller and simpler inputs to isolate potential issues.
237
+
- Test the request with smaller and simpler inputs to isolate potential issues.
Copy file name to clipboardExpand all lines: pages/gpu/how-to/use-nvidia-mig-technology.mdx
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -15,7 +15,7 @@ categories:
15
15
16
16
<Messagetype="note">
17
17
* Scaleway offers MIG-compatible GPU Instances such as H100 PCIe GPU Instances
18
-
* NVIDIA uses the term *GPU instance* to designate a MIG partition of a GPU (MIG= Multi-Instance GPU)
18
+
* NVIDIA uses the term *GPU instance* to designate an MIG partition of a GPU (MIG= Multi-Instance GPU)
19
19
* To avoid confusion, we will use the term GPU Instance in this document to designate the Scaleway GPU Instance, and *MIG partition* in the context of the MIG feature.
20
20
</Message>
21
21
@@ -151,10 +151,10 @@ Refer to the official documentation for more information about the supported [MI
151
151
*`-cgi 9,19,19,19`: this flag specifies the MIG partition configuration. The numbers following the flag represent the MIG partitions for each of the four MIG device slices. In this case, there are four slices with configurations 9, 19, 19, and 19 compute instances each. These numbers correspond to the profile IDs retrieved previously. Note that you can use either of the following:
152
152
* Profile ID (e.g. 9, 14, 5)
153
153
* Short name of the profile (e.g. `3g.40gb`)
154
-
* Full profile name of the instance (e.g. `MIG 3g.40gb`)
154
+
* Full profile name of the Instance (e.g. `MIG 3g.40gb`)
155
155
*`-C`: this flag automatically creates the corresponding compute instances for the MIG partitions.
156
156
157
-
The command instructs the `nvidia-smi` tool to set up a MIG configuration where the GPU is divided into four slices, each containing different numbers of MIG partition configurations as specified: an MIG 3g.40gb (Profile ID 9) for the first slice, and an MIG 1g.10gb (Profile ID 19) for each of the remaining three slices.
157
+
The command instructs the `nvidia-smi` tool to set up an MIG configuration where the GPU is divided into four slices, each containing different numbers of MIG partition configurations as specified: an MIG 3g.40gb (Profile ID 9) for the first slice, and an MIG 1g.10gb (Profile ID 19) for each of the remaining three slices.
158
158
159
159
<Messagetype="note">
160
160
- Running CUDA workloads on the GPU requires the creation of MIG partitions along with their corresponding compute instances. Just enabling MIG mode on the GPU is not enough to achieve this.
Copy file name to clipboardExpand all lines: pages/gpu/reference-content/gpu-instances-bandwidth-overview.mdx
+1-2Lines changed: 1 addition & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,9 +13,8 @@ categories:
13
13
- compute
14
14
---
15
15
16
-
17
16
Scaleway GPU Instances are designed to deliver **high-performance computing** for AI/ML workloads, rendering, scientific simulations, and visualization tasks.
18
-
This guide provides a detailed overview of their **internet and Block Storage bandwidth capabilities** to help you choose the right instance for your GPU-powered workloads.
17
+
This guide provides a detailed overview of their **internet and Block Storage bandwidth capabilities** to help you choose the right Instance for your GPU-powered workloads.
Copy file name to clipboardExpand all lines: pages/instances/api-cli/using-cloud-init.mdx
+2-4Lines changed: 2 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -30,7 +30,7 @@ Cloud-config files are special scripts designed to be run by the cloud-init proc
30
30
31
31
You can give provisioning instructions to cloud-init using the `cloud-init` key of the `user_data` facility.
32
32
33
-
For `user_data` to be effective, it has to be added prior to the creation of the instance since `cloud-init` gets activated early in the first phases of the boot process.
33
+
For `user_data` to be effective, it has to be added prior to the creation of the Instance since `cloud-init` gets activated early in the first phases of the boot process.
34
34
35
35
***Server ID** refers to the unique identification string of your server. It will be displayed when you create your server. You can also recover it from the list of your servers, by typing `scw instance server list`.
36
36
@@ -88,6 +88,4 @@ Subcommands:
88
88
89
89
````
90
90
91
-
For detailed information on cloud-init, refer to the official cloud-init [documentation](http://cloudinit.readthedocs.io/en/latest/index.html).
92
-
93
-
91
+
For detailed information on cloud-init, refer to the official cloud-init [documentation](http://cloudinit.readthedocs.io/en/latest/index.html).
0 commit comments