Skip to content

Commit 9aae0b9

Browse files
authored
Update fixing-common-issues.mdx
Add details to on 429: Too many request error
1 parent f94215e commit 9aae0b9

File tree

1 file changed

+16
-3
lines changed

1 file changed

+16
-3
lines changed

ai-data/generative-apis/troubleshooting/fixing-common-issues.mdx

Lines changed: 16 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13,11 +13,24 @@ dates:
1313

1414
Below are common issues that you may encounter when using Generative APIs, their causes, and recommended solutions.
1515

16-
## 504: Timeout
16+
## 429: Too Many Request - You exceeded your current quota of requests/tokens per minute
1717

1818
### Cause
19-
- The query is too long.
20-
- The model goes into an infinite loop while processing the input.
19+
- You performed too many API requests over a given minute
20+
- You consumed too much tokens (input and output) with your API requests over a given minute
21+
22+
### Solution
23+
- [Ask our support](https://console.scaleway.com/support/tickets/create) to raise your quota
24+
- Smooth out your API requests rate by limiting the number of API requests you perform in parallel
25+
- Reduce the size of the input or output tokens processed by your API requests
26+
- Use [Managed Inference](/ai-data/managed-inference/), where these quota do not apply (your throughput will be only limited by the amount of Inference Deployment your provision)
27+
28+
29+
## 504: Gateway Timeout
30+
31+
### Cause
32+
- The query is too long to process (even if context-length stays [between supported context window and maximum tokens](https://www.scaleway.com/en/docs/ai-data/generative-apis/reference-content/supported-models/))
33+
- The model goes into an infinite loop while processing the input (which is a known structural issue with several AI models)
2134

2235
### Solution
2336
- Set a stricter **maximum token limit** to prevent overly long responses.

0 commit comments

Comments
 (0)