Skip to content

Commit b8bd443

Browse files
authored
feat(genapi): add billing free tiers example to faq
1 parent c1b227c commit b8bd443

File tree

1 file changed

+37
-1
lines changed

1 file changed

+37
-1
lines changed

pages/generative-apis/faq.mdx

Lines changed: 37 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,43 @@ Our Generative APIs support a range of popular models, including:
2121

2222
## How does the free tier work?
2323
The free tier allows you to process up to 1,000,000 tokens without incurring any costs. After reaching this limit, you will be charged per million tokens processed. Free tier usage is calculated by adding all input and output tokens consumed from all models used.
24-
For more information, refer to our [pricing page](https://www.scaleway.com/en/pricing/model-as-a-service/#generative-apis).
24+
For more information, refer to our [pricing page](https://www.scaleway.com/en/pricing/model-as-a-service/#generative-apis) or access your bills by token types and models in [billing section from Scaleway Console](https://console.scaleway.com/billing/payment) (past and previsional during the current month).
25+
26+
Note that when your consumption exceeds free tier, you will be billed for each additional token consumed by model and token types. The minimum billing unit is 1 million tokens. Here are two examples for low volume consumptions:
27+
28+
Example 1: Free Tier only
29+
30+
| Model | Token type | Tokens consumed | Price | Bill |
31+
|-----------------|-----------------|-----------------|-----------------|-----------------|
32+
| `llama-3.3-70b-instruct` | Input | 500k | 0.90€/million tokens | 0.00€ |
33+
| `llama-3.3-70b-instruct` | Output | 200k | 0.90€/million tokens | 0.00€ |
34+
| `mistral-small-3.1-24b-instruct-2503` | Input | 100k | 0.15€/million tokens | 0.00€ |
35+
| `mistral-small-3.1-24b-instruct-2503` | Output | 100k | 0.35€/million tokens | 0.00€ |
36+
37+
Total tokens consumed: `900k`
38+
Total bill: `0.00€`
39+
40+
Example 2: Exceeding Free Tier
41+
42+
| Model | Token type | Tokens consumed | Price | Billed consumption | Bill |
43+
|-----------------|-----------------|-----------------|-----------------|-----------------|-----------------|
44+
| `llama-3.3-70b-instruct` | Input | 800k | 0.90€/million tokens | 1 million tokens | 0.00€ (Free Tier application)|
45+
| `llama-3.3-70b-instruct` | Output | 2 500k | 0.90€/million tokens | 3 million tokens | 2.70€ |
46+
| `mistral-small-3.1-24b-instruct-2503` | Input | 100k | 0.15€/million tokens | 1 million tokens | 0.15€ |
47+
| `mistral-small-3.1-24b-instruct-2503` | Output | 100k | 0.35€/million tokens | 1 million tokens | 0.35€ |
48+
49+
Total tokens consumed: `900k`
50+
Total billed consumption: `6 million tokens`
51+
Total bill: `3.20€`
52+
53+
Note that in this example, the first line where free tier applies will not display in your current Scaleway bills by model, but will instead be listed under `Generative APIs Free Tier - First 1M tokens for free`.
54+
55+
## What is a token and how are they counted?
56+
A token is the minimum unit of content that is seen and processed by a model. Hence token definitions depends on input types:
57+
- For text, on average, `1` token corresponds to `~4` characters, and thus `0.75` words (as words are on average 5 characters long)
58+
- For images, `1` token corresponds to a square of pixels. For example, [pixtral-12b-2409 model](https://www.scaleway.com/en/docs/managed-inference/reference-content/pixtral-12b-2409/#frequently-asked-questions) image tokens of `16x16` pixels (16 pixel height, and 16 pixel width, hence `256` pixels in total).
59+
60+
Exact tokens count and definition depends on [tokenizers](https://huggingface.co/learn/llm-course/en/chapter2/4) used by each models. When this difference is significant (such as for image processing), you can find detailed information in each model documentation (for instance in [`pixtral-12b-2409` size limit documentation](https://www.scaleway.com/en/docs/managed-inference/reference-content/pixtral-12b-2409/#frequently-asked-questions)). Otherwise, when the model is open, you can find this information in the model files on platforms such as Hugging Face, usually in the `tokenizer_config.json` file.
2561

2662
## How can I monitor my token consumption?
2763
You can see your token consumption in [Scaleway Cockpit](/cockpit/). You can access it from the Scaleway console under the [Metrics tab](https://console.scaleway.com/generative-api/metrics).

0 commit comments

Comments
 (0)