Skip to content

Commit 134f6b2

Browse files
bene2k1RoRoJ
andauthored
Update changelog/june2025/2025-06-25-generative-apis-changed-llama-33-70b-maximum-context-up.mdx
Co-authored-by: Rowena Jones <[email protected]>
1 parent c7cf3ca commit 134f6b2

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

changelog/june2025/2025-06-25-generative-apis-changed-llama-33-70b-maximum-context-up.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,5 +8,5 @@ product: generative-apis
88

99
Llama 3.3 70B maximum context is [now reduced to 100k tokens](https://www.scaleway.com/en/docs/generative-apis/reference-content/supported-models/) (from 130k tokens previously).
1010
This update will improve average throughput and time to first token.
11-
[Managed Inference](https://www.scaleway.com/en/docs/managed-inference/reference-content/model-catalog/) can still be used to support 130k tokens context length.
11+
[Managed Inference](https://www.scaleway.com/en/docs/managed-inference/reference-content/model-catalog/) can still be used to support context lengths of 130k tokens.
1212

0 commit comments

Comments
 (0)