Skip to content

Commit 038be81

Browse files
ofrancChangelog botbene2k1RoRoJ
authored
feat(changelog): generative-apis-changed-llama-33-70b-maximum-context-up 2025-06-25 (#5184)
* feat(changelog): add new entry * Update changelog/june2025/2025-06-25-generative-apis-changed-llama-33-70b-maximum-context-up.mdx Co-authored-by: Rowena Jones <[email protected]> --------- Co-authored-by: Changelog bot <[email protected]> Co-authored-by: Benedikt Rollik <[email protected]> Co-authored-by: Rowena Jones <[email protected]>
1 parent 0e2fb6b commit 038be81

File tree

1 file changed

+12
-0
lines changed

1 file changed

+12
-0
lines changed
Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,12 @@
1+
---
2+
title: Llama 3.3 70B maximum context update
3+
status: changed
4+
date: 2025-06-25
5+
category: ai-data
6+
product: generative-apis
7+
---
8+
9+
Llama 3.3 70B maximum context is [now reduced to 100k tokens](https://www.scaleway.com/en/docs/generative-apis/reference-content/supported-models/) (from 130k tokens previously).
10+
This update will improve average throughput and time to first token.
11+
[Managed Inference](https://www.scaleway.com/en/docs/managed-inference/reference-content/model-catalog/) can still be used to support context lengths of 130k tokens.
12+

0 commit comments

Comments
 (0)