We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4fe5657 commit dc98cd1Copy full SHA for dc98cd1
pages/managed-inference/reference-content/llama-3.1-8b-instruct.mdx
@@ -34,9 +34,9 @@ meta/llama-3.1-8b-instruct:bf16
34
| Instance type | Max context length |
35
| ------------- |-------------|
36
| L4 | 96k (FP8), 27k (BF16) |
37
-| L40S | 96k (FP8), 27k (BF16) |
38
-| H100 | 128k (FP8, BF16)
39
-| H100-2 | 128k (FP8, BF16)
+| L40S | 128k (FP8, BF16) |
+| H100 | 128k (FP8, BF16) |
+| H100-2 | 128k (FP8, BF16) |
40
41
## Model introduction
42
0 commit comments