@@ -18,7 +18,7 @@ A quick overview of available models in Scaleway's catalog and their core attrib
1818| ------------| ----------| --------------| ------------| -----------| ---------|
1919| [ ` gpt-oss-120b ` ] ( #gpt-oss-120b ) | OpenAI | 128k | Text | H100 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) |
2020| [ ` whisper-large-v3 ` ] ( #whisper-large-v3 ) | OpenAI | - | Audio transcription | L4, L40S, H100, H100-SXM-2 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) |
21- | [ ` qwen3-235b-a22b-instruct-2507 ` ] ( #qwen3-235b-a22b-instruct-2507 ) | Qwen | 40k | Text | H100-2 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) |
21+ | [ ` qwen3-235b-a22b-instruct-2507 ` ] ( #qwen3-235b-a22b-instruct-2507 ) | Qwen | 250k | Text | H100-SXM-2 (40k), H100-SXM-4 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) |
2222| [ ` gemma-3-27b-it ` ] ( #gemma-3-27b-it ) | Google | 40k | Text, Vision | H100, H100-2 | [ Gemma] ( https://ai.google.dev/gemma/terms ) |
2323| [ ` llama-3.3-70b-instruct ` ] ( #llama-33-70b-instruct ) | Meta | 128k | Text | H100 (15k), H100-2 | [ Llama 3.3 Community] ( https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct ) |
2424| [ ` llama-3.1-70b-instruct ` ] ( #llama-31-70b-instruct ) | Meta | 128k | Text | H100 (15k), H100-2 | [ Llama 3.1 Community] ( https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct/blob/main/LICENSE ) |
@@ -40,7 +40,7 @@ A quick overview of available models in Scaleway's catalog and their core attrib
4040| [ ` molmo-72b-0924 ` ] ( #molmo-72b-0924 ) | Allen AI | 50k | Text, Vision | H100-2 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) and [ Twonyi Qianwen license] ( https://huggingface.co/Qwen/Qwen2-72B/blob/main/LICENSE ) |
4141| [ ` qwen3-coder-30b-a3b-instruct ` ] ( #qwen3-coder-30b-a3b-instruct ) | Qwen | 128k | Code | L40S, H100, H100-2 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) |
4242| [ ` qwen2.5-coder-32b-instruct ` ] ( #qwen25-coder-32b-instruct ) | Qwen | 32k | Code | H100, H100-2 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) |
43- | [ ` bge-multilingual-gemma2 ` ] ( #bge-multilingual-gemma2 ) | BAAI | 4k | Embeddings | L4, L40S, H100, H100-2 | [ Gemma] ( https://ai.google.dev/gemma/terms ) |
43+ | [ ` bge-multilingual-gemma2 ` ] ( #bge-multilingual-gemma2 ) | BAAI | 8k | Embeddings | L4, L40S, H100, H100-2 | [ Gemma] ( https://ai.google.dev/gemma/terms ) |
4444| [ ` sentence-t5-xxl ` ] ( #sentence-t5-xxl ) | Sentence transformers | 512 | Embeddings | L4 | [ Apache 2.0] ( https://www.apache.org/licenses/LICENSE-2.0 ) |
4545
4646\* Maximum context length is only mentioned when instances VRAM size limits context length. Otherwise, maximum context length is the one defined by the model.
0 commit comments