Skip to content

Commit 25aef4d

Browse files
authored
Create mistral-small-24b-instruct-2501.mdx
1 parent 5007743 commit 25aef4d

File tree

1 file changed

+77
-0
lines changed

1 file changed

+77
-0
lines changed
Lines changed: 77 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,77 @@
1+
---
2+
meta:
3+
title: Understanding the Mistral-small-24b-base-2501 model
4+
description: Deploy your own secure Mistral-small-24b-base-2501 model with Scaleway Managed Inference. Privacy-focused, fully managed.
5+
content:
6+
h1: Understanding the Mistral-small-24b-base-2501 model
7+
paragraph: This page provides information on the Mistral-small-24b-base-2501 model
8+
tags:
9+
dates:
10+
validation: 2025-03-04
11+
posted: 2025-03-04
12+
categories:
13+
- ai-data
14+
---
15+
16+
## Model overview
17+
18+
| Attribute | Details |
19+
|-----------------|------------------------------------|
20+
| Provider | [Mistral](https://mistral.ai/technology/#models) |
21+
| Compatible Instances | L40S, H100, H100-2 (FP8) |
22+
| Context size | 32K tokens |
23+
24+
## Model name
25+
26+
```bash
27+
mistral/mistral-small-24b-instruct-2501:fp8
28+
```
29+
30+
## Compatible Instances
31+
32+
| Instance type | Max context length |
33+
| ------------- |-------------|
34+
| L40 | 20k (FP8) |
35+
| H100 | 32k (FP8) |
36+
| H100-2 | 32k (FP8) |
37+
38+
## Model introduction
39+
40+
Mistral Small 24B Instruct is a state-of-the-art transformer model of 24B parameters, built by Mistral.
41+
This model is open-weight and distributed under the Apache 2.0 license.
42+
43+
## Why is it useful?
44+
45+
- Mistral Small 24B offers a large context window of up to 32k tokens and provide both conversational and reasoning capabilities.
46+
- This model supports multiple languages, including English, French, German, Spanish, Italian, Chinese, Japanese, Korean, Portuguese, Dutch, and Polish.
47+
- It superseeds Mistral Nemo Instruct, although its tokens throughput is slightly lower.
48+
49+
## How to use it
50+
51+
### Sending Inference requests
52+
53+
To perform inference tasks with your Mistral model deployed at Scaleway, use the following command:
54+
55+
```bash
56+
curl -s \
57+
-H "Authorization: Bearer <IAM API key>" \
58+
-H "Content-Type: application/json" \
59+
--request POST \
60+
--url "https://<Deployment UUID>.ifr.fr-par.scaleway.com/v1/chat/completions" \
61+
--data '{"model":"mistral/mistral-small-24b-instruct-2501:fp8", "messages":[{"role": "user","content": "Tell me about Scaleway."}], "top_p": 1, "temperature": 0.7, "stream": false}'
62+
```
63+
64+
Make sure to replace `<IAM API key>` and `<Deployment UUID>` with your actual [IAM API key](/iam/how-to/create-api-keys/) and the Deployment UUID you are targeting.
65+
66+
<Message type="note">
67+
Ensure that the `messages` array is properly formatted with roles (system, user, assistant) and content.
68+
</Message>
69+
70+
### Receiving Managed Inference responses
71+
72+
Upon sending the HTTP request to the public or private endpoints exposed by the server, you will receive inference responses from the managed Managed Inference server.
73+
Process the output data according to your application's needs. The response will contain the output generated by the LLM model based on the input provided in the request.
74+
75+
<Message type="note">
76+
Despite efforts for accuracy, the possibility of generated text containing inaccuracies or [hallucinations](/managed-inference/concepts/#hallucinations) exists. Always verify the content generated independently.
77+
</Message>

0 commit comments

Comments
 (0)