Skip to content

Commit f84a1d6

Browse files
committed
fix(inf): add beta info re node number
1 parent 7784799 commit f84a1d6

File tree

4 files changed

+10
-1
lines changed

4 files changed

+10
-1
lines changed

pages/managed-inference/concepts.mdx

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -71,6 +71,10 @@ LLMs are advanced artificial intelligence systems capable of understanding and g
7171
These models, such as Llama-3, are trained on vast amounts of data to learn the patterns and structures of language, enabling them to generate coherent and contextually relevant responses to queries or prompts.
7272
LLMs have applications in natural language processing, text generation, translation, and other tasks requiring sophisticated language understanding and production.
7373

74+
## Node number
75+
76+
The node number (or node count) defines the number of nodes, or Instances, that are running your Managed Inference deployment. [Increasing the node number](/managed-inference/how-to/configure-autoscaling/) scales your deployment, so that it can handle more load.
77+
7478
## Prompt
7579

7680
In the context of generative AI models, a prompt refers to the input provided to the model to generate a desired response.

pages/managed-inference/how-to/configure-autoscaling.mdx

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -15,6 +15,10 @@ categories:
1515

1616
You can scale your Managed Inference deployment up or down to match it to the incoming load of your deployment.
1717

18+
<Message type="important">
19+
This feature is currently in [Public Beta](https://www.scaleway.com/betas/).
20+
</Message>
21+
1822
<Macro id="requirements" />
1923

2024
- A Scaleway account logged into the [console](https://console.scaleway.com)

pages/managed-inference/how-to/create-deployment.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ dates:
2828
</Message>
2929
- Choose the geographical **region** for the deployment.
3030
- Specify the GPU Instance type to be used with your deployment.
31-
4. Choose the number of nodes for your deployment.
31+
4. Choose the number of nodes for your deployment. Note that this feature is currently in [Public Beta](https://www.scaleway.com/betas/).
3232
<Message type="note">
3333
High availability is only guaranteed with two or more nodes.
3434
</Message>

pages/managed-inference/quickstart.mdx

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -43,6 +43,7 @@ Here are some of the key features of Scaleway Managed Inference:
4343
</Message>
4444
- Choose the geographical **region** for the deployment.
4545
- Specify the GPU Instance type to be used with your deployment.
46+
- Choose the number of nodes for your deployment. Note that this feature is currently in [Public Beta](https://www.scaleway.com/betas/).
4647
4. Enter a **name** for the deployment, along with optional tags to aid in organization.
4748
5. Configure the **network** settings for the deployment:
4849
- Enable **Private Network** for secure communication and restricted availability within Private Networks. Choose an existing Private Network from the drop-down list, or create a new one.

0 commit comments

Comments
 (0)