You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: pages/managed-inference/how-to/create-deployment.mdx
+4Lines changed: 4 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -28,6 +28,10 @@ dates:
28
28
Some models may require acceptance of an end-user license agreement. If prompted, review the terms and conditions and accept the license accordingly.
29
29
</Message>
30
30
- Choose the geographical **region** for the deployment.
31
+
- For custom models: Choose the model quantization.
32
+
<Messagetype="tip">
33
+
Each model comes with a default quantization. Select lower bits quantization to improve performance and enable the model to run on smaller GPU nodes, while potentially reducing precision.
34
+
</Message>
31
35
- Specify the GPU Instance type to be used with your deployment.
32
36
5. Choose the number of nodes for your deployment. Note that this feature is currently in [Public Beta](https://www.scaleway.com/betas/).
0 commit comments