You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: pages/managed-inference/how-to/create-deployment.mdx
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -29,7 +29,7 @@ dates:
29
29
- Choose the geographical **region** for the deployment.
30
30
- For custom models: Choose the model quantization.
31
31
<Messagetype="tip">
32
-
Each model comes with a default quantization. Select lower bits quantization to improve performance and enable model to run on smaller GPU Nodes, while potentially reducing precision.
32
+
Each model comes with a default quantization. Select lower bits quantization to improve performance and enable the model to run on smaller GPU nodes, while potentially reducing precision.
33
33
</Message>
34
34
- Specify the GPU Instance type to be used with your deployment.
35
35
4. Enter a **name** for the deployment, and optional tags.
0 commit comments