Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions pages/managed-inference/how-to/create-deployment.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,10 @@ dates:
Some models may require acceptance of an end-user license agreement. If prompted, review the terms and conditions and accept the license accordingly.
</Message>
- Choose the geographical **region** for the deployment.
- For custom models: Choose the model quantization.
<Message type="tip">
Each model comes with a default quantization. Select lower bits quantization to improve performance and enable model to run on smaller GPU Nodes, while potentially reducing precision.
</Message>
- Specify the GPU Instance type to be used with your deployment.
4. Enter a **name** for the deployment, and optional tags.
5. Configure the **network connectivity** settings for the deployment:
Expand Down
Loading