Skip to content

Commit 7728b4d

Browse files
Merge pull request #334 from ssalgadodev/patch-8
Update model-catalog-overview.md
2 parents 78b2e56 + 064dee1 commit 7728b4d

File tree

1 file changed

+11
-11
lines changed

1 file changed

+11
-11
lines changed

articles/ai-studio/how-to/model-catalog-overview.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -57,10 +57,10 @@ The deployment options and features available for each model vary, as described
5757

5858
Features | Managed compute | Serverless API (pay-as-you-go)
5959
--|--|--
60-
Deployment experience and billing | Model weights are deployed to dedicated virtual machines with managed online endpoints. A managed online endpoint, which can have one or more deployments, makes available a REST API for inference. You're billed for the virtual machine core hours that the deployments use. | Access to models is through a deployment that provisions an API to access the model. The API provides access to the model that Microsoft hosts and manages, for inference. You're billed for inputs and outputs to the APIs, typically in tokens. Pricing information is provided before you deploy.
60+
Deployment experience and billing | Model weights are deployed to dedicated virtual machines with managed compute. A managed compute, which can have one or more deployments, makes available a REST API for inference. You're billed for the virtual machine core hours that the deployments use. | Access to models is through a deployment that provisions an API to access the model. The API provides access to the model that Microsoft hosts and manages, for inference. You're billed for inputs and outputs to the APIs, typically in tokens. Pricing information is provided before you deploy.
6161
API authentication | Keys and Microsoft Entra authentication. | Keys only.
6262
Content safety | Use Azure AI Content Safety service APIs. | Azure AI Content Safety filters are available integrated with inference APIs. Azure AI Content Safety filters are billed separately.
63-
Network isolation | [Configure managed networks for Azure AI Studio hubs](configure-managed-network.md). | Endpoints follow your hub's public network access (PNA) flag setting. For more information, see the [Network isolation for models deployed via Serverless APIs](#network-isolation-for-models-deployed-via-serverless-apis) section later in this article.
63+
Network isolation | [Configure managed networks for Azure AI Studio hubs](configure-managed-network.md). | Managed compute follow your hub's public network access (PNA) flag setting. For more information, see the [Network isolation for models deployed via Serverless APIs](#network-isolation-for-models-deployed-via-serverless-apis) section later in this article.
6464

6565
Model | Managed compute | Serverless API (pay-as-you-go)
6666
--|--|--
@@ -74,7 +74,7 @@ Other models | Available | Not available
7474

7575
<!-- docutune:enable -->
7676

77-
:::image type="content" source="../media/explore/platform-service-cycle.png" alt-text="Diagram that shows models as a service and the service cycle of real-time endpoints." lightbox="../media/explore/platform-service-cycle.png":::
77+
:::image type="content" source="../media/explore/platform-service-cycle.png" alt-text="Diagram that shows models as a service and the service cycle of managed computes." lightbox="../media/explore/platform-service-cycle.png":::
7878

7979
## Managed compute
8080

@@ -94,7 +94,7 @@ The registries build on top of a highly scalable and enterprise-ready infrastruc
9494

9595
### Deployment of models for inference with managed compute
9696

97-
Models available for deployment to managed compute can be deployed to Azure Machine Learning online endpoints for real-time inference. Deploying to managed compute requires you to have a virtual machine quota in your Azure subscription for the specific products that you need to optimally run the model. Some models allow you to deploy to a [temporarily shared quota for model testing](deploy-models-open.md).
97+
Models available for deployment to managed compute can be deployed to Azure Machine Learning managed compute for real-time inference. Deploying to managed compute requires you to have a virtual machine quota in your Azure subscription for the specific products that you need to optimally run the model. Some models allow you to deploy to a [temporarily shared quota for model testing](deploy-models-open.md).
9898

9999
Learn more about deploying models:
100100

@@ -151,25 +151,25 @@ Pay-as-you-go billing is available only to users whose Azure subscription belong
151151

152152
### Network isolation for models deployed via serverless APIs
153153

154-
Endpoints for models deployed as serverless APIs follow the PNA flag setting of the AI Studio hub that has the project in which the deployment exists. To help secure your MaaS endpoint, disable the PNA flag on your AI Studio hub. You can help secure inbound communication from a client to your endpoint by using a private endpoint for the hub.
154+
Managed computes for models deployed as serverless APIs follow the public network access flag setting of the AI Studio hub that has the project in which the deployment exists. To help secure your managed compute, disable the public network access flag on your AI Studio hub. You can help secure inbound communication from a client to your managed compute by using a private endpoint for the hub.
155155

156-
To set the PNA flag for the AI Studio hub:
156+
To set the public network access flag for the AI Studio hub:
157157

158158
* Go to the [Azure portal](https://ms.portal.azure.com/).
159159
* Search for the resource group to which the hub belongs, and select your AI Studio hub from the resources listed for this resource group.
160160
* On the hub overview page, on the left pane, go to **Settings** > **Networking**.
161-
* On the **Public access** tab, you can configure settings for the PNA flag.
161+
* On the **Public access** tab, you can configure settings for the public network access flag.
162162
* Save your changes. Your changes might take up to five minutes to propagate.
163163

164164
#### Limitations
165165

166-
* If you have an AI Studio hub with a private endpoint created before July 11, 2024, new MaaS endpoints added to projects in this hub won't follow the networking configuration of the hub. Instead, you need to create a new private endpoint for the hub and create new serverless API deployments in the project so that the new deployments can follow the hub's networking configuration.
166+
* If you have an AI Studio hub with a managed compute created before July 11, 2024, managed computes added to projects in this hub won't follow the networking configuration of the hub. Instead, you need to create a new managed compute for the hub and create new serverless API deployments in the project so that the new deployments can follow the hub's networking configuration.
167167

168-
* If you have an AI Studio hub with MaaS deployments created before July 11, 2024, and you enable a private endpoint on this hub, the existing MaaS deployments won't follow the hub's networking configuration. For serverless API deployments in the hub to follow the hub's networking configuration, you need to create the deployments again.
168+
* If you have an AI Studio hub with MaaS deployments created before July 11, 2024, and you enable a managed compute on this hub, the existing MaaS deployments won't follow the hub's networking configuration. For serverless API deployments in the hub to follow the hub's networking configuration, you need to create the deployments again.
169169

170-
* Currently, [Azure OpenAI On Your Data](/azure/ai-services/openai/concepts/use-your-data) support isn't available for MaaS deployments in private hubs, because private hubs have the PNA flag disabled.
170+
* Currently, [Azure OpenAI On Your Data](/azure/ai-services/openai/concepts/use-your-data) support isn't available for MaaS deployments in private hubs, because private hubs have the public network access flag disabled.
171171

172-
* Any network configuration change (for example, enabling or disabling the PNA flag) might take up to five minutes to propagate.
172+
* Any network configuration change (for example, enabling or disabling the public network access flag) might take up to five minutes to propagate.
173173

174174
## Related content
175175

0 commit comments

Comments
 (0)