Skip to content

Commit 6166a30

Browse files
committed
additional data collection preview references
1 parent 28058f8 commit 6166a30

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

articles/machine-learning/how-to-deploy-models-llama.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -401,7 +401,7 @@ Follow these steps to deploy a model such as `Llama-2-7b-chat` to a real-time en
401401
402402
1. Select the **Virtual machine** and the **Instance count** that you want to assign to the deployment.
403403
1. Select if you want to create this deployment as part of a new endpoint or an existing one. Endpoints can host multiple deployments while keeping resource configuration exclusive for each of them. Deployments under the same endpoint share the endpoint URI and its access keys.
404-
1. Indicate if you want to enable **Inferencing data collection (preview)**.
404+
1. Indicate if you want to enable **Inferencing data collection**.
405405
1. Indicate if you want to enable **Package Model (preview)**.
406406
1. Select **Deploy**. After a few moments, the endpoint's **Details** page opens up.
407407
1. Wait for the endpoint creation and deployment to finish. This step can take a few minutes.

articles/machine-learning/prompt-flow/how-to-deploy-for-real-time-inference.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -99,7 +99,7 @@ This step allows you to configure the basic settings of the deployment.
9999
|Deployment name| - Within the same endpoint, deployment name should be unique. <br> - If you select an existing endpoint, and input an existing deployment name, then that deployment will be overwritten with the new configurations. |
100100
|Virtual machine| The VM size to use for the deployment. For the list of supported sizes, see [Managed online endpoints SKU list](../reference-managed-online-endpoints-vm-sku-list.md).|
101101
|Instance count| The number of instances to use for the deployment. Specify the value on the workload you expect. For high availability, we recommend that you set the value to at least 3. We reserve an extra 20% for performing upgrades. For more information, see [managed online endpoints quotas](../how-to-manage-quotas.md#azure-machine-learning-online-endpoints-and-batch-endpoints)|
102-
|Inference data collection (preview)| If you enable this, the flow inputs and outputs will be auto collected in an Azure Machine Learning data asset, and can be used for later monitoring. To learn more, see [how to monitor generative ai applications.](how-to-monitor-generative-ai-applications.md)|
102+
|Inference data collection| If you enable this, the flow inputs and outputs will be auto collected in an Azure Machine Learning data asset, and can be used for later monitoring. To learn more, see [how to monitor generative ai applications.](how-to-monitor-generative-ai-applications.md)|
103103
|Application Insights diagnostics| If you enable this, system metrics during inference time (such as token count, flow latency, flow request, and etc.) will be collected into workspace default Application Insights. To learn more, see [prompt flow serving metrics](#view-prompt-flow-endpoints-specific-metrics-optional).|
104104

105105

0 commit comments

Comments
 (0)