You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
-[Conversation data for images and multi-modalities](./evaluate-sdk.md#conversation-support-for-images-and-multi-modal-text-and-image).
93
+
-[Conversation data for images and multi-modalities](./evaluate-sdk.md#conversation-support-for-images-and-multi-modal-text-and-image)
94
94
95
95
To learn more about input data formats for evaluating agents, see [Evaluate Azure AI agents](./agent-evaluate-sdk.md#evaluate-azure-ai-agents) and [Evaluate other agents](./agent-evaluate-sdk.md#evaluating-other-agents).
After you build a prompt flow and test it properly, you can deploy it as an online endpoint. Deployments are hosted in an endpoint. They can receive data from clients and send responses in real time.
25
+
After you build a prompt flow and test it, you can deploy it as an online endpoint. Deployments are hosted in an endpoint. They can receive data from clients and send responses in real time.
26
26
27
27
You can invoke the endpoint for real-time inference for chat, a copilot, or another generative AI application. Prompt flows support endpoint deployment from a flow or a bulk test run.
28
28
@@ -40,9 +40,9 @@ In this article, you learn how to deploy a flow as a managed online endpoint for
40
40
41
41
To deploy a prompt flow as an online endpoint, you need:
42
42
43
-
- An Azure subscription. If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/pricing/purchase-options/azure-account?cid=msft_learn) before you begin.
43
+
- An Azure subscription. If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/pricing/purchase-options/azure-account?cid=msft_learn).
44
44
- An Azure AI Foundry project.
45
-
- A `Microsoft.PolicyInsights` resource provider registered in your subscription. For more information on how to register a resource provider, see [Register a resource provider](/azure/azure-resource-manager/management/resource-providers-and-types#register-resource-provider-1).
45
+
- A `Microsoft.PolicyInsights` resource provider registered in your subscription. For more information, see [Register a resource provider](/azure/azure-resource-manager/management/resource-providers-and-types#register-resource-provider-1).
46
46
47
47
## Create an online deployment
48
48
@@ -85,7 +85,7 @@ For information about how to deploy a base model, see [Deploy models with Azure
85
85
86
86
### Requirements text file
87
87
88
-
Optionally, you can specify extra packages that you need in `requirements.txt`. You can find `requirements.txt` in the root folder of your flow folder. When you deploy a prompt flow to a managed online endpoint in the UI, by default, the deployment uses the environment that was created based on the base image specified in `flow.dag.yaml` and the dependencies specified in `requirements.txt` of the flow.
88
+
Optionally, you can specify extra packages that you need in `requirements.txt`. You can find `requirements.txt` in the root folder of your flow folder. When you deploy a prompt flow to a managed online endpoint in the UI, by default, the deployment uses the environment that was created based on the base image specified in `flow.dag.yaml` and the dependencies specified in `requirements.txt`.
89
89
90
90
The base image specified in `flow.dag.yaml` is created based on the prompt flow base image `mcr.microsoft.com/azureml/promptflow/promptflow-runtime-stable:<newest_version>`. To see the latest version, see [this list](https://mcr.microsoft.com/v2/azureml/promptflow/promptflow-runtime-stable/tags/list). If you don't specify the base image in `flow.dag.yaml`, the deployment uses the default base image `mcr.microsoft.com/azureml/promptflow/promptflow-runtime-stable:latest`.
91
91
@@ -123,7 +123,7 @@ This setting identifies the authentication method for the endpoint. Key-based au
123
123
124
124
The endpoint needs to access Azure resources for inferencing, such as Azure Container Registry or your Azure AI Foundry hub connections. You can allow the endpoint permission to access Azure resources by giving permission to its managed identity.
125
125
126
-
System-assigned identity is automatically created after your endpoint is created. The user creates the user-assigned identity. For more information, see [Managed identities for Azure resources](/azure/active-directory/managed-identities-azure-resources/overview).
126
+
System-assigned identity is created after your endpoint is created. The user creates the user-assigned identity. For more information, see [Managed identities for Azure resources](/azure/active-directory/managed-identities-azure-resources/overview).
- A prompt flow ready for deployment. If you don't have one, see [Develop a prompt flow](flow-develop.md).
45
-
- Azure role-based access controls are used to grant access to operations in the Azure AI Foundry portal. To perform the steps in this article, your user account must be assigned the Azure AI Developer role on the resource group. For more information on permissions, see [Role-based access control for Azure AI Foundry](../concepts/rbac-azure-ai-foundry.md).
45
+
- Azure role-based access controls are used to grant access to operations in the Azure AI Foundry portal. For this article, your user account must be assigned the Azure AI Developer role on the resource group. For more information, see [Role-based access control for Azure AI Foundry](../concepts/rbac-azure-ai-foundry.md).
46
46
47
47
# [Python SDK](#tab/python)
48
48
@@ -56,7 +56,9 @@ pip install -U azure-ai-ml
56
56
57
57
## Requirements for monitoring metrics
58
58
59
-
Generative pretrained transformer (GPT) language models generate monitoring metrics that are configured with specific evaluation instructions, or *prompt templates*. These models act as evaluator models for sequence-to-sequence tasks. Use of this technique to generate monitoring metrics shows strong empirical results and high correlation with human judgment when compared to standard generative AI evaluation metrics. For more information about prompt flow evaluation, see [Submit a batch test and evaluate a flow](./flow-bulk-test-evaluation.md) and [Observability in generative AI](../concepts/observability.md).
59
+
Generative pretrained transformer (GPT) language models generate monitoring metrics that are configured with specific evaluation instructions, or *prompt templates*. These models act as evaluator models for sequence-to-sequence tasks.
60
+
61
+
Using this technique to generate monitoring metrics shows strong empirical results and high correlation with human judgment when compared to standard generative AI evaluation metrics. For more information about prompt flow evaluation, see [Submit a batch test and evaluate a flow](./flow-bulk-test-evaluation.md) and [Observability in generative AI](../concepts/observability.md).
60
62
61
63
The following GPT models generate monitoring metrics. These GPT models are supported with monitoring and configured as your Azure OpenAI resource:
62
64
@@ -96,11 +98,11 @@ The parameters that are configured in your data asset dictate what metrics you c
96
98
| Groundedness | Required | Required | Required|
97
99
| Relevance | Required | Required | Required|
98
100
99
-
For more information on the specific data mapping requirements for each metric, see [Query and response metric requirements](evaluate-generative-ai-app.md#query-and-response-metric-requirements).
101
+
For information on the specific data mapping requirements for each metric, see [Query and response metric requirements](evaluate-generative-ai-app.md#query-and-response-metric-requirements).
100
102
101
103
## Set up monitoring for a prompt flow
102
104
103
-
To set up monitoring for your prompt flow application, first deploy your prompt flow application with inferencing data collection. Then you can configure monitoring for the deployed application.
105
+
To set up monitoring for your prompt flow application, deploy your prompt flow application with inferencing data collection. Then configure monitoring for the deployed application.
104
106
105
107
### Deploy your prompt flow application with inferencing data collection
106
108
@@ -119,7 +121,7 @@ In this section, you learn how to deploy your prompt flow with inferencing data
119
121
120
122
1. Confirm that your flow runs successfully and that the required inputs and outputs are configured for the [metrics that you want to assess](#supported-metrics-for-monitoring).
121
123
122
-
The minimum required parameters are question/inputs and answer/outputs. Supplying the minimum parameters provides only two metrics: _coherence_ and _fluency_. You must configure your flow as described in [Requirements for monitoring metrics](#requirements-for-monitoring-metrics). This example uses `question` (Question) and `chat_history` (Context) as the flow inputs, and `answer` (Answer) as the flow output.
124
+
The minimum required parameters are question/inputs and answer/outputs. Supplying the minimum parameters provides only two metrics: _coherence_ and _fluency_. Configure your flow as described in [Requirements for monitoring metrics](#requirements-for-monitoring-metrics). This example uses `question` (Question) and `chat_history` (Context) as the flow inputs, and `answer` (Answer) as the flow output.
123
125
124
126
1. Select **Deploy** to begin deploying your flow.
0 commit comments