You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We support AzureOpenAI or OpenAI [reasoning models](../../../ai-services/openai/how-to/reasoning.md) and non-reasoning models for the LLM-judge depending on the evaluators:
58
58
@@ -65,7 +65,7 @@ For complex evaluation that requires refined reasoning, we recommend a strong re
65
65
66
66
## Intent resolution
67
67
68
-
`IntentResolutionEvaluator` measures how well the system identifies and understands a user's request, including how well it scopes the user’s intent, asks clarifying questions, and reminds end users of its scope of capabilities. Higher score means better identification of user intent.
68
+
`IntentResolutionEvaluator` measures how well the system identifies and understands a user's request, including how well it scopes the user's intent, asks clarifying questions, and reminds end users of its scope of capabilities. Higher score means better identification of user intent.
69
69
70
70
### Intent resolution example
71
71
@@ -99,11 +99,9 @@ The numerical score is on a Likert scale (integer 1 to 5) and a higher score is
99
99
}
100
100
}
101
101
102
-
103
-
104
102
```
105
103
106
-
If you're building agents outside of Azure AI Agent Service, this evaluator accepts a schema typical for agent messages. To learn more, see our sample notebook for [Intent Resolution](https://aka.ms/intentresolution-sample).
104
+
If you're building agents outside of Azure AI Foundry Agent Service, this evaluator accepts a schema typical for agent messages. To learn more, see our sample notebook for [Intent Resolution](https://aka.ms/intentresolution-sample).
107
105
108
106
## Tool call accuracy
109
107
@@ -113,17 +111,20 @@ If you're building agents outside of Azure AI Agent Service, this evaluator acce
113
111
- the counts of missing or excessive calls.
114
112
115
113
#### Tool call evaluation support
116
-
`ToolCallAccuracyEvaluator` supports evaluation in Azure AI Agent for the following tools:
117
-
1. File Search
118
-
2. Azure AI Search
119
-
3. Bing Grounding
120
-
4. Bing Custom Search
121
-
5. SharePoint Grounding
122
-
6. Code Interpreter
123
-
7. Fabric Data Agent
124
-
8. OpenAPI
125
-
9. Function Tool (user-defined tools)
126
-
However, if a non-supported tool is used in the agent run, it outputs a "pass" and a reason that evaluating the invoked tool(s) isn't supported, for ease of filtering out these cases. It is recommended that you wrap non-supported tools as user-defined tools to enable evaluation.
114
+
115
+
`ToolCallAccuracyEvaluator` supports evaluation in Azure AI Foundry Agent Service for the following tools:
116
+
117
+
- File Search
118
+
- Azure AI Search
119
+
- Bing Grounding
120
+
- Bing Custom Search
121
+
- SharePoint Grounding
122
+
- Code Interpreter
123
+
- Fabric Data Agent
124
+
- OpenAPI
125
+
- Function Tool (user-defined tools)
126
+
127
+
However, if a non-supported tool is used in the agent run, it outputs a "pass" and a reason that evaluating the invoked tool(s) isn't supported, for ease of filtering out these cases. It's recommended that you wrap non-supported tools as user-defined tools to enable evaluation.
127
128
128
129
### Tool call accuracy example
129
130
@@ -270,7 +271,7 @@ If you're building agents outside of Azure AI Agent Service, this evaluator acce
270
271
271
272
## Task adherence
272
273
273
-
In various task-oriented AI systems such as agentic systems, it's important to assess whether the agent has stayed on track to complete a given task instead of making inefficient or out-of-scope steps. `TaskAdherenceEvaluator` measures how well an agent’s response adheres to their assigned tasks, according to their task instruction (extracted from system message and user query), and available tools. Higher score means better adherence of the system instruction to resolve the given task.
274
+
In various task-oriented AI systems such as agentic systems, it's important to assess whether the agent has stayed on track to complete a given task instead of making inefficient or out-of-scope steps. `TaskAdherenceEvaluator` measures how well an agent's response adheres to their assigned tasks, according to their task instruction (extracted from system message and user query), and available tools. Higher score means better adherence of the system instruction to resolve the given task.
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/create-resource-terraform.md
+36-16Lines changed: 36 additions & 16 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -15,21 +15,23 @@ ai-usage: ai-assisted
15
15
#customer intent: As a Terraform user, I want to see how to configure Azure AI Foundry using Terraform, so I can automate my setup.
16
16
---
17
17
18
-
# Use Terraform to create Azure AI Foundry resource
18
+
# Use Terraform to manage Azure AI Foundry resources
19
19
20
-
In this article, you use Terraform to create an [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs) resource. You learn how to use Terraform to create AI Foundry management configurations including projects, deployments, and connections.
20
+
In this article, you use Terraform to manage an [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs) resource. You learn how to use Terraform to manage AI Foundry management configurations including projects, deployments, and connections.
21
21
22
-
The examples used in article use the [AzAPI](/azure/developer/terraform/overview-azapi-provider) Terraform provider. Similar [AzureRM](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/ai_services) provider support is available via the classic `AzureRM_AIServices` module (using the `aiservices` kind as its value), but is limited in functionality to resource and deployment creation.
22
+
You can use either the Terraform [AzAPI Provider](/azure/developer/terraform/overview-azapi-provider) or [AzureRM Provider](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/cognitive_account) to manage Azure AI Foundry resources. While the AzAPI provider lets you access all Foundry control plane configurations including preview features, the AzureRM variant is limited to core management capabilities:
23
+
24
+
|Action|AzAPI Provider|AzureRM Provider|
25
+
| --- | --- | --- |
26
+
| Create a resource group | ✅ | ✅ |
27
+
| Create an AI Foundry resource | ✅ | ✅ |
28
+
| Configure deployments | ✅ | ✅ |
29
+
| Configure projects | ✅ | - |
30
+
| Configure a connection to knowledge and tools | ✅ | - |
31
+
| Configure a capability host (for advanced tool configurations like [Agent standard setup](../agents/concepts/capability-hosts.md)) | ✅ | - |
@@ -39,10 +41,7 @@ The examples used in article use the [AzAPI](/azure/developer/terraform/overview
39
41
40
42
## Implement a basic AI Foundry configuration using Terraform code
41
43
42
-
> [!NOTE]
43
-
> The sample code for this article is located in the [Azure Terraform GitHub repo](https://github.com/Azure/terraform/tree/master/quickstart/101-azure-ai-foundry). You can view the log file containing the [test results from current and previous versions of Terraform](https://github.com/Azure/terraform/tree/master/quickstart/101-azure-ai-foundry/TestRecord.md). You may need to update the resource provider versions used in the template to use the latest available versions.
44
-
>
45
-
> See more [articles and sample code showing how to use Terraform to manage Azure resources](/azure/terraform)
44
+
# [AzAPI Provider](#tab/azapi)
46
45
47
46
1. Create a directory in which to test and run the sample Terraform code and make it the current directory.
48
47
@@ -58,6 +57,24 @@ The examples used in article use the [AzAPI](/azure/developer/terraform/overview
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/develop/agent-evaluate-sdk.md
+14-13Lines changed: 14 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -39,7 +39,7 @@ pip install azure-ai-evaluation
39
39
40
40
## Evaluate Azure AI agents
41
41
42
-
If you use [Foundry Agent Service](../../../ai-services/agents/overview.md), you can seamlessly evaluate your agents via our converter support for Azure AI agents and Semantic Kernel's Chat Completion and Azure AI agents. This list of evaluators are supported for evaluation data returned by the converter: `IntentResolution`, `ToolCallAccuracy`, `TaskAdherence`, `Relevance`, `Groundedness`.
42
+
If you use [Foundry Agent Service](../../../ai-services/agents/overview.md), you can seamlessly evaluate your agents using our converter support for Azure AI agents and Semantic Kernelagents. The following evaluators are supported for evaluation data returned by the converter: `IntentResolution`, `ToolCallAccuracy`, `TaskAdherence`, `Relevance`, and`Groundedness`.
43
43
44
44
> [!NOTE]
45
45
> If you are building other agents that output a different schema, you can convert them into the general openai-style [agent message schema](#agent-message-schema) and use the above evaluators.
@@ -48,15 +48,17 @@ If you use [Foundry Agent Service](../../../ai-services/agents/overview.md), you
48
48
49
49
#### Tool call evaluation support
50
50
`ToolCallAccuracyEvaluator` supports evaluation in Azure AI Agent for the following tools:
51
-
1. File Search
52
-
2. Azure AI Search
53
-
3. Bing Grounding
54
-
4. Bing Custom Search
55
-
5. SharePoint Grounding
56
-
6. Code Interpreter
57
-
7. Fabric Data Agent
58
-
8. OpenAPI
59
-
9. Function Tool (user-defined tools)
51
+
52
+
- File Search
53
+
- Azure AI Search
54
+
- Bing Grounding
55
+
- Bing Custom Search
56
+
- SharePoint Grounding
57
+
- Code Interpreter
58
+
- Fabric Data Agent
59
+
- OpenAPI
60
+
- Function Tool (user-defined tools)
61
+
60
62
However, if a non-supported tool is used in the agent run, it outputs a "pass" and a reason that evaluating the invoked tool(s) isn't supported, for ease of filtering out these cases. It is recommended that you wrap non-supported tools as user-defined tools to enable evaluation.
61
63
62
64
Here's an example that shows you how to seamlessly build and evaluate an Azure AI agent. Separately from evaluation, Azure AI Foundry Agent Service requires `pip install azure-ai-projects azure-identity`, an Azure AI project connection string, and the supported models.
@@ -215,10 +217,10 @@ reasoning_model_config = {
215
217
"api_version": os.getenv("AZURE_API_VERSION"),
216
218
}
217
219
218
-
# Evaluators you may want to use reasoning models with
220
+
# Evaluators you might want to use with reasoning models
219
221
quality_evaluators = {evaluator.__name__: evaluator(model_config=reasoning_model_config, is_reasoning_model=True) for evaluator in [IntentResolutionEvaluator, TaskAdherenceEvaluator, ToolCallAccuracyEvaluator]}
220
222
221
-
# Other evaluators you may NOT want to use reasoning models
223
+
# Other evaluators you might NOT want to use with reasoning models
222
224
quality_evaluators.update({ evaluator.__name__: evaluator(model_config=model_config) for evaluator in [CoherenceEvaluator, FluencyEvaluator, RelevanceEvaluator]})
223
225
224
226
## Using Azure AI Foundry (non-Hub) project endpoint, example: AZURE_AI_PROJECT=https://your-account.services.ai.azure.com/api/projects/your-project
@@ -234,7 +236,6 @@ for name, evaluator in quality_and_safety_evaluators.items():
0 commit comments