You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/evaluation-evaluators/agent-evaluators.md
+84-2Lines changed: 84 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -112,15 +112,97 @@ If you're building agents outside of Azure AI Agent Service, this evaluator acce
112
112
- the correctness of parameters used in tool calls;
113
113
- the counts of missing or excessive calls.
114
114
115
-
> [!NOTE]
116
-
> `ToolCallAccuracyEvaluator` only supports Azure AI Agent's Function Tool evaluation, but doesn't support Built-in Tool evaluation. The agent run must have at least one Function Tool call and no Built-in Tool calls made to be evaluated.
115
+
#### Tool call evaluation support
116
+
`ToolCallAccuracyEvaluator` supports evaluation in Azure AI Agent for the following tools:
117
+
1. File Search
118
+
2. Azure AI Search
119
+
3. Bing Grounding
120
+
4. Bing Custom Search
121
+
5. SharePoint Grounding
122
+
6. Code Interpreter
123
+
7. Fabric Data Agent
124
+
8. OpenAPI
125
+
9. Function Tool (user-defined tools)
126
+
However, if a non-supported tool is used in the agent run, it outputs a "pass" and a reason that evaluating the invoked tool(s) isn't supported, for ease of filtering out these cases. It is recommended that you wrap non-supported tools as user-defined tools to enable evaluation.
117
127
118
128
### Tool call accuracy example
119
129
120
130
```python
121
131
from azure.ai.evaluation import ToolCallAccuracyEvaluator
If you use [Foundry Agent Service](../../../ai-services/agents/overview.md), you can seamlessly evaluate your agents via our converter support for Azure AI agent threads and runs. We support this list of evaluators for Azure AI agent messages from our converter:
42
+
If you use [Foundry Agent Service](../../../ai-services/agents/overview.md), you can seamlessly evaluate your agents via our converter support for Azure AI agents and Semantic Kernel's Chat Completion and Azure AI agents. This list of evaluators accept agent messages returnd by our converter:
43
43
44
-
### Evaluators supported for evaluation data converter
If you are building other agents with a different schema, you can convert them into the general openai-style [agent message schema](#agent-message-schema) and use the above evaluators.
47
+
48
+
More generally, if you can parse the agent messages into the [required data formats](./evaluate-sdk.md#data-requirements-for-built-in-evaluators), you can also use the following evaluators:
> `ToolCallAccuracyEvaluator` only supports Foundry Agent's Function Tool evaluation (user-defined Python functions), but doesn't support other Tool evaluation. If an agent run invoked a tool other than Function Tool, it outputs a "pass" and a reason that evaluating the invoked tool(s) isn't supported.
52
+
53
+
#### Tool call evaluation support
54
+
`ToolCallAccuracyEvaluator` supports evaluation in Azure AI Agent for the following tools:
55
+
1. File Search
56
+
2. Azure AI Search
57
+
3. Bing Grounding
58
+
4. Bing Custom Search
59
+
5. SharePoint Grounding
60
+
6. Code Interpreter
61
+
7. Fabric Data Agent
62
+
8. OpenAPI
63
+
9. Function Tool (user-defined tools)
64
+
However, if a non-supported tool is used in the agent run, it outputs a "pass" and a reason that evaluating the invoked tool(s) isn't supported, for ease of filtering out these cases. It is recommended that you wrap non-supported tools as user-defined tools to enable evaluation.
51
65
52
66
Here's an example that shows you how to seamlessly build and evaluate an Azure AI agent. Separately from evaluation, Azure AI Foundry Agent Service requires `pip install azure-ai-projects azure-identity`, an Azure AI project connection string, and the supported models.
0 commit comments