You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/evaluation-metrics-built-in.md
+11-16Lines changed: 11 additions & 16 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -28,24 +28,20 @@ In the development and deployment of generative AI models and applications, the
28
28
29
29
:::image type="content" source="../media/evaluations/risk-safety-evaluators.png" alt-text="Diagram of the risk and safety evaluators detailed in the following metric list." lightbox="../media/evaluations/risk-safety-evaluators.png":::
30
30
31
-
Metrics:
32
-
33
31
-[**Hateful and Unfair Content**](#hateful-and-unfair-content-definition-and-severity-scale): It measures the presence of any language that reflects hate towards or unfair representations of individuals and social groups based on factors including, but not limited to, race, ethnicity, nationality, gender, sexual orientation, religion, immigration status, ability, personal appearance, and body size. Unfairness occurs when AI systems treat or represent social groups inequitably, creating or contributing to societal inequities.
34
32
-[**Sexual Content**](#sexual-content-definition-and-severity-scale): It measures the presence of any language pertaining to anatomical organs and genitals, romantic relationships, acts portrayed in erotic terms, pregnancy, physical sexual acts (including assault or sexual violence), prostitution, pornography, and sexual abuse.
35
33
-[**Violent Content**](#violent-content-definition-and-severity-scale): It includes language pertaining to physical actions intended to hurt, injure, damage, or kill someone or something. It also includes descriptions of weapons (and related entities such as manufacturers and associations).
36
34
-[**Self-harm-related Content**](#self-harm-related-content-definition-and-severity-scale): It measures the presence of any language pertaining to physical actions intended to hurt, injure, or damage one's body or kill oneself.
37
35
-[**Protected Material Content**](#protected-material-definition-and-label): It measures the presence of any text that is under copyright, including song lyrics, recipes, and articles. The evaluation uses the Azure AI Content Safety Protected Material for Text service to perform the classification.
38
36
-[**Direct Attack Jailbreak (UPIA)**](#jailbreak-vulnerability-definition-and-label): It measures to what extent the response fell for the jailbreak attempt. Direct attack jailbreak attempts (user prompt injected attack [UPIA]) inject prompts in the user role turn of conversations or queries to generative AI applications. Jailbreaks occur when a model response bypasses the restrictions placed on it or when an LLM deviates from the intended task or topic.
39
37
-[**Indirect Attack Jailbreak (XPIA)**](#indirect-attack-definition-and-label): It measures to what extent the response fell for the indirect jailbreak attempt. Indirect attacks, also known as cross-domain prompt injected attacks (XPIA), occur when jailbreak attacks are injected into the context of a document or source that might result in altered, unexpected behavior on the part of the LLM.
40
-
-**Code vulnerability**: It measures whether AI generates code with security vulnerabilities, such as code injection, tar-slip, SQL injections, stack trace exposure and other risks across Python, Java, C++, C#, Go, JavaScript, and SQL.
41
-
-**Ungrounded attributes**: It measures the frequency and severity of an application generating text responses that contain ungrounded inferences about personal attributes, such as their demographics or emotional state.
38
+
-[**Code Vulnerability**](#code-vulnerability-definition-and-label): It measures whether AI generates code with security vulnerabilities, such as code injection, tar-slip, SQL injections, stack trace exposure and other risks across Python, Java, C++, C#, Go, JavaScript, and SQL.
39
+
-[**Ungrounded Attributes**](#ungrounded-attributes-definition-and-label): It measures the frequency and severity of an application generating text responses that contain ungrounded inferences about personal attributes, such as their demographics or emotional state.
42
40
43
41
-**Performance and Quality Evaluators**: Assess the accuracy, groundedness, relevance, and overall quality of generated content.
44
42
45
43
:::image type="content" source="../media/evaluations/quality-evaluators.png" alt-text="Diagram of the performance and quality evaluators detailed in the following metric list." lightbox="../media/evaluations/quality-evaluators.png":::
46
44
47
-
Metrics:
48
-
49
45
-**Agent Evaluators**:
50
46
-**Intent Resolution**: It measures how well the agent identifies and clarifies user intent, including asking for clarifications and staying within scope.
51
47
-**Tool Call Accuracy**: It measures the agent’s proficiency in selecting appropriate tools, and accurately extracting and processing inputs.
@@ -183,11 +179,11 @@ Self-harm-related content includes language pertaining to actions intended to hu
183
179
184
180
### Protected material definition and label
185
181
186
-
**Definition**:
182
+
#### Protected material definition
187
183
188
184
Protected material is any text that is under copyright, including song lyrics, recipes, and articles. Protected material evaluation uses the Azure AI Content Safety Protected Material for Text service to perform the classification.
189
185
190
-
**Label:**
186
+
#### Protected material evaluation label
191
187
192
188
|Label | Definition |
193
189
| --- | --- |
@@ -210,11 +206,11 @@ You can do this with functionality and attack datasets generated with the [direc
210
206
211
207
### Indirect attack definition and label
212
208
213
-
**Definition**:
209
+
#### Indirect attack definition
214
210
215
211
Indirect attacks, also known as cross-domain prompt injected attacks (XPIA), are when jailbreak attacks are injected into the context of a document or source that might result in an altered, unexpected behavior. *Evaluating indirect attack* is an AI-assisted evaluator and doesn't require comparative measurement like evaluating direct attacks. Generate an indirect attack jailbreak injected dataset with the [indirect attack simulator](../how-to/develop/simulator-interaction-data.md#simulating-jailbreak-attacks) then evaluate with the `IndirectAttackEvaluator`.
216
212
217
-
**Label:**
213
+
#### Indirect attack evaluation label
218
214
219
215
|Label | Definition |
220
216
| --- | --- |
@@ -223,11 +219,11 @@ Indirect attacks, also known as cross-domain prompt injected attacks (XPIA), are
223
219
224
220
### Code vulnerability definition and label
225
221
226
-
**Definition**:
222
+
#### Code vulnerability definition
227
223
228
224
Code vulnerability represents security vulnerabilities in generated code (code completion) across the following programming languages: Python, Java, C++, C#, Go, JavaScript, and SQL.
229
225
230
-
**Label:**
226
+
#### Code vulnerability evaluation label
231
227
232
228
|Label | Definition |
233
229
| --- | --- |
@@ -266,11 +262,11 @@ Example of a result output:
266
262
267
263
### Ungrounded attributes definition and label
268
264
269
-
**Definition**:
265
+
#### Ungrounded attributes definition
270
266
271
267
Ungrounded attributes are ungrounded inferences in generated text about a person's attributes, such as their demographics or emotional state, based on given context such as chat history or meeting transcript.
272
268
273
-
**Label:**
269
+
#### Ungrounded attributes evaluation label
274
270
275
271
|Label | Definition |
276
272
| --- | --- |
@@ -562,7 +558,6 @@ Currently certain AI-assisted evaluators are available only in the following reg
562
558
563
559
| Region | Hate and unfairness, Sexual, Violent, Self-harm, Indirect attack | Groundedness Pro | Protected material |
564
560
|--|--|--|--|
565
-
| UK South | deprecated as of 12/1/2024 | N/A | N/A |
566
561
| East US 2 | Supported | Supported | Supported |
567
562
| Sweden Central | Supported | Supported | N/A |
568
563
| US North Central | Supported | N/A | N/A |
@@ -575,4 +570,4 @@ Currently certain AI-assisted evaluators are available only in the following reg
575
570
-[Evaluate with the Azure AI evaluate SDK](../how-to/develop/evaluate-sdk.md)
576
571
-[Evaluate your generative AI apps with the Azure AI Foundry portal](../how-to/evaluate-generative-ai-app.md)
577
572
-[View the evaluation results](../how-to/evaluate-results.md)
578
-
-[Transparency Note for Azure AI Foundry safety evaluations](safety-evaluations-transparency-note.md)
573
+
-[Transparency Note for Azure AI Foundry safety evaluations](safety-evaluations-transparency-note.md)
0 commit comments