Skip to content

Commit e286d39

Browse files
Merge pull request #3920 from minthigpen/patch-6
Update evaluation-metrics-built-in.md
2 parents 58591ee + e15c858 commit e286d39

File tree

1 file changed

+11
-16
lines changed

1 file changed

+11
-16
lines changed

articles/ai-foundry/concepts/evaluation-metrics-built-in.md

Lines changed: 11 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -28,24 +28,20 @@ In the development and deployment of generative AI models and applications, the
2828

2929
:::image type="content" source="../media/evaluations/risk-safety-evaluators.png" alt-text="Diagram of the risk and safety evaluators detailed in the following metric list." lightbox="../media/evaluations/risk-safety-evaluators.png":::
3030

31-
Metrics:
32-
3331
- [**Hateful and Unfair Content**](#hateful-and-unfair-content-definition-and-severity-scale): It measures the presence of any language that reflects hate towards or unfair representations of individuals and social groups based on factors including, but not limited to, race, ethnicity, nationality, gender, sexual orientation, religion, immigration status, ability, personal appearance, and body size. Unfairness occurs when AI systems treat or represent social groups inequitably, creating or contributing to societal inequities.
3432
- [**Sexual Content**](#sexual-content-definition-and-severity-scale): It measures the presence of any language pertaining to anatomical organs and genitals, romantic relationships, acts portrayed in erotic terms, pregnancy, physical sexual acts (including assault or sexual violence), prostitution, pornography, and sexual abuse.
3533
- [**Violent Content**](#violent-content-definition-and-severity-scale): It includes language pertaining to physical actions intended to hurt, injure, damage, or kill someone or something. It also includes descriptions of weapons (and related entities such as manufacturers and associations).
3634
- [**Self-harm-related Content**](#self-harm-related-content-definition-and-severity-scale): It measures the presence of any language pertaining to physical actions intended to hurt, injure, or damage one's body or kill oneself.
3735
- [**Protected Material Content**](#protected-material-definition-and-label): It measures the presence of any text that is under copyright, including song lyrics, recipes, and articles. The evaluation uses the Azure AI Content Safety Protected Material for Text service to perform the classification.
3836
- [**Direct Attack Jailbreak (UPIA)**](#jailbreak-vulnerability-definition-and-label): It measures to what extent the response fell for the jailbreak attempt. Direct attack jailbreak attempts (user prompt injected attack [UPIA]) inject prompts in the user role turn of conversations or queries to generative AI applications. Jailbreaks occur when a model response bypasses the restrictions placed on it or when an LLM deviates from the intended task or topic.
3937
- [**Indirect Attack Jailbreak (XPIA)**](#indirect-attack-definition-and-label): It measures to what extent the response fell for the indirect jailbreak attempt. Indirect attacks, also known as cross-domain prompt injected attacks (XPIA), occur when jailbreak attacks are injected into the context of a document or source that might result in altered, unexpected behavior on the part of the LLM.
40-
- **Code vulnerability**: It measures whether AI generates code with security vulnerabilities, such as code injection, tar-slip, SQL injections, stack trace exposure and other risks across Python, Java, C++, C#, Go, JavaScript, and SQL.
41-
- **Ungrounded attributes**: It measures the frequency and severity of an application generating text responses that contain ungrounded inferences about personal attributes, such as their demographics or emotional state.
38+
- [**Code Vulnerability**](#code-vulnerability-definition-and-label): It measures whether AI generates code with security vulnerabilities, such as code injection, tar-slip, SQL injections, stack trace exposure and other risks across Python, Java, C++, C#, Go, JavaScript, and SQL.
39+
- [**Ungrounded Attributes**](#ungrounded-attributes-definition-and-label): It measures the frequency and severity of an application generating text responses that contain ungrounded inferences about personal attributes, such as their demographics or emotional state.
4240

4341
- **Performance and Quality Evaluators**: Assess the accuracy, groundedness, relevance, and overall quality of generated content.
4442

4543
:::image type="content" source="../media/evaluations/quality-evaluators.png" alt-text="Diagram of the performance and quality evaluators detailed in the following metric list." lightbox="../media/evaluations/quality-evaluators.png":::
4644

47-
Metrics:
48-
4945
- **Agent Evaluators**:
5046
- **Intent Resolution**: It measures how well the agent identifies and clarifies user intent, including asking for clarifications and staying within scope.
5147
- **Tool Call Accuracy**: It measures the agent’s proficiency in selecting appropriate tools, and accurately extracting and processing inputs.
@@ -183,11 +179,11 @@ Self-harm-related content includes language pertaining to actions intended to hu
183179

184180
### Protected material definition and label
185181

186-
**Definition**:
182+
#### Protected material definition
187183

188184
Protected material is any text that is under copyright, including song lyrics, recipes, and articles. Protected material evaluation uses the Azure AI Content Safety Protected Material for Text service to perform the classification.
189185

190-
**Label:**
186+
#### Protected material evaluation label
191187

192188
|Label | Definition |
193189
| --- | --- |
@@ -210,11 +206,11 @@ You can do this with functionality and attack datasets generated with the [direc
210206

211207
### Indirect attack definition and label
212208

213-
**Definition**:
209+
#### Indirect attack definition
214210

215211
Indirect attacks, also known as cross-domain prompt injected attacks (XPIA), are when jailbreak attacks are injected into the context of a document or source that might result in an altered, unexpected behavior. *Evaluating indirect attack* is an AI-assisted evaluator and doesn't require comparative measurement like evaluating direct attacks. Generate an indirect attack jailbreak injected dataset with the [indirect attack simulator](../how-to/develop/simulator-interaction-data.md#simulating-jailbreak-attacks) then evaluate with the `IndirectAttackEvaluator`.
216212

217-
**Label:**
213+
#### Indirect attack evaluation label
218214

219215
|Label | Definition |
220216
| --- | --- |
@@ -223,11 +219,11 @@ Indirect attacks, also known as cross-domain prompt injected attacks (XPIA), are
223219

224220
### Code vulnerability definition and label
225221

226-
**Definition**:
222+
#### Code vulnerability definition
227223

228224
Code vulnerability represents security vulnerabilities in generated code (code completion) across the following programming languages: Python, Java, C++, C#, Go, JavaScript, and SQL.
229225

230-
**Label:**
226+
#### Code vulnerability evaluation label
231227

232228
|Label | Definition |
233229
| --- | --- |
@@ -266,11 +262,11 @@ Example of a result output:
266262

267263
### Ungrounded attributes definition and label
268264

269-
**Definition**:
265+
#### Ungrounded attributes definition
270266

271267
Ungrounded attributes are ungrounded inferences in generated text about a person's attributes, such as their demographics or emotional state, based on given context such as chat history or meeting transcript.
272268

273-
**Label:**
269+
#### Ungrounded attributes evaluation label
274270

275271
|Label | Definition |
276272
| --- | --- |
@@ -562,7 +558,6 @@ Currently certain AI-assisted evaluators are available only in the following reg
562558

563559
| Region | Hate and unfairness, Sexual, Violent, Self-harm, Indirect attack | Groundedness Pro | Protected material |
564560
|--|--|--|--|
565-
| UK South | deprecated as of 12/1/2024 | N/A | N/A |
566561
| East US 2 | Supported | Supported | Supported |
567562
| Sweden Central | Supported | Supported | N/A |
568563
| US North Central | Supported | N/A | N/A |
@@ -575,4 +570,4 @@ Currently certain AI-assisted evaluators are available only in the following reg
575570
- [Evaluate with the Azure AI evaluate SDK](../how-to/develop/evaluate-sdk.md)
576571
- [Evaluate your generative AI apps with the Azure AI Foundry portal](../how-to/evaluate-generative-ai-app.md)
577572
- [View the evaluation results](../how-to/evaluate-results.md)
578-
- [Transparency Note for Azure AI Foundry safety evaluations](safety-evaluations-transparency-note.md)
573+
- [Transparency Note for Azure AI Foundry safety evaluations](safety-evaluations-transparency-note.md)

0 commit comments

Comments
 (0)