Skip to content

Commit 272b6c3

Browse files
committed
Content safety update
Updated model catalog content safety page and created include file for content safety harms and categories
1 parent e88aa0c commit 272b6c3

File tree

2 files changed

+40
-2
lines changed

2 files changed

+40
-2
lines changed
Lines changed: 30 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,30 @@
1+
---
2+
title: include file
3+
description: include file
4+
ms.reviewer: osiotugo
5+
ms.author: osiotugo
6+
ms.service: azure-ai-studio
7+
ms.topic: include
8+
ms.date: 01/28/2025
9+
ms.custom: include
10+
---
11+
12+
## Understand harm categories
13+
14+
### Harm categories
15+
16+
| Category | Description |API term |
17+
| --------- | ------------------- | --- |
18+
| Hate and Fairness | Hate and fairness harms refer to any content that attacks or uses discriminatory language with reference to a person or identity group based on certain differentiating attributes of these groups. <br><br>This includes, but is not limited to:<ul><li>Race, ethnicity, nationality</li><li>Gender identity groups and expression</li><li>Sexual orientation</li><li>Religion</li><li>Personal appearance and body size</li><li>Disability status</li><li>Harassment and bullying</li></ul> | `Hate` |
19+
| Sexual | Sexual describes language related to anatomical organs and genitals, romantic relationships and sexual acts, acts portrayed in erotic or affectionate terms, including those portrayed as an assault or a forced sexual violent act against one’s will. <br><br> This includes but is not limited to:<ul><li>Vulgar content</li><li>Prostitution</li><li>Nudity and Pornography</li><li>Abuse</li><li>Child exploitation, child abuse, child grooming</li></ul> | `Sexual` |
20+
| Violence | Violence describes language related to physical actions intended to hurt, injure, damage, or kill someone or something; describes weapons, guns, and related entities. <br><br>This includes, but isn't limited to: <ul><li>Weapons</li><li>Bullying and intimidation</li><li>Terrorist and violent extremism</li><li>Stalking</li></ul> | `Violence` |
21+
| Self-Harm | Self-harm describes language related to physical actions intended to purposely hurt, injure, damage one’s body or kill oneself. <br><br> This includes, but isn't limited to: <ul><li>Eating Disorders</li><li>Bullying and intimidation</li></ul> | `SelfHarm` |
22+
23+
### Severity levels
24+
25+
| Level | Description |
26+
| --- | ---|
27+
|Safe |Content might be related to violence, self-harm, sexual, or hate categories but the terms are used in general, journalistic, scientific, medical, and similar professional contexts, which are appropriate for most audiences. |
28+
|Low |Content that expresses prejudiced, judgmental, or opinionated views, includes offensive use of language, stereotyping, use cases exploring a fictional world (for example, gaming, literature) and depictions at low intensity.|
29+
|Medium |Content that uses offensive, insulting, mocking, intimidating, or demeaning language towards specific identity groups, includes depictions of seeking and executing harmful instructions, fantasies, glorification, promotion of harm at medium intensity. |
30+
|High |Content that displays explicit and severe harmful instructions, actions, damage, or abuse; includes endorsement, glorification, or promotion of severe harmful acts, extreme or illegal forms of harm, radicalization, or nonconsensual power exchange or abuse. |

articles/ai-studio/how-to/model-catalog-content-safety.md

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,9 +20,9 @@ In this article, learn about content safety capabilities for models from the mod
2020

2121
## Content filter defaults
2222

23-
Azure AI uses a default configuration of [Azure AI Content Safety](/azure/ai-services/content-safety/overview) content filters that detect harmful content across four categories hate, self-harm, sexual, and violence for models deployed via serverless APIs. To learn more about content filtering (preview), see [Harm categories in Azure AI Content Safety](/azure/ai-services/content-safety/concepts/harm-categories).
23+
Azure AI uses a default configuration of [Azure AI Content Safety](/azure/ai-services/content-safety/overview) content filters that detect harmful content across four categories hate and fairness, self-harm, sexual, and violence for models deployed via serverless APIs. To learn more about content filtering (preview), see [Harm categories in Azure AI Content Safety](/azure/ai-services/content-safety/concepts/harm-categories).
2424

25-
The default content filtering configuration for text models is set to filter at the medium severity threshold, filtering any detected content at this level or higher. For image models, the default content filtering configuration is set at the low configuration threshold, filtering at this level or higher. Models deployed using the [Azure AI model inference service]() can create configurable filters by clicking the **Content filters** tab within the **Safety + security** page.
25+
The default content filtering configuration for text models is set to filter at the medium severity threshold, filtering any detected content at this level or higher. For image models, the default content filtering configuration is set at the low configuration threshold, filtering at this level or higher. Models deployed using the [Azure AI model inference service](/articles/ai-foundry/model-inference/how-to/configure-content-filters.md) can create configurable filters by clicking the **Content filters** tab within the **Safety + security** page.
2626

2727
> [!TIP]
2828
> Content filtering (preview) is not available for certain model types that are deployed via serverless APIs. These model types include embedding models and time series models.
@@ -34,6 +34,14 @@ Content filtering (preview) occurs synchronously as the service processes prompt
3434

3535
Suppose you decide to use an API other than the [Azure AI Model Inference API](/azure/ai-studio/reference/reference-model-inference-api) to work with a model that is deployed via a serverless API. In such a situation, content filtering (preview) isn't enabled unless you implement it separately by using Azure AI Content Safety. To get started with Azure AI Content Safety, see [Quickstart: Analyze text content](/azure/ai-services/content-safety/quickstart-text). You run a higher risk of exposing users to harmful content if you don't use content filtering (preview) when working with models that are deployed via serverless APIs.
3636

37+
[!INCLUDE [content-safety-harm-categories](/articles/ai-studio/ai-services/includes/content-safety-harm-categories.md)]
38+
3739
## How charges are calculated
3840

3941
Pricing details are viewable at [Azure AI Content Safety pricing](https://azure.microsoft.com/pricing/details/cognitive-services/content-safety/). Charges are incurred when the Azure AI Content Safety validates the prompt or completion. If Azure AI Content Safety blocks the prompt or completion, you're charged for both the evaluation of the content and the inference calls.
42+
43+
## Related content
44+
45+
- [How to configure content filters (preview) for models in Azure AI services](/articles/ai-foundry/model-inference/how-to/configure-content-filters.md)
46+
- [Azure AI Content Safety Overview](/articles/ai-services/content-safety/overview.md)
47+
- [Model catalog and collections in Azure AI Foundry portal](/articles/ai-studio/how-to/model-catalog-overview.md)

0 commit comments

Comments
 (0)