Skip to content

Commit cae81c4

Browse files
committed
acrolinx
1 parent c877173 commit cae81c4

File tree

1 file changed

+8
-8
lines changed

1 file changed

+8
-8
lines changed

articles/ai-studio/ai-services/content-safety-overview.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -23,12 +23,12 @@ You can use Azure AI Content Safety for many scenarios:
2323

2424
**Text content**:
2525
- Moderate text content: This feature scans and moderates text content, identifying and categorizing it based on different levels of severity to ensure appropriate responses.
26-
- Groundedness detection: This filter determines if the AI's responses are based on trusted, user-provided sources, ensuring that the answers are "grounded" in the intended material. This is helpful for improving the reliability and factual accuracy of responses.
27-
- Protected material detection for text: This identifies protected text material, such as known song lyrics, articles, or other content, ensuring that the AI doesn’t output these without permission.
26+
- Groundedness detection: This filter determines if the AI's responses are based on trusted, user-provided sources, ensuring that the answers are "grounded" in the intended material. Groundedness detection is helpful for improving the reliability and factual accuracy of responses.
27+
- Protected material detection for text: This feature identifies protected text material, such as known song lyrics, articles, or other content, ensuring that the AI doesn’t output this content without permission.
2828
- Protected material detection for code: Detects code segments in the model's output that match known code from public repositories, helping to prevent uncredited or unauthorized reproduction of source code.
29-
- Prompt shields: These provide a unified API to address "Jailbreak" and "Indirect Attacks":
30-
- Jailbreak Attacks: These are attempts by users to manipulate the AI into bypassing its safety protocols or ethical guidelines. Examples include prompts designed to trick the AI into giving inappropriate responses or performing tasks it was programmed to avoid.
31-
- Indirect Attacks: Also known as Cross-Domain Prompt Injection Attacks, these involve embedding malicious prompts within documents that the AI may process. For example, if a document contains hidden instructions, the AI might inadvertently follow them, leading to unintended or unsafe outputs.
29+
- Prompt shields: This feature provides a unified API to address "Jailbreak" and "Indirect Attacks":
30+
- Jailbreak Attacks: Attempts by users to manipulate the AI into bypassing its safety protocols or ethical guidelines. Examples include prompts designed to trick the AI into giving inappropriate responses or performing tasks it was programmed to avoid.
31+
- Indirect Attacks: Also known as Cross-Domain Prompt Injection Attacks, indirect attacks involve embedding malicious prompts within documents that the AI might process. For example, if a document contains hidden instructions, the AI might inadvertently follow them, leading to unintended or unsafe outputs.
3232

3333
**Image content**:
3434
- Moderate image content: Similar to text moderation, this feature filters and assesses image content to detect inappropriate or harmful visuals.
@@ -43,7 +43,7 @@ You can use Azure AI Content Safety for many scenarios:
4343
### Harm categories
4444

4545
| Category | Description |API term |
46-
| --------- | ------------------- | -- |
46+
| --------- | ------------------- | --- |
4747
| Hate and Fairness | Hate and fairness harms refer to any content that attacks or uses discriminatory language with reference to a person or identity group based on certain differentiating attributes of these groups. <br><br>This includes, but is not limited to:<ul><li>Race, ethnicity, nationality</li><li>Gender identity groups and expression</li><li>Sexual orientation</li><li>Religion</li><li>Personal appearance and body size</li><li>Disability status</li><li>Harassment and bullying</li></ul> | `Hate` |
4848
| Sexual | Sexual describes language related to anatomical organs and genitals, romantic relationships and sexual acts, acts portrayed in erotic or affectionate terms, including those portrayed as an assault or a forced sexual violent act against one’s will. <br><br> This includes but is not limited to:<ul><li>Vulgar content</li><li>Prostitution</li><li>Nudity and Pornography</li><li>Abuse</li><li>Child exploitation, child abuse, child grooming</li></ul> | `Sexual` |
4949
| Violence | Violence describes language related to physical actions intended to hurt, injure, damage, or kill someone or something; describes weapons, guns, and related entities. <br><br>This includes, but isn't limited to: <ul><li>Weapons</li><li>Bullying and intimidation</li><li>Terrorist and violent extremism</li><li>Stalking</li></ul> | `Violence` |
@@ -52,7 +52,7 @@ You can use Azure AI Content Safety for many scenarios:
5252
### Severity levels
5353

5454
| Level | Description |
55-
| -- | --|
55+
| --- | ---|
5656
|
5757
Safe |Content might be related to violence, self-harm, sexual, or hate categories but the terms are used in general, journalistic, scientific, medical, and similar professional contexts, which are appropriate for most audiences. |
5858
|Low |Content that expresses prejudiced, judgmental, or opinionated views, includes offensive use of language, stereotyping, use cases exploring a fictional world (for example, gaming, literature) and depictions at low intensity.|
@@ -62,7 +62,7 @@ Safe |Content might be related to violence, self-harm, sexual, or hate categorie
6262
## Other Content Safety features
6363

6464
| Feature | Functionality | Concepts guide |
65-
| :- | :-- | --|
65+
|:--- |:--- | ---|
6666
| [Groundedness detection](/rest/api/contentsafety/text-groundedness-detection-operations/detect-groundedness-options) (preview) | Detects whether the text responses of large language models (LLMs) are grounded in the source materials provided by the users. | [Groundedness detection concepts](/azure/ai-services/content-safety/concepts/groundedness)|
6767
| [Protected material text detection](/rest/api/contentsafety/text-operations/detect-text-protected-material) | Scans AI-generated text for known text content (for example, song lyrics, articles, recipes, selected web content). | [Protected material concepts](/azure/ai-services/content-safety/concepts/protected-material)|
6868
| Custom categories (standard) API (preview) | Lets you create and train your own custom content categories and scan text for matches. | [Custom categories concepts](/azure/ai-services/content-safety/concepts/custom-categories)|

0 commit comments

Comments
 (0)