|
| 1 | +--- |
| 2 | +title: "Mitigate false results in Azure AI Content Safety" |
| 3 | +titleSuffix: Azure AI services |
| 4 | +description: Learn techniques to improve the performance of Azure AI Content Safety models by handling false positives and false negatives. |
| 5 | +#services: cognitive-services |
| 6 | +author: PatrickFarley |
| 7 | +manager: nitinme |
| 8 | +ms.service: azure-ai-content-safety |
| 9 | +ms.topic: how-to |
| 10 | +ms.date: 09/18/2024 |
| 11 | +ms.author: pafarley |
| 12 | +#customer intent: As a user, I want to improve the performance of Azure AI Content Safety so that I can ensure accurate content moderation. |
| 13 | +--- |
| 14 | + |
| 15 | +# Mitigate false results in Azure AI Content Safety |
| 16 | + |
| 17 | +This guide provides a step-by-step process for handling false positives and false negatives from Azure AI Content Safety models. |
| 18 | + |
| 19 | +False positives are when the system incorrectly flags non-harmful content as harmful; false negatives are when harmful content is not flagged as harmful. Address these instances to ensure the integrity and reliability of your content moderation process, including responsible generative AI deployment. |
| 20 | + |
| 21 | +## Prerequisites |
| 22 | + |
| 23 | +* An Azure subscription - [Create one for free](https://azure.microsoft.com/free/cognitive-services/) |
| 24 | +* Once you have your Azure subscription, <a href="https://aka.ms/acs-create" title="Create a Content Safety resource" target="_blank">create a Content Safety resource </a> in the Azure portal to get your key and endpoint. Enter a unique name for your resource, select your subscription, and select a resource group, supported region (see [Region availability](/azure/ai-services/content-safety/overview#region-availability)), and supported pricing tier. Then select **Create**. |
| 25 | + |
| 26 | +## Review and verification |
| 27 | + |
| 28 | +Conduct an initial assessment to confirm that the flagged content is really a false positive or false negative. This can involve: |
| 29 | +- Checking the context of the flagged content. |
| 30 | +- Comparing the flagged content against the content safety risk categories and severity definitions: |
| 31 | + - If you're using content safety in Azure OpenAI, see the [Azure OpenAI content filtering doc](/azure/ai-services/openai/concepts/content-filter). |
| 32 | + - If you're using the Azure AI Content Safety standalone API, see the [Harm categories doc](/azure/ai-services/content-safety/concepts/harm-categories?tabs=warning) or the [Prompt Shields doc](/azure/ai-services/content-safety/concepts/jailbreak-detection), depending on which API you're using. |
| 33 | + |
| 34 | +## Customize your severity settings |
| 35 | + |
| 36 | +If your assessment confirms that you found a false positive or false negative, you can try customizing your severity settings to mitigate the issue. The settings depend on which platform you're using. |
| 37 | + |
| 38 | +#### [Content Safety standalone API](#tab/standalone-api) |
| 39 | + |
| 40 | +If you're using the Azure AI Content Safety standalone API directly, try experimenting by setting the severity threshold at different levels for [harm categories](/azure/ai-services/content-safety/concepts/harm-categories?tabs=definitions) based on API output. Alternatively, if you prefer the no-code approach, you can try out those settings in [Content Safety Studio](https://contentsafety.cognitive.azure.com/) or Azure AI Studio’s [Content Safety page](https://ai.azure.com/explore/contentsafety). Instructions can be found [here](/azure/ai-studio/quickstarts/content-safety?tabs=moderate-text-content). |
| 41 | + |
| 42 | +In addition to adjusting the severity levels for false negatives, you can also use blocklists. More information on using blocklists for text moderation can be found in [Use blocklists for text moderation](/azure/ai-services/content-safety/how-to/use-blocklist?tabs=windows%2Crest). |
| 43 | + |
| 44 | + |
| 45 | +#### [Azure OpenAI](#tab/azure-openai-studio) |
| 46 | + |
| 47 | +Read the [Configurability](/en-us/azure/ai-services/openai/concepts/content-filter?tabs=warning%2Cuser-prompt%2Cpython-new#configurability-preview) documentation, as some content filtering configurations may require approval through the process mentioned there. |
| 48 | + |
| 49 | +Follow the steps in the documentation to update configurations to handle false positives or negatives: [How to use content filters (preview) with Azure OpenAI Service](/azure/ai-services/openai/how-to/content-filters). |
| 50 | + |
| 51 | +In addition to adjusting the severity levels for false negatives, you can also use blocklists. Detailed instruction can be found in [How to use blocklists with Azure OpenAI Service](/azure/ai-services/openai/how-to/use-blocklists). |
| 52 | + |
| 53 | +#### [Azure AI Studio](#tab/azure-ai-studio) |
| 54 | + |
| 55 | +Read the [Configurability](/azure/ai-studio/concepts/content-filtering#configurability-preview) documentation, as some content filtering configurations may require approval through the process mentioned there. |
| 56 | + |
| 57 | +Follow the steps in the documentation to update configurations to handle false positives or negatives: [Azure AI Studio content filtering](/azure/ai-studio/concepts/content-filtering#create-a-content-filter). |
| 58 | + |
| 59 | +In addition to adjusting the severity levels for false negatives, you can also use blocklists. Detailed instruction can be found in [Azure AI Studio content filtering](/azure/ai-studio/concepts/content-filtering#use-a-blocklist-as-a-filter). |
| 60 | + |
| 61 | +--- |
| 62 | + |
| 63 | +## Create a custom category based on your own RAI policy |
| 64 | + |
| 65 | +Sometimes you might need to create a custom category to ensure the filtering aligns with your specific Responsible AI policy, as prebuilt categories or content filtering may not be enough. |
| 66 | + |
| 67 | +Refer to the [Custom categories documentation](/azure/ai-services/content-safety/concepts/custom-categories) to build your own categories with the Azure AI Content Safety API. |
| 68 | + |
| 69 | +## Document issues and send feedback to Azure |
| 70 | + |
| 71 | +If, after you’ve tried all the steps mentioned above, Azure AI Content Safety still can't resolve the false positives or negatives, there is likely a policy definition or model issue that needs further attention. |
| 72 | + |
| 73 | +Document the details of the false positives and/or false negatives by providing the following information to the [Content safety support team ](mailto:[email protected]): |
| 74 | +- Description of the flagged content. |
| 75 | +- Context in which the content was posted. |
| 76 | +- Reason given by Azure AI Content Safety for the flagging (if positive). |
| 77 | +- Explanation of why the content is a false positive or negative. |
| 78 | +- Any adjustments already attempted by adjusting severity settings or using custom categories. |
| 79 | +- Screenshots or logs of the flagged content and system responses. |
| 80 | + |
| 81 | +This documentation helps in escalating the issue to the appropriate teams for resolution. |
| 82 | + |
| 83 | +## Related content |
| 84 | + |
| 85 | +- [Azure AI Content Safety overview](/azure/ai-services/content-safety/overview) |
| 86 | +- [Harm categories](/azure/ai-services/content-safety/concepts/harm-categories?tabs=warning) |
0 commit comments