Detect and block violence, hate, sexual, and self-harm content. Configure severity thresholds for your specific use case, and adhere to your responsible AI policies.
Create unique content filters tailored to your requirements using custom categories. Quickly train a new custom category by providing examples of content you need to block.
Safeguard your AI applications against prompt injection attacks and jailbreak attempts. Identify and mitigate both direct and indirect threats with prompt shields.
Detect and mitigate harmful content in user-generated and AI-generated inputs and outputs—including text, images, and mixed media—all with Azure AI Content Safety.
Use cases
Safeguard your AI applications
Develop guardrails for generative AI applications
Use AI to monitor human-generated content
Security
Embedded security and compliance
34,000
Full-time equivalent engineers dedicated to security initiatives at Microsoft.
15,000
Partners with specialized security expertise.
>100
Compliance certifications, including over 50 specific to global regions and countries.
Use Azure AI Content Safety with other Azure AI products to create advanced guardrails for generative AI or to develop comprehensive solutions with built-in responsible AI tooling.
Azure OpenAI
Build your own copilot and generative AI applications with cutting-edge language and vision models.
See how customers are protecting their applications with Azure AI Content Safety
"The South Australia Department for Education launched a generative AI-powered educational chatbot to safely incorporate AI technology into classrooms."
"ASOS is using generative AI to elevate their customer experience, enabling users to discover new fashion looks effortlessly. Azure AI Content Safety helps to ensure top-quality interactions and outputs."
"Unity developed Muse Chat to enhance their game creation process. To ensure responsible use, Unity uses Azure OpenAI content filters powered by Azure AI Content Safety."
"IWill Therapy used generative AI to create a Hindi-speaking chatbot that provides cognitive behavioral therapy across India. The solution employs Azure AI Content Safety to detect and filter potentially harmful content."
Get expert insights on how customizing AI models is helping the world's leading companies drive more value in this new report from Microsoft and MIT Technology Review Insights.
Content Safety models have been specifically trained and tested in the following languages: English, German, Spanish, Japanese, French, Italian, Portuguese, and Chinese. The service can work in other languages as well, but the quality might vary. In all cases, you should do your own testing to ensure that it works for your application.
Custom categories currently work well in English only. You can use other languages with your own dataset, but the quality might vary.
The system monitors across four harm categories: hate, sexual, violence, and self-harm.
Yes, you can adjust severity thresholds for each harm category filter.
Yes, you can use the Azure AI Content Safety custom categories API to create your own content filters. By providing examples, you can train the filter to detect and block undesired content specific to your defined custom categories.
Prompt shields enhance the security of generative AI systems by defending against prompt injection attacks:
Direct prompt attacks (jailbreaks): Users try to manipulate the AI system and bypass safety protocols by creating prompts that attempt to alter system rules or trick the model into executing restricted actions.
Indirect attacks: Third-party content, like documents or emails, contains hidden instructions to exploit the AI system, such as embedded commands an AI might unknowingly execute.
Groundedness detection identifies and corrects the ungrounded outputs of generative AI models, ensuring they’re based on provided source materials. This helps to prevent the generation of fabricated or false information. Using a custom language model, groundedness detection evaluates claims against source data and mitigates AI hallucinations.
Protected material detection for text identifies and blocks known text content, such as lyrics, articles, recipes, and selected web content, from appearing in AI-generated outputs.
Protected material detection for code detects and prevents the output of known code. It checks for matches against public source code in GitHub repositories. Additionally, the code referencing capability powered by GitHub Copilot enables developers to locate repositories for exploring and discovering relevant code.
The content filtering system inside Azure OpenAI is powered by Azure AI Content Safety. It’s designed to detect and prevent the output of harmful content in both input prompts and output completions. It works alongside core models, including GPT and DALL-E.