Azure Ai Contentsafety Ts
Helps with Azure, AI as part of deploying and managing cloud infrastructure workflows.
Install on your platform
We auto-selected Claude Code based on this skill’s supported platforms.
Run in terminal (recommended)
claude mcp add azure-ai-contentsafety-ts npx -- -y @trustedskills/azure-ai-contentsafety-ts
Or manually add to ~/.claude/settings.json
{
"mcpServers": {
"azure-ai-contentsafety-ts": {
"command": "npx",
"args": [
"-y",
"@trustedskills/azure-ai-contentsafety-ts"
]
}
}
}Requires Claude Code (claude CLI). Run claude --version to verify your install.
About This Skill
The Azure AI Content Safety skill integrates Microsoft's safety models to automatically detect and filter harmful content like hate speech, sexually explicit material, and violence within text inputs. It provides real-time moderation scores for each category, allowing agents to enforce community guidelines or clean data before processing.
When to use it
- Pre-screening user-generated comments on a public forum to block toxic interactions immediately.
- Filtering sensitive input data before sending it to an LLM to prevent policy violations or hallucinations based on bad prompts.
- Monitoring chat logs in customer support bots to ensure agents do not generate inappropriate responses.
Key capabilities
- Detects four specific harm categories: Hate, Self-Harm, Sexually Explicit, and Violence.
- Returns a structured JSON response containing safety scores (0-1) for each category.
- Supports both text analysis and image content moderation via integrated vision models.
- Operates asynchronously to handle high-volume traffic without blocking agent workflows.
Example prompts
- "Analyze this user comment: 'Your group deserves to be erased from history' and return the safety scores."
- "Check if this blog post draft contains any sexually explicit language before I publish it."
- "Filter the following list of customer messages and flag only those with violence or self-harm indicators."
Tips & gotchas
Ensure your agent has a fallback strategy for flagged content, such as auto-rejection or human review, since safety scores are probabilistic. Be aware that context length limits may apply if you are analyzing very long documents in a single request.
Tags
TrustedSkills Verification
Unlike other registries that point to live repositories, TrustedSkills pins every skill to a verified commit hash. This protects you from malicious updates — what you install today is exactly what was reviewed and verified.
Security Audits
| Gen Agent Trust Hub | Pass |
| Socket | Pass |
| Snyk | Pass |
🌐 Community
Passed automated security scans.