azure-ai-contentsafety-ts
Moderate harmful content with Azure AI Content Safety
Protect your platform from harmful user-generated content including hate speech, violence, sexual content, and self-harm material. This skill integrates Azure AI Content Safety REST API to automatically analyze and flag inappropriate text and images with configurable severity thresholds.
下載技能 ZIP
在 Claude 中上傳
前往 設定 → 功能 → 技能 → 上傳技能
開啟並開始使用
測試它
正在使用「azure-ai-contentsafety-ts」。 Analyze a comment for hate speech and violence
預期結果:
Text flagged with severity 4 (Medium) for Hate category. Recommendation: Block or require human review before publication.
正在使用「azure-ai-contentsafety-ts」。 Check an uploaded profile image
預期結果:
Image analysis complete. All categories at severity 0 (Safe). Content approved for publication.
正在使用「azure-ai-contentsafety-ts」。 Moderate chat message with custom blocklist
預期結果:
Message blocked: Contains prohibited term "cheat-code-hack" from gaming-blocklist. Severity 6 (High) for Violence category also detected.
安全審計
低風險Static analyzer produced 70 false positive findings by misidentifying markdown documentation as executable code. The SKILL.md file contains TypeScript code examples using markdown code fences (```), not shell backticks. The skill legitimately requires network access to Azure Content Safety APIs and environment variables for API credentials. No malicious patterns detected after manual review.
低風險問題 (2)
風險因素
🌐 網路存取 (2)
🔑 環境變數 (3)
品質評分
你能建構什麼
Social Platform Content Moderation
Automatically scan user posts, comments, and uploaded images before publication to detect and block harmful content that violates community guidelines.
Educational Forum Safety
Protect students in online learning environments by filtering hate speech, bullying content, and self-harm discussions with appropriate severity thresholds.
E-commerce Review Filtering
Moderate product reviews and seller communications to maintain platform quality standards and prevent abusive or inappropriate content from appearing.
試試這些提示
Analyze this text for harmful content and tell me if it should be allowed: "[INSERT TEXT]"
Check if this image contains harmful or inappropriate content that should be blocked from our platform: [PROVIDE IMAGE PATH OR URL]
Help me create a blocklist for my gaming platform that blocks cheating-related terms, hate speech, and harassment. Add these specific terms: [LIST TERMS]
I need to moderate these 50 user submissions with a maximum allowed severity of 2 (low). Flag anything above this threshold and check against my "gaming-community" blocklist: [LIST CONTENT]
最佳實務
- Always use the isUnexpected() type guard to handle API errors gracefully and prevent crashes from unexpected responses
- Set category-specific severity thresholds based on your community guidelines - hate speech may need stricter limits than general violence
- Maintain audit logs of all moderation decisions with timestamps, severity scores, and action taken for compliance and appeals
避免
- Do not block content solely based on severity 2 (Low) - this may result in excessive false positives and user frustration
- Never store API keys directly in code - always use environment variables or Azure Key Vault for credential management
- Avoid making moderation decisions on empty text or corrupted images - validate input before sending to the API
常見問題
What categories of harmful content does Azure Content Safety detect?
How do severity levels work and what should I block?
Can I add custom words or phrases to block?
What image formats are supported for analysis?
How do I authenticate with the Azure Content Safety API?
What happens if the API is unavailable or returns an error?
開發者詳情
作者
sickn33授權
MIT
儲存庫
https://github.com/sickn33/antigravity-awesome-skills/tree/main/skills/azure-ai-contentsafety-ts引用
main
檔案結構
📄 SKILL.md