content-moderation
Moderate text, images, and video using Vettly's content moderation API via MCP server.
Installation
npx clawhub@latest install content-moderationView the full skill documentation and source below.
Documentation
Content Moderation
Moderate user-generated content using Vettly's AI-powered content moderation API. This skill uses the @vettly/mcp MCP server to check text, images, and video against configurable moderation policies with auditable decisions.
Setup
Add the @vettly/mcp MCP server to your configuration:
{
"mcpServers": {
"vettly": {
"command": "npx",
"args": ["-y", "@vettly/mcp"],
"env": {
"VETTLY_API_KEY": "your-api-key"
}
}
}
}
Get an API key at [vettly.dev]().
Available Tools
moderate_content
Check text, image, or video content against a Vettly moderation policy. Returns a safety assessment with category scores, the action taken, provider used, latency, and cost.
Parameters:
content(required) - The content to moderate (text string, or URL for images/video)policyId(required) - The policy ID to use for moderationcontentType(optional, default:text) - Type of content:text,image, orvideo
validate_policy
Validate a Vettly policy YAML without saving it. Returns validation results with any syntax or configuration errors. Use this to test policy changes before deploying them.
Parameters:
yamlContent(required) - The YAML policy content to validate
list_policies
List all moderation policies available in your Vettly account. Takes no parameters. Use this to discover available policy IDs before moderating content.
get_usage_stats
Get usage statistics for your Vettly account including request counts, costs, and moderation outcomes.
Parameters:
days(optional, default:30) - Number of days to include in statistics (1-365)
get_recent_decisions
Get recent moderation decisions with optional filtering by outcome, content type, or policy.
Parameters:
limit(optional, default:10) - Number of decisions to return (1-50)flagged(optional) - Filter to only flagged content (true) or safe content (false)policyId(optional) - Filter by specific policy IDcontentType(optional) - Filter by content type:text,image, orvideo
When to Use
- Moderate user-generated content (comments, posts, uploads) before publishing
- Test and validate moderation policy YAML configs during development
- Audit recent moderation decisions to review flagged content
- Monitor moderation costs and usage across your account
- Compare moderation results across different policies
Examples
Moderate a user comment
Moderate this user comment for my community forum policy:
"I hate this product, it's the worst thing I've ever used and the developers should be ashamed"
Call list_policies to find available policies, then moderate_content with the appropriate policy ID and return the safety assessment.
Validate a policy before deploying
Validate this moderation policy YAML:
categories:
- name: toxicity
threshold: 0.8
action: flag
- name: spam
threshold: 0.6
action: block
Call validate_policy and report any syntax or configuration errors.
Review recent flagged content
Show me all flagged content from the last week
Call get_recent_decisions with flagged: true to retrieve recent moderation decisions that were flagged.
Tips
- Always call
list_policiesfirst if you don't know which policy ID to use - Use
validate_policyto test policy changes before deploying to production - Use
get_usage_statsto monitor costs and catch unexpected spikes - Filter
get_recent_decisionsbycontentTypeorpolicyIdto narrow results - For image and video moderation, pass the content URL rather than raw data