Documentation Index
Fetch the complete documentation index at: https://volvox-mintlify-seo-audit-1777851731.mintlify.app/llms.txt
Use this file to discover all available pages before exploring further.
Moderation (Active Sentry)
Active Sentry is Volvox.Bot’s moderation engine. It runs automatically, and you can configure it from the dashboard.What it does
- Auto-moderation — Automatically detect and act on spam, slurs, invite links, and more
- Content Safety — Score incoming messages for toxicity, spam, harassment, hate speech, sexual content, violence, and self-harm. The bot runs your configured actions when a score meets its threshold
- Warnings — Issue warnings that accumulate and trigger escalating actions
- Bans & Kicks — Full ban and kick support with reason tracking
- Timeouts — Discord native timeouts with configurable durations
- Cases — Every moderation action creates a case with full context
How to use
- Open the dashboard → Moderation for your server
- Review active cases, filter by type or member
- Use Config → Moderation & Safety → Moderation to set up classic auto-mod rules
- Use Config → Moderation & Safety → Content Safety to choose the detection model, thresholds, and actions
- Use Config → AI & Automation → Triage to choose the supported classifier and response models for AI triage
Configure escalation rules so that repeated warnings automatically escalate to timeouts or bans.
Configuration
All moderation settings are in Config → Moderation & Safety → Moderation:| Setting | Description |
|---|---|
| Auto-mod enabled | Toggle automatic moderation |
| Spam threshold | Messages per minute before flagging |
| Word filter | Custom list of blocked words |
| Invite blocking | Block Discord invite links |
| Log channel | Where the bot posts moderation actions |
| Setting | Description |
|---|---|
| Detection model | The supported provider/model used to score incoming messages |
| Incident report channel | Where the bot posts flagged AI moderation reports |
| Instant enforcement | Global auto-delete for messages that trigger AI moderation |
| Sensitivity thresholds | Per-category score required before action runs |
| Response actions | Choose any combination of flag, delete, warn, timeout, kick, or ban per category. Leave a category empty to skip category-specific responses. warn creates a warning record, sends the warning DM when enabled, and checks your escalation rules. |
ai_automod.<action> for each action that actually runs. The bot only records ai_automod.none when a threshold is hit and no action runs. Global instant enforcement (autoDelete) or an incident report channel (flagChannelId) can still produce delete or flag actions, even when you haven’t chosen a category-specific response action.
Triage model settings are in Config → AI & Automation → Triage. The classifier and response engines use the same supported model dropdown as Content Safety; if saved config contains an invalid or unsupported model string, the bot logs a warning, tries supported configured values in the fixed precedence order, and uses the default AI model if none are supported.
Changes apply instantly — no bot restart needed.