Skip to main content

Documentation Index

Fetch the complete documentation index at: https://volvox-mintlify-seo-audit-1777851731.mintlify.app/llms.txt

Use this file to discover all available pages before exploring further.

Moderation (Active Sentry)

Active Sentry is Volvox.Bot’s moderation engine. It runs automatically, and you can configure it from the dashboard.

What it does

  • Auto-moderation — Automatically detect and act on spam, slurs, invite links, and more
  • Content Safety — Score incoming messages for toxicity, spam, harassment, hate speech, sexual content, violence, and self-harm. The bot runs your configured actions when a score meets its threshold
  • Warnings — Issue warnings that accumulate and trigger escalating actions
  • Bans & Kicks — Full ban and kick support with reason tracking
  • Timeouts — Discord native timeouts with configurable durations
  • Cases — Every moderation action creates a case with full context

How to use

  1. Open the dashboard → Moderation for your server
  2. Review active cases, filter by type or member
  3. Use Config → Moderation & Safety → Moderation to set up classic auto-mod rules
  4. Use Config → Moderation & Safety → Content Safety to choose the detection model, thresholds, and actions
  5. Use Config → AI & Automation → Triage to choose the supported classifier and response models for AI triage
Configure escalation rules so that repeated warnings automatically escalate to timeouts or bans.

Configuration

All moderation settings are in Config → Moderation & Safety → Moderation:
SettingDescription
Auto-mod enabledToggle automatic moderation
Spam thresholdMessages per minute before flagging
Word filterCustom list of blocked words
Invite blockingBlock Discord invite links
Log channelWhere the bot posts moderation actions
AI moderation settings are in Config → Moderation & Safety → Content Safety:
SettingDescription
Detection modelThe supported provider/model used to score incoming messages
Incident report channelWhere the bot posts flagged AI moderation reports
Instant enforcementGlobal auto-delete for messages that trigger AI moderation
Sensitivity thresholdsPer-category score required before action runs
Response actionsChoose any combination of flag, delete, warn, timeout, kick, or ban per category. Leave a category empty to skip category-specific responses. warn creates a warning record, sends the warning DM when enabled, and checks your escalation rules.
The bot records every triggered AI moderation response in the audit log as ai_automod.<action> for each action that actually runs. The bot only records ai_automod.none when a threshold is hit and no action runs. Global instant enforcement (autoDelete) or an incident report channel (flagChannelId) can still produce delete or flag actions, even when you haven’t chosen a category-specific response action. Triage model settings are in Config → AI & Automation → Triage. The classifier and response engines use the same supported model dropdown as Content Safety; if saved config contains an invalid or unsupported model string, the bot logs a warning, tries supported configured values in the fixed precedence order, and uses the default AI model if none are supported. Changes apply instantly — no bot restart needed.