Encyclopedia
Content Moderation
Content moderation in AI companions is the safety layer that detects and filters disallowed, harmful, or policy-violating content. It’s different from general SFW/NSFW modes: moderation actively reviews inputs/outputs and applies rules, age gates, or redirections to safer phrasing.
Strong moderation protects users and platforms while preserving conversational flow—for example, refusing a prompt but offering a compliant alternative that keeps the scene moving.

