Skip to main content
The Moderation feature provides a comprehensive suite of capabilities designed to manage and enforce message moderation rules across various types of messages, ensuring your platform remains safe and compliant for all users.

Key Benefits

  • Automated Content Filtering - Automatically detect and handle inappropriate content before it reaches users
  • Customizable Rules - Create rules tailored to your platform’s specific community guidelines
  • Multi-format Support - Moderate text messages, images, and videos
  • Real-time Protection - Messages are moderated instantly as they are sent
  • Manual Review - Users can report content, and moderators can review flagged messages

How It Works

Quick Navigation

Getting Started

1

Access Moderation Settings

Navigate to the CometChat Dashboard and go to Moderation section
2

Create Moderation Rules

Set up rules to automatically detect and handle inappropriate content. See Rules Management
3

Configure Flag Reasons

Customize the reasons users can select when reporting messages. Go to Moderation > Advanced Settings
4

Integrate with Your App

Enable moderation features in your UI Kit or implement using the SDK/REST API

Platform Integration

Moderation is supported across all CometChat platforms. Choose your integration method:

UI Kits

UI Kits provide built-in support for message moderation and the Report Message feature:

Chat SDKs

Implement message flagging directly using CometChat Chat SDKs:

Rules Management

This feature enables you to define and manage a set of moderation rules tailored to address inappropriate messages under various conditions. You can establish specific criteria that determine what constitutes unacceptable behavior or content, such as the use of offensive language, unsafe content, or sharing sensitive information. By customizing these rules, you ensure that the moderation system effectively identifies and manages messages that violate your platform’s standards, thereby maintaining a safe and respectful environment for all users. For more detailed management, refer to the Rules Management section.

Lists Management

This feature allows you to create and manage comprehensive lists of keywords or regex patterns that are used for message moderation. These lists serve as a vital component in identifying and handling inappropriate content. Once created, these keyword lists can be linked to various moderation rules when creating or updating rules, ensuring that the moderation system effectively detects and manages content that violates your standards. For more detailed management, refer to the Lists Management section.

Flagged Messages

This feature enables you to access and manage all messages that require moderation review. You can view a complete list of messages that have been automatically flagged by the rule engine for policy violations or manually reported by users for inappropriate content. For more details, refer to the Flagged Messages section.

Blocked Messages

This feature allows you to retrieve all the violated messages. You can retrieve a comprehensive list of messages that have been blocked due to violations of moderation rules. Additionally, you can perform searches within this list to find specific messages or filter results based on date ranges. For more details, refer to the Blocked Messages section.

Available Moderation Rules

Our platform offers a wide range of moderation rules to help you detect and manage various types of risky, sensitive, or inappropriate content.
NameDescription
Word Pattern MatchIdentifies profane or offensive words using word matching
Contact Details RemovalDetects and removes phone numbers from text
Email DetectionDetects and removes email addresses from messages
Spam Detection (English)Detects spam messages in English
Scam Detection (English)Detects scam or fraudulent text in English
Platform Circumvention (English)Identifies attempts to bypass platform rules
Toxicity Detection (English)Detects toxic or harmful language in text
Explicit or Inappropriate ContentDetects explicit sexual descriptions, graphic violence, or unsuitable text
Privacy and Sensitive InfoIdentifies sensitive personal information shared without consent
Hate and HarassmentDetects hateful or harassing language toward individuals or groups
Self-Harm or Suicidal ContentDetects content suggesting self-harm or suicidal thoughts
Impersonation or FraudDetects deceptive attempts to impersonate individuals or organizations
Violent or Terroristic ThreatsDetects content promoting violence or extremism
Non-Consensual Sexual ContentDetects sexual exploitation, grooming, or non-consensual content
Spam and ScamIdentifies spam, phishing attempts, and fraudulent schemes