kojott/content-moderation-vision

AI-powered content moderation for images using MiniCPM-V-2.6 - analyzes visual content and returns structured safety scores with detailed classifications

Public
30 runs

Content Moderation Vision AI

An advanced content moderation system powered by MiniCPM-V-2.6 that analyzes images for inappropriate content and returns structured safety assessments.

Features

  • Intelligent Content Analysis: Uses state-of-the-art vision-language model MiniCPM-V-2.6
  • Structured JSON Responses: Returns detailed safety scores (0-4) with classifications
  • Comprehensive Categories: Detects SAFE, ADULT_THEMES, NSFW, HATE_CONTENT, VIOLENCE, HARMFUL content
  • Custom Prompts: Supports both automated moderation and custom image analysis
  • GPU Optimized: Efficient CUDA acceleration with automatic memory management

Content Classifications

  • Score 0 (SAFE): Completely safe content - no concerns for any audience
  • Score 1 (ADULT_THEMES): Minor adult themes, revealing but clothed content
  • Score 2 (NSFW): Moderate concerns requiring age-appropriate context
  • Score 3 (INAPPROPRIATE): Explicit content with visible private parts, hate symbols
  • Score 4 (HARMFUL): Severely harmful content requiring immediate action

API Usage

Content Moderation Mode (Default)

import replicate

output = replicate.run(
    "kojott/content-moderation-vision",
    input={"image": "https://example.com/image.jpg"}
)
print(output)  # Returns structured JSON

Custom Prompt Mode

output = replicate.run(
    "kojott/content-moderation-vision",
    input={
        "image": "https://example.com/image.jpg",
        "prompt": "Describe what you see in this image"
    }
)

Response Format

{
  "score": 0,
  "classification": "SAFE",
  "description": "A beautiful landscape photo showing mountains and trees",
  "concerns": [],
  "safe_for_children": true,
  "requires_restriction": false,
  "admin_notes": "Natural landscape content, completely appropriate"
}

Parameters

  • image (required): Image file to analyze
  • prompt (optional): Custom analysis prompt. If empty, uses content moderation mode
  • temperature (0.0-1.0): Controls randomness in generation (default: 0.1)
  • top_p (0.0-1.0): Nucleus sampling parameter (default: 0.9)

Technical Details

  • Base Model: MiniCPM-V-2.6 (OpenBMB)
  • Framework: PyTorch with CUDA acceleration
  • Memory: Optimized for GPU efficiency with automatic cleanup
  • Response Time: Typically 2-5 seconds per image
  • Supported Formats: JPEG, PNG, WebP, and other PIL-compatible formats

Use Cases

  • Social Media Platforms: Automated content screening
  • E-commerce: Product image validation
  • Educational Platforms: Child-safe content verification
  • Community Forums: User-generated content moderation
  • Dating Apps: Profile photo screening

Model Performance

This model balances accuracy with practical deployment needs, avoiding over-censorship while effectively identifying genuinely harmful content. It’s designed for real-world applications where human-centered judgment is essential.

Built with reliability and production deployment in mind, featuring comprehensive error handling and fallback responses.