AI Slop Detector: Multi-Agent Quality Assessment System

Published:

Overview

A multi-agent system that identifies “AI slop”—low-quality, incoherent model outputs that evade basic quality filters. Built for the AWS Small Language Build Day Hackathon, the system uses four specialized small language models (1-3B parameters) with LoRA adapters to serve as an early warning mechanism before expensive moderation stages.

Technical Approach

  • Multi-Agent Architecture: 4 fine-tuned SLMs targeting specific quality signals
    • Genericity Detector: Identifies vague, semantically shallow language
    • Substance Analyzer: Measures information density and topical grounding
    • Style Analyzer: Captures synthetic text patterns and lexical regularity
    • Repetition Detector: Flags templated artifacts and autoregressive repetition
  • Models: Qwen 2.5 (1.5B) + 4 LoRA adapters (~5MB each)
  • Ensemble Voting: Integrates judgments across evaluators
  • Compute: AWS Trainium (trn1.2xlarge) with Neuron SDK
  • Stack: PyTorch, Transformers, PEFT

Key Advantages

  • Interpretability: Specialized evaluators expose decision drivers
  • Efficiency: Real-time evaluation capability, ~10× cheaper than large-model inference
  • Generalization: Tests across model families (GPT vs Qwen)
  • Transparency: Trustable monitoring pipeline

What I Learned

  • Designing multi-agent architectures where specialized models collaborate on complex classification tasks
  • Fine-tuning small language models efficiently using LoRA adapters for targeted quality assessment
  • Working with AWS Trainium accelerators and the Neuron SDK for cost-effective training
  • Balancing model size, inference speed, and accuracy in production-oriented quality detection systems
  • Implementing ensemble voting strategies to combine multiple evaluation signals into robust decisions

Status: CompletedTimeline: Nov 2025