BAILU - Lightweight AI-Generated Image Detector

BAILU is a highly efficient deepfake detection model designed to identify AI-generated images from various image generation models. With only 2M parameters (~8MB), it achieves 95.88% overall accuracy by analyzing artifacts/signatures unique to AI generation pipelines.

🌍 Why Open-Source Matters for Deepfake Detection

This model was only possible because companies like Black Forest Labs and Stability AI release their models publicly. Private, closed-source models create detection blind spots—we cannot defend against what we cannot study. We strongly encourage all AI companies to open-source their models to enable:

  • Effective deepfake detection research
  • Transparency in AI development
  • Collaborative safety measures
  • Public trust through verifiable defenses

🎯 Key Features

  • Ultra-Lightweight: 2M parameters, ~8MB model size - runs on CPU or GPU
  • Multi-VAE Detection: Trained to detect artifacts from FLUX.1, FLUX.2, SDXL, and Stable Diffusion 1.5
  • High Accuracy: 95.88% overall accuracy (97.75% AI detection rate, 94.00% real detection)
  • Fast Inference: <10ms per image on modern GPUs
  • Open-Source Advocacy: Built to demonstrate the importance of open-source model transparency

📊 Performance Metrics

Metric Score
Overall Validation Accuracy 95.88% (767/800)
Loss 0.2645

Tested on balanced dataset of 400 AI-generated and 400 real images

🎓 Training Details

  • Hardware: NVIDIA RTX 5090
  • Training Time: ~110 hours
  • Data Augmentation: Random crops, flips, compression, resizing
  • Optimizer: AdamW (lr=1e-4, weight_decay=1e-4)
  • Scheduler: CosineAnnealingLR (T_max=50)
  • Loss: Binary Cross-Entropy with Logits

Detection must keep pace with generation. That requires open access.

⚠️ Important Limitations

  • Not foolproof: Adversarial attacks and new model architectures may evade detection (We plan to train model capable of detecting adversarial attacks later.)
  • No attribution: Cannot identify which specific AI model created an image
  • Temporal degradation: Effectiveness may decrease as new AI models emerge

Disclaimer: This tool is for research and educational purposes. Results should not be used as sole evidence in legal or high-stakes decisions without human expert verification.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Space using LoliRimuru/BAILU 1