Spaces:
Running
Running
Update: Edited & AI-Generated Content Detection β Project Plan
π Phase 1: Rule-Based Image Detection (In Progress)
We're implementing three core techniques to individually flag edited or AI-generated images:
- ELA (Error Level Analysis): Highlights inconsistencies via JPEG recompression.
- FFT (Frequency Analysis): Uses 2D Fourier Transform to detect unnatural image frequency patterns.
- Metadata Analysis: Parses EXIF data to catch clues like editing software tags.
These give us visual + interpretable results for each image, and currently offer ~60β70% accuracy on typical AI-edited content.
Phase 2: AI vs Human Detection System (Coming Soon)
Goal: Build an AI model that classifies whether content is AI- or human-made β initially focusing on images, and later expanding to text.
Data Strategy:
- Scraping large volumes of recent AI-gen images (e.g. SDXL, Gibbli, MidJourney).
- Balancing with high-quality human images.
Model Plan:
- Use ELA, FFT, and metadata as feature extractors.
- Feed these into a CNN or ensemble model.
- Later, unify into a full web-based platform (upload β get AI/human probability).