Beyond Semantic Features: Pixel-level Mapping for Generalized AI-Generated Image Detection
DOI:
https://doi.org/10.1609/aaai.v40i42.40927Abstract
The rapid evolution of generative technologies necessitates reliable methods for detecting AI-generated images. A critical limitation of current detectors is their failure to generalize to images from unseen generative models, as they often overfit to source-specific semantic cues rather than learning universal generative artifacts. To overcome this, we introduce a simple yet remarkably effective pixel-level mapping pre-processing step to disrupt the pixel value distribution of images and break the fragile, non-essential semantic patterns that detectors commonly exploit as shortcuts. This forces the detector to focus on more fundamental and generalizable high-frequency traces inherent to the image generation process. Through comprehensive experiments on GAN and diffusion-based generators, we show that our approach significantly boosts the cross-generator performance of state-of-the-art detectors. Extensive analysis further verifies our hypothesis that the disruption of semantic cues is the key to generalization.Downloads
Published
2026-03-14
How to Cite
Zhou, C., Wang, J., Li, Y., Li, L., Cao, J., & Tang, S. (2026). Beyond Semantic Features: Pixel-level Mapping for Generalized AI-Generated Image Detection. Proceedings of the AAAI Conference on Artificial Intelligence, 40(42), 36101–36109. https://doi.org/10.1609/aaai.v40i42.40927
Issue
Section
AAAI Technical Track on Philosophy and Ethics of AI