Rethinking Reverse Distillation for Multi-Modal Anomaly Detection
DOI:
https://doi.org/10.1609/aaai.v38i8.28687Keywords:
DMKM: Anomaly/Outlier Detection, ML: Unsupervised & Self-Supervised LearningAbstract
In recent years, there has been significant progress in employing color images for anomaly detection in industrial scenarios, but it is insufficient for identifying anomalies that are invisible in RGB images alone. As a supplement, introducing extra modalities such as depth and surface normal maps can be helpful to detect these anomalies. To this end, we present a novel Multi-Modal Reverse Distillation (MMRD) paradigm that consists of a frozen multi-modal teacher encoder to generate distillation targets and a learnable student decoder targeting to restore multi-modal representations from the teacher. Specifically, the teacher extracts complementary visual features from different modalities via a siamese architecture and then parameter-freely fuses these information from multiple levels as the targets of distillation. For the student, it learns modality-related priors from the teacher representations of normal training data and performs interaction between them to form multi-modal representations for target reconstruction. Extensive experiments show that our MMRD outperforms recent state-of-the-art methods on both anomaly detection and localization on MVTec-3D AD and Eyecandies benchmarks. Codes will be available upon acceptance.Downloads
Published
2024-03-24
How to Cite
Gu, Z., Zhang, J., Liu, L., Chen, X., Peng, J., Gan, Z., Jiang, G., Shu, A., Wang, Y., & Ma, L. (2024). Rethinking Reverse Distillation for Multi-Modal Anomaly Detection. Proceedings of the AAAI Conference on Artificial Intelligence, 38(8), 8445-8453. https://doi.org/10.1609/aaai.v38i8.28687
Issue
Section
AAAI Technical Track on Data Mining & Knowledge Management