Robust Feature Rectification of Pretrained Vision Models for Object Recognition

Authors

  • Shengchao Zhou NLPR, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences
  • Gaofeng Meng NLPR, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences CAIR, HK Institute of Science and Innovation, Chinese Academy of Sciences
  • Zhaoxiang Zhang NLPR, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences CAIR, HK Institute of Science and Innovation, Chinese Academy of Sciences
  • Richard Yi Da Xu FSC1209, Kowloon Tong Campus, Hong Kong Baptist University
  • Shiming Xiang NLPR, Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v37i3.25492

Keywords:

CV: Object Detection & Categorization, CV: Applications

Abstract

Pretrained vision models for object recognition often suffer a dramatic performance drop with degradations unseen during training. In this work, we propose a RObust FEature Rectification module (ROFER) to improve the performance of pretrained models against degradations. Specifically, ROFER first estimates the type and intensity of the degradation that corrupts the image features. Then, it leverages a Fully Convolutional Network (FCN) to rectify the features from the degradation by pulling them back to clear features. ROFER is a general-purpose module that can address various degradations simultaneously, including blur, noise, and low contrast. Besides, it can be plugged into pretrained models seamlessly to rectify the degraded features without retraining the whole model. Furthermore, ROFER can be easily extended to address composite degradations by adopting a beam search algorithm to find the composition order. Evaluations on CIFAR-10 and Tiny-ImageNet demonstrate that the accuracy of ROFER is 5% higher than that of SOTA methods on different degradations. With respect to composite degradations, ROFER improves the accuracy of a pretrained CNN by 10% and 6% on CIFAR-10 and Tiny-ImageNet respectively.

Downloads

Published

2023-06-26

How to Cite

Zhou, S., Meng, G., Zhang, Z., Xu, R. Y. D., & Xiang, S. (2023). Robust Feature Rectification of Pretrained Vision Models for Object Recognition. Proceedings of the AAAI Conference on Artificial Intelligence, 37(3), 3796-3804. https://doi.org/10.1609/aaai.v37i3.25492

Issue

Section

AAAI Technical Track on Computer Vision III