iFAN: Image-Instance Full Alignment Networks for Adaptive Object Detection

Authors

  • Chenfan Zhuang Malong Technologies, Shenzhen, China
  • Xintong Han Malong Technologies, Shenzhen, China
  • Weilin Huang Malong Technologies, Shenzhen, China
  • Matthew Scott Malong Technologies, Shenzhen, China

DOI:

https://doi.org/10.1609/aaai.v34i07.7015

Abstract

Training an object detector on a data-rich domain and applying it to a data-poor one with limited performance drop is highly attractive in industry, because it saves huge annotation cost. Recent research on unsupervised domain adaptive object detection has verified that aligning data distributions between source and target images through adversarial learning is very useful. The key is when, where and how to use it to achieve best practice. We propose Image-Instance Full Alignment Networks (iFAN) to tackle this problem by precisely aligning feature distributions on both image and instance levels: 1) Image-level alignment: multi-scale features are roughly aligned by training adversarial domain classifiers in a hierarchically-nested fashion. 2) Full instance-level alignment: deep semantic information and elaborate instance representations are fully exploited to establish a strong relationship among categories and domains. Establishing these correlations is formulated as a metric learning problem by carefully constructing instance pairs. Above-mentioned adaptations can be integrated into an object detector (e.g. Faster R-CNN), resulting in an end-to-end trainable framework where multiple alignments can work collaboratively in a coarse-to-fine manner. In two domain adaptation tasks: synthetic-to-real (SIM10K → Cityscapes) and normal-to-foggy weather (Cityscapes → Foggy Cityscapes), iFAN outperforms the state-of-the-art methods with a boost of 10%+ AP over the source-only baseline.

Downloads

Published

2020-04-03

How to Cite

Zhuang, C., Han, X., Huang, W., & Scott, M. (2020). iFAN: Image-Instance Full Alignment Networks for Adaptive Object Detection. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07), 13122-13129. https://doi.org/10.1609/aaai.v34i07.7015

Issue

Section

AAAI Technical Track: Vision