Few-Shot Object Detection via Variational Feature Aggregation

Authors

  • Jiaming Han Wuhan University
  • Yuqiang Ren Tencent
  • Jian Ding Wuhan University
  • Ke Yan Tencent
  • Gui-Song Xia Wuhan University

DOI:

https://doi.org/10.1609/aaai.v37i1.25153

Keywords:

CV: Object Detection & Categorization

Abstract

As few-shot object detectors are often trained with abundant base samples and fine-tuned on few-shot novel examples, the learned models are usually biased to base classes and sensitive to the variance of novel examples. To address this issue, we propose a meta-learning framework with two novel feature aggregation schemes. More precisely, we first present a Class-Agnostic Aggregation (CAA) method, where the query and support features can be aggregated regardless of their categories. The interactions between different classes encourage class-agnostic representations and reduce confusion between base and novel classes. Based on the CAA, we then propose a Variational Feature Aggregation (VFA) method, which encodes support examples into class-level support features for robust feature aggregation. We use a variational autoencoder to estimate class distributions and sample variational features from distributions that are more robust to the variance of support examples. Besides, we decouple classification and regression tasks so that VFA is performed on the classification branch without affecting object localization. Extensive experiments on PASCAL VOC and COCO demonstrate that our method significantly outperforms a strong baseline (up to 16%) and previous state-of-the-art methods (4% in average).

Downloads

Published

2023-06-26

How to Cite

Han, J., Ren, Y., Ding, J., Yan, K., & Xia, G.-S. (2023). Few-Shot Object Detection via Variational Feature Aggregation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(1), 755-763. https://doi.org/10.1609/aaai.v37i1.25153

Issue

Section

AAAI Technical Track on Computer Vision I