MTLDesc: Looking Wider to Describe Better

Authors

  • Changwei Wang Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence University of Chinese Academy of Sciences
  • Rongtao Xu Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence University of Chinese Academy of Sciences
  • Yuyang Zhang Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence University of Chinese Academy of Sciences
  • Shibiao Xu School of Artificial Intelligence, Beijing University of Posts and Telecommunications
  • Weiliang Meng Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence Zhejiang Lab University of Chinese Academy of Sciences
  • Bin Fan University of Science and Technology Beijing
  • Xiaopeng Zhang Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence University of Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v36i2.20138

Keywords:

Computer Vision (CV), Machine Learning (ML)

Abstract

Limited by the locality of convolutional neural networks, most existing local features description methods only learn local descriptors with local information and lack awareness of global and surrounding spatial context. In this work, we focus on making local descriptors ``look wider to describe better'' by learning local Descriptors with More Than Local information (MTLDesc). Specifically, we resort to context augmentation and spatial attention mechanism to make the descriptors obtain non-local awareness. First, Adaptive Global Context Augmented Module and Diverse Local Context Augmented Module are proposed to construct robust local descriptors with context information from global to local. Second, we propose the Consistent Attention Weighted Triplet Loss to leverage spatial attention awareness in both optimization and matching of local descriptors. Third, Local Features Detection with Feature Pyramid is proposed to obtain more stable and accurate keypoints localization. With the above innovations, the performance of the proposed MTLDesc significantly surpasses the current state-of-the-art local descriptors on HPatches, Aachen Day-Night localization and InLoc indoor localization benchmarks. Our code is available at https://github.com/vignywang/MTLDesc.

Downloads

Published

2022-06-28

How to Cite

Wang, C., Xu, R., Zhang, Y., Xu, S., Meng, W., Fan, B., & Zhang, X. (2022). MTLDesc: Looking Wider to Describe Better. Proceedings of the AAAI Conference on Artificial Intelligence, 36(2), 2388-2396. https://doi.org/10.1609/aaai.v36i2.20138

Issue

Section

AAAI Technical Track on Computer Vision II