Point-PEFT: Parameter-Efficient Fine-Tuning for 3D Pre-trained Models

Authors

  • Yiwen Tang Northwestern Polytechnical University Shanghai AI Laboratory
  • Ray Zhang Shanghai AI Laboratory
  • Zoey Guo Shanghai AI Laboratory
  • Xianzheng Ma Shanghai AI Laboratory
  • Bin Zhao Northwestern Polytechnical University Shanghai AI Laboratory
  • Zhigang Wang Shanghai AI Laboratory
  • Dong Wang Shanghai AI Laboratory
  • Xuelong Li Northwestern Polytechnical University Shanghai AI Laboratory

DOI:

https://doi.org/10.1609/aaai.v38i6.28323

Keywords:

CV: 3D Computer Vision, ML: Transfer, Domain Adaptation, Multi-Task Learning

Abstract

The popularity of pre-trained large models has revolutionized downstream tasks across diverse fields, such as language, vision, and multi-modality. To minimize the adaption cost for downstream tasks, many Parameter-Efficient Fine-Tuning (PEFT) techniques are proposed for language and 2D image pre-trained models. However, the specialized PEFT method for 3D pre-trained models is still under-explored. To this end, we introduce Point-PEFT, a novel framework for adapting point cloud pre-trained models with minimal learnable parameters. Specifically, for a pre-trained 3D model, we freeze most of its parameters, and only tune the newly added PEFT modules on downstream tasks, which consist of a Point-prior Prompt and a Geometry-aware Adapter. The Point-prior Prompt adopts a set of learnable prompt tokens, for which we propose to construct a memory bank with domain-specific knowledge, and utilize a parameter-free attention to enhance the prompt tokens. The Geometry-aware Adapter aims to aggregate point cloud features within spatial neighborhoods to capture fine-grained geometric information through local interactions. Extensive experiments indicate that our Point-PEFT can achieve better performance than the full fine-tuning on various downstream tasks, while using only 5% of the trainable parameters, demonstrating the efficiency and effectiveness of our approach. Code is released at https://github.com/Ivan-Tang-3D/Point-PEFT.

Published

2024-03-24

How to Cite

Tang, Y., Zhang, R., Guo, Z., Ma, X., Zhao, B., Wang, Z., … Li, X. (2024). Point-PEFT: Parameter-Efficient Fine-Tuning for 3D Pre-trained Models. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6), 5171–5179. https://doi.org/10.1609/aaai.v38i6.28323

Issue

Section

AAAI Technical Track on Computer Vision V