One Step Learning, One Step Review
DOI:
https://doi.org/10.1609/aaai.v38i11.29159Keywords:
ML: Transfer, Domain Adaptation, Multi-Task Learning, CV: Learning & Optimization for CV, CV: Representation Learning for Vision, ML: Representation Learning, CV: Large Vision ModelsAbstract
Visual fine-tuning has garnered significant attention with the rise of pre-trained vision models. The current prevailing method, full fine-tuning, suffers from the issue of knowledge forgetting as it focuses solely on fitting the downstream training set. In this paper, we propose a novel weight rollback-based fine-tuning method called OLOR (One step Learning, One step Review). OLOR combines fine-tuning with optimizers, incorporating a weight rollback term into the weight update term at each step. This ensures consistency in the weight range of upstream and downstream models, effectively mitigating knowledge forgetting and enhancing fine-tuning performance. In addition, a layer-wise penalty is presented to employ penalty decay and the diversified decay rate to adjust the weight rollback levels of layers for adapting varying downstream tasks. Through extensive experiments on various tasks such as image classification, object detection, semantic segmentation, and instance segmentation, we demonstrate the general applicability and state-of-the-art performance of our proposed OLOR. Code is available at https://github.com/rainbow-xiao/OLOR-AAAI-2024.Downloads
Published
2024-03-24
How to Cite
Huang, X., Li, Q., Li, X., & Gao, X. (2024). One Step Learning, One Step Review. Proceedings of the AAAI Conference on Artificial Intelligence, 38(11), 12644-12652. https://doi.org/10.1609/aaai.v38i11.29159
Issue
Section
AAAI Technical Track on Machine Learning II