@article{Liu_Cai_Guo_Chen_2021, title={TransTailor: Pruning the Pre-trained Model for Improved Transfer Learning}, volume={35}, url={https://ojs.aaai.org/index.php/AAAI/article/view/17046}, DOI={10.1609/aaai.v35i10.17046}, abstractNote={The increasing of pre-trained models has significantly facilitated the performance on limited data tasks with transfer learning. However, progress on transfer learning mainly focuses on optimizing the weights of pre-trained models, which ignores the structure mismatch between the model and the target task. This paper aims to improve the transfer performance from another angle - in addition to tuning the weights, we tune the structure of pre-trained models, in order to better match the target task. To this end, we propose TransTailor, targeting at pruning the pre-trained model for improved transfer learning. Different from traditional pruning pipelines, we prune and fine-tune the pre-trained model according to the target-aware weight importance, generating an optimal sub-model tailored for a specific target task. In this way, we transfer a more suitable sub-structure that can be applied during fine-tuning to benefit the final performance. Extensive experiments on multiple pre-trained models and datasets demonstrate that TransTailor outperforms the traditional pruning methods and achieves competitive or even better performance than other state-of-the-art transfer learning methods while using a smaller model. Notably, on the Stanford Dogs dataset, TransTailor can achieve 2.7% accuracy improvement over other transfer methods with 20% fewer FLOPs.}, number={10}, journal={Proceedings of the AAAI Conference on Artificial Intelligence}, author={Liu, Bingyan and Cai, Yifeng and Guo, Yao and Chen, Xiangqun}, year={2021}, month={May}, pages={8627-8634} }