RoSA: Enhancing Parameter-Efficient Fine-Tuning via RoPE-aware Selective Adaptation in Large Language Models

Authors

  • Dayan Pan School of Computer Science and Engineering, Beihang University, Beijing, China MOE Engineering Research Center of Advanced Computer Application Technology, Beihang University, China
  • Jingyuan Wang School of Computer Science and Engineering, Beihang University, Beijing, China School of Economics and Management, Beihang University, Beijing, China MOE Engineering Research Center of Advanced Computer Application Technology, Beihang University, China
  • Yilong Zhou School of Computer Science and Engineering, Beihang University, Beijing, China MOE Engineering Research Center of Advanced Computer Application Technology, Beihang University, China
  • Jiawei Cheng School of Computer Science and Engineering, Beihang University, Beijing, China MOE Engineering Research Center of Advanced Computer Application Technology, Beihang University, China Department of Data Science, City University of Hong Kong, Hong Kong, China
  • Pengyue Jia Department of Data Science, City University of Hong Kong, Hong Kong, China
  • Xiangyu Zhao Department of Data Science, City University of Hong Kong, Hong Kong, China

DOI:

https://doi.org/10.1609/aaai.v40i18.38589

Abstract

Fine-tuning large language models is essential for task-specific adaptation, yet it remains computationally prohibitive. Parameter-Efficient Fine-Tuning (PEFT) methods have emerged as a solution, but current approaches typically ignore the distinct roles of model components and the heterogeneous importance across layers, thereby limiting adaptation efficiency. Motivated by the observation that Rotary Position Embeddings (RoPE) induce critical activations in the low-frequency dimensions of attention states, we propose RoPE-aware Selective Adaptation (RoSA), a novel PEFT framework that allocates trainable parameters in a more targeted and effective manner. RoSA comprises a RoPE-aware Attention Enhancement (RoAE) module, which selectively enhances the low-frequency components of RoPE-influenced attention states, and a Dynamic Layer Selection (DLS) strategy that adaptively identifies and updates the most critical layers based on LayerNorm gradient norms. By combining dimension-wise enhancement with layer-wise adaptation, RoSA achieves more targeted and efficient fine-tuning. Extensive experiments on fifteen commonsense and arithmetic benchmarks demonstrate that RoSA outperforms mainstream PEFT methods under comparable trainable parameters.

Downloads

Published

2026-03-14

How to Cite

Pan, D., Wang, J., Zhou, Y., Cheng, J., Jia, P., & Zhao, X. (2026). RoSA: Enhancing Parameter-Efficient Fine-Tuning via RoPE-aware Selective Adaptation in Large Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 40(18), 15600–15608. https://doi.org/10.1609/aaai.v40i18.38589

Issue

Section

AAAI Technical Track on Data Mining & Knowledge Management II