DP-MemArc: Differential Privacy Transfer Learning for Memory Efficient Language Models
DOI:
https://doi.org/10.1609/aaai.v39i25.34830Abstract
Large language models have repeatedly shown outstanding performance across diverse applications. However, deploying these models can inadvertently risk user privacy. The significant memory demands during training pose a major challenge in terms of resource consumption. This substantial size places a heavy load on memory resources, raising considerable practical concerns. In this paper, we introduce DP-MemArc, a novel training framework aimed at reducing the memory costs of large language models while emphasizing the protection of user data privacy. DP-MemArc incorporates side network or reversible network designs to support a variety of differential privacy memory-efficient fine-tuning schemes. Our approach not only achieves about 2.5 times in memory optimization but also ensures robust privacy protection, keeping user data secure and confidential. Extensive experiments have demonstrated that DP-MemArc effectively provides differential privacy-efficient fine-tuning across different task scenarios.Downloads
Published
2025-04-11
How to Cite
Liu, Y., Peng, X., Zhang, Y., Ke, X., Deng, S., Cao, J., … Du, T. (2025). DP-MemArc: Differential Privacy Transfer Learning for Memory Efficient Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 39(25), 26317–26325. https://doi.org/10.1609/aaai.v39i25.34830
Issue
Section
AAAI Technical Track on Philosophy and Ethics of AI