SMMF: Square-Matricized Momentum Factorization for Memory-Efficient Optimization

Authors

  • Kwangryeol Park Ulsan National Institute of Science and Technology
  • Seulki Lee Ulsan National Institute of Science and Technology

DOI:

https://doi.org/10.1609/aaai.v39i19.34186

Abstract

We propose SMMF (Square-Matricized Momentum Factorization), a memory-efficient optimizer that reduces the memory requirement of the widely used adaptive learning rate optimizers, such as Adam, by up to 96%. SMMF enables flexible and efficient factorization of an arbitrary rank (shape) of the first and second momentum tensors during optimization, based on the proposed square-matricization and one-time single matrix factorization. From this, it becomes effectively applicable to any rank (shape) of momentum tensors, i.e., bias, matrix, and any rank-d tensors, prevalent in various deep model architectures, such as CNNs (high rank) and Transformers (low rank), in contrast to existing memory-efficient optimizers that applies only to a particular (rank-2) momentum tensor, e.g., linear layers. We conduct a regret bound analysis of SMMF, which shows that it converges similarly to non-memory-efficient adaptive learning rate optimizers, such as AdamNC, providing a theoretical basis for its competitive optimization capability. In our experiment, SMMF takes up to 96% less memory compared to state-of-the-art memoryefficient optimizers, e.g., Adafactor, CAME, and SM3, while achieving comparable model performance on various CNN and Transformer tasks.

Published

2025-04-11

How to Cite

Park, K., & Lee, S. (2025). SMMF: Square-Matricized Momentum Factorization for Memory-Efficient Optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 39(19), 19848-19856. https://doi.org/10.1609/aaai.v39i19.34186

Issue

Section

AAAI Technical Track on Machine Learning V