MHA2MLA-VLM: Enabling DeepSeek’s Economical Multi-Head Latent Attention Across Vision-Language Models

Authors

  • Xiaoran Fan Fudan University
  • Zhichao Sun Fudan University
  • Tao Ji Fudan University
  • Lixing Shen Hikvision Inc
  • Tao Gui Fudan University Shanghai Innovation Institute Pengcheng Laboratory

DOI:

https://doi.org/10.1609/aaai.v40i36.40319

Abstract

As vision-language models (VLMs) tackle increasingly complex and multimodal tasks, the rapid growth of Key-Value (KV) cache imposes significant memory and computational bottlenecks during inference. While Multi-Head Latent Attention (MLA) offers an effective means to compress the KV cache and accelerate inference, adapting existing VLMs to the MLA architecture without costly pretraining remains largely unexplored. In this work, we present \textbf{MHA2MLA-VLM}, a parameter-efficient and multimodal-aware framework for converting off-the-shelf VLMs to MLA. Our approach features two core techniques: (1) a modality-adaptive partial-RoPE strategy that supports both traditional and multimodal settings by selectively masking nonessential dimensions, and (2) a modality-decoupled low-rank approximation method that independently compresses the visual and textual KV spaces. Furthermore, we introduce parameter-efficient fine-tuning to minimize adaptation cost and demonstrate that minimizing output activation error, rather than parameter distance, substantially reduces performance loss. Extensive experiments on three representative VLMs show that MHA2MLA-VLM restores original model performance with minimal supervised data, significantly reduces KV cache footprint, and integrates seamlessly with KV quantization.

Downloads

Published

2026-03-14

How to Cite

Fan, X., Sun, Z., Ji, T., Shen, L., & Gui, T. (2026). MHA2MLA-VLM: Enabling DeepSeek’s Economical Multi-Head Latent Attention Across Vision-Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 40(36), 30638-30646. https://doi.org/10.1609/aaai.v40i36.40319

Issue

Section

AAAI Technical Track on Natural Language Processing I