What Makes Quantization for Large Language Model Hard? An Empirical Study from the Lens of Perturbation

Authors

  • Zhuocheng Gong Wangxuan Institute of Computer Technology, Peking University
  • Jiahao Liu Meituan
  • Jingang Wang Meituan
  • Xunliang Cai Meituan
  • Dongyan Zhao Wangxuan Institute of Computer Technology, Peking University State Key Laboratory of Media Convergence Production Technology and Systems
  • Rui Yan Gaoling School of Artificial Intelligence, Renmin University of China

DOI:

https://doi.org/10.1609/aaai.v38i16.29765

Keywords:

NLP: (Large) Language Models

Abstract

Quantization has emerged as a promising technique for improving the memory and computational efficiency of large language models (LLMs). Though the trade-off between performance and efficiency is well-known, there is still much to be learned about the relationship between quantization and LLM performance. To shed light on this relationship, we propose a new perspective on quantization, viewing it as perturbations added to the weights and activations of LLMs. We call this approach ``the lens of perturbation". Using this lens, we conduct experiments with various artificial perturbations to explore their impact on LLM performance. Our findings reveal several connections between the properties of perturbations and LLM performance, providing insights into the failure cases of uniform quantization and suggesting potential solutions to improve the robustness of LLM quantization. To demonstrate the significance of our findings, we implement a simple non-uniform quantization approach based on our insights. Our experiments show that this approach achieves minimal performance degradation on both 4-bit weight quantization and 8-bit quantization for weights and activations. These results validate the correctness of our approach and highlight its potential to improve the efficiency of LLMs without sacrificing performance.

Downloads

Published

2024-03-24

How to Cite

Gong, Z., Liu, J., Wang, J., Cai, X., Zhao, D., & Yan, R. (2024). What Makes Quantization for Large Language Model Hard? An Empirical Study from the Lens of Perturbation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16), 18082-18089. https://doi.org/10.1609/aaai.v38i16.29765

Issue

Section

AAAI Technical Track on Natural Language Processing I