TY - JOUR
AU - Bu, Yuheng
AU - Gao, Weihao
AU - Zou, Shaofeng
AU - Veeravalli, Venugopal
PY - 2020/04/03
Y2 - 2024/03/05
TI - Information-Theoretic Understanding of Population Risk Improvement with Model Compression
JF - Proceedings of the AAAI Conference on Artificial Intelligence
JA - AAAI
VL - 34
IS - 04
SE - AAAI Technical Track: Machine Learning
DO - 10.1609/aaai.v34i04.5730
UR - https://ojs.aaai.org/index.php/AAAI/article/view/5730
SP - 3300-3307
AB - <p>We show that model compression can improve the population risk of a pre-trained model, by studying the tradeoff between the decrease in the generalization error and the increase in the empirical risk with model compression. We first prove that model compression reduces an information-theoretic bound on the generalization error; this allows for an interpretation of model compression as a regularization technique to avoid overfitting. We then characterize the increase in empirical risk with model compression using rate distortion theory. These results imply that the population risk could be improved by model compression if the decrease in generalization error exceeds the increase in empirical risk. We show through a linear regression example that such a decrease in population risk due to model compression is indeed possible. Our theoretical results further suggest that the Hessian-weighted <em>K</em>-means clustering compression approach can be improved by regularizing the distance between the clustering centers. We provide experiments with neural networks to support our theoretical assertions.</p>
ER -