DAMix: Exploiting Deep Autoregressive Model Zoo for Improving Lossless Compression Generalization

Authors

  • Qishi Dong Hong Kong Baptist University Huawei Noah’s Ark Lab
  • Fengwei Zhou Huawei Noah's Ark Lab
  • Ning Kang Huawei Noah's Ark Lab
  • Chuanlong Xie Beijing Normal University Huawei Noah’s Ark Lab
  • Shifeng Zhang Huawei Noah's Ark Lab
  • Jiawei Li Huawei Noah's Ark Lab
  • Heng Peng Hong Kong Baptist University
  • Zhenguo Li Huawei Noah's Ark Lab

DOI:

https://doi.org/10.1609/aaai.v37i4.25543

Keywords:

DMKM: Data Compression, ML: Bayesian Learning, ML: Ensemble Methods, ML: Probabilistic Methods

Abstract

Deep generative models have demonstrated superior performance in lossless compression on identically distributed data. However, in real-world scenarios, data to be compressed are of various distributions and usually cannot be known in advance. Thus, commercially expected neural compression must have strong Out-of-Distribution (OoD) generalization capabilities. Compared with traditional compression methods, deep learning methods have intrinsic flaws for OoD generalization. In this work, we make the attempt to tackle this challenge via exploiting a zoo of Deep Autoregressive models (DAMix). We build a model zoo consisting of autoregressive models trained on data from diverse distributions. In the test phase, we select useful expert models by a simple model evaluation score and adaptively aggregate the predictions of selected models. By assuming the outputs from each expert model are biased in favor of their training distributions, a von Mises-Fisher based filter is proposed to recover the value of unbiased predictions that provides more accurate density estimations than a single model. We derive the posterior of unbiased predictions as well as concentration parameters in the filter, and a novel temporal Stein variational gradient descent for sequential data is proposed to adaptively update the posterior distributions. We evaluate DAMix on 22 image datasets, including in-distribution and OoD data, and demonstrate that making use of unbiased predictions has up to 45.6% improvement over the single model trained on ImageNet.

Downloads

Published

2023-06-26

How to Cite

Dong, Q., Zhou, F., Kang, N., Xie, C., Zhang, S., Li, J., Peng, H., & Li, Z. (2023). DAMix: Exploiting Deep Autoregressive Model Zoo for Improving Lossless Compression Generalization. Proceedings of the AAAI Conference on Artificial Intelligence, 37(4), 4250-4258. https://doi.org/10.1609/aaai.v37i4.25543

Issue

Section

AAAI Technical Track on Data Mining and Knowledge Management