Generalized Variational Inference via Optimal Transport

Authors

  • Jinjin Chi College of Computer Science and Technology, Jilin University, China Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, China
  • Zhichao Zhang College of Computer Science and Technology, Jilin University, China Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, China
  • Zhiyao Yang College of Computer Science and Technology, Jilin University, China Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, China
  • Jihong Ouyang College of Computer Science and Technology, Jilin University, China Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, China
  • Hongbin Pei MOE KLINNS Lab, School of Cyber Science and Engineering, Xi'an Jiaotong University, China

DOI:

https://doi.org/10.1609/aaai.v38i10.29035

Keywords:

ML: Bayesian Learning

Abstract

Variational Inference (VI) has gained popularity as a flexible approximate inference scheme for computing posterior distributions in Bayesian models. Original VI methods use Kullback-Leibler (KL) divergence to construct variational objectives. However, KL divergence has zero-forcing behavior and is completely agnostic to the metric of the underlying data distribution, resulting in bad approximations. To alleviate this issue, we propose a new variational objective by using Optimal Transport (OT) distance, which is a metric-aware divergence, to measure the difference between approximate posteriors and priors. The superior performance of OT distance enables us to learn more accurate approximations. We further enhance the objective by gradually including the OT term using a hyperparameter λ for over-parameterized models. We develop a Variational inference method with OT (VOT) which presents a gradient-based black-box framework for solving Bayesian models, even when the density function of approximate distribution is not available. We provide the consistency analysis of approximate posteriors and demonstrate the practical effectiveness on Bayesian neural networks and variational autoencoders.

Published

2024-03-24

How to Cite

Chi, J., Zhang, Z., Yang, Z., Ouyang, J., & Pei, H. (2024). Generalized Variational Inference via Optimal Transport. Proceedings of the AAAI Conference on Artificial Intelligence, 38(10), 11534-11542. https://doi.org/10.1609/aaai.v38i10.29035

Issue

Section

AAAI Technical Track on Machine Learning I