Mean Field Inference in Dependency Networks: An Empirical Study


  • Daniel Lowd University of Oregon
  • Arash Shamaei University of Oregon


Dependency networks are a compelling alternative to Bayesian networks for learning joint probability distributions from data and using them to compute probabilities. A dependency network consists of a set of conditional probability distributions, each representing the probability of a single variable given its Markov blanket. Running Gibbs sampling with these conditional distributions produces a joint distribution that can be used to answer queries, but suffers from the traditional slowness of sampling-based inference. In this paper, we observe that the mean field update equation can be applied to dependency networks, even though the conditional probability distributions may be inconsistent with each other. In experiments with learning and inference on 12 datasets, we demonstrate that mean field inference in dependency networks offers similar accuracy to Gibbs sampling but with orders of magnitude improvements in speed. Compared to Bayesian networks learned on the same data, dependency networks offer higher accuracy at greater amounts of evidence. Furthermore, mean field inference is consistently more accurate in dependency networks than in Bayesian networks learned on the same data.




How to Cite

Lowd, D., & Shamaei, A. (2011). Mean Field Inference in Dependency Networks: An Empirical Study. Proceedings of the AAAI Conference on Artificial Intelligence, 25(1), 404-410. Retrieved from



AAAI Technical Track: Machine Learning