Nested Variational Inference

Open Access
Authors
Publication date 2022
Host editors
  • M. Ranzato
  • A. Beygelzimer
  • Y. Dauphin
  • P.S. Liang
  • J. Wortman Vaughan
Book title 35th Conference on Neural Information Processing Systems (NeurIPS 2021)
Book subtitle online, 6-14 December 2021
ISBN
  • 9781713845393
Series Advances in Neural Information Processing Systems
Event NeurIPS 2021
Volume | Issue number 25
Pages (from-to) 20423-20435
Publisher San Diego, CA: Neural Information Processing Systems Foundation
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
We develop nested variational inference (NVI), a family of methods that learn proposals for nested importance samplers by minimizing an forward or reverse KL divergence at each level of nesting. NVI is applicable to many commonly-used importance sampling strategies and provides a mechanism for learning intermediate densities, which can serve as heuristics to guide the sampler. Our experiments apply NVI to (a) sample from a multimodal distribution using a learned annealing path (b) learn heuristics that approximate the likelihood of future observations in a hidden Markov model and (c) to perform amortized inference in hierarchical deep generative models. We observe that optimizing nested objectives leads to improved sample quality in terms of log average weight and effective sample size.
Document type Conference contribution
Language English
Published at https://proceedings.neurips.cc/paper_files/paper/2021/hash/ab49b208848abe14418090d95df0d590-Abstract.html
Other links https://www.proceedings.com/63069.html
Downloads
Permalink to this page
Back