Argmax Flows and Multinomial Diffusion: Learning Categorical Distributions

Open Access
Authors
Publication date 2022
Host editors
  • M. Ranzato
  • A. Beygelzimer
  • Y. Dauphin
  • P.S. Liang
  • J. Wortman Vaughan
Book title 35th Conference on Neural Information Processing Systems (NeurIPS 2021)
Book subtitle online, 6-14 December 2021
ISBN
  • 9781713845393
Series Advances in Neural Information Processing Systems
Event NeurIPS 2021
Volume | Issue number 15
Pages (from-to) 12454-12465
Publisher San Diego, CA: Neural Information Processing Systems Foundation
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Generative flows and diffusion models have been predominantly trained on ordinal data, for example natural images. This paper introduces two extensions of flows and diffusion for categorical data such as language or image segmentation: Argmax Flows and Multinomial Diffusion. Argmax Flows are defined by a composition of a continuous distribution (such as a normalizing flow), and an argmax function. To optimize this model, we learn a probabilistic inverse for the argmax that lifts the categorical data to a continuous space. Multinomial Diffusion gradually adds categorical noise in a diffusion process, for which the generative denoising process is learned. We demonstrate that our method outperforms existing dequantization approaches on text modelling and modelling on image segmentation maps in log-likelihood.
Document type Conference contribution
Note With supplemental file
Language English
Published at https://papers.nips.cc/paper/2021/hash/67d96d458abdef21792e6d8e590244e7-Abstract.html
Other links https://www.proceedings.com/63069.html
Downloads
Supplementary materials
Permalink to this page
Back