Self Normalizing Flows
| Authors | |
|---|---|
| Publication date | 2021 |
| Journal | Proceedings of Machine Learning Research |
| Event | 38th International Conference on Machine Learning |
| Volume | Issue number | 139 |
| Pages (from-to) | 5378-5387 |
| Organisations |
|
| Abstract |
Efficient gradient computation of the Jacobian determinant term is a core problem in many machine learning settings, and especially so in the normalizing flow framework. Most proposed flow models therefore either restrict to a function class with easy evaluation of the Jacobian determinant, or an efficient estimator thereof. However, these restrictions limit the performance of such density models, frequently requiring significant depth to reach desired performance levels. In this work, we propose Self Normalizing Flows, a flexible framework for training normalizing flows by replacing expensive terms in the gradient by learned approximate inverses at each layer. This reduces the computational complexity of each layer’s exact update from O(D3) to O(D2), allowing for the training of flow architectures which were otherwise computationally infeasible, while also providing efficient sampling. We show experimentally that such models are remarkably stable and optimize to similar data likelihood values as their exact gradient counterparts, while training more quickly and surpassing the performance of functionally constrained counterparts.
|
| Document type | Article |
| Note | International Conference on Machine Learning, 18-24 July 2021, Virtual. - With supplementary file. |
| Language | English |
| Published at | https://arxiv.org/abs/2011.07248 https://proceedings.mlr.press/v139/keller21a.html |
| Other links | https://youtu.be/6Q3b3MergqI |
| Downloads |
keller21a
(Final published version)
|
| Supplementary materials | |
| Permalink to this page | |