Abstract | ||
---|---|---|
A promising class of generative models maps points from a simple distribution to a complex distribution through an invertible neural network. Likelihood-based training of these models requires restricting their architectures to allow cheap computation of Jacobian determinants. Alternatively, the Jacobian trace can be used if the transformation is specified by an ordinary differential equation. In this paper, we use Hutchinsonu0027s trace estimator to give a scalable unbiased estimate of the log-density. The result is a continuous-time invertible generative model with unbiased density estimation and one-pass sampling, while allowing unrestricted neural network architectures. We demonstrate our approach on high-dimensional density estimation, image generation, and variational inference, achieving the state-of-the-art among exact likelihood methods with efficient sampling. |
Year | Venue | Field |
---|---|---|
2018 | international conference on learning representations | Density estimation,Mathematical optimization,Ordinary differential equation,Jacobian matrix and determinant,Algorithm,Sampling (statistics),Invertible matrix,Artificial neural network,Mathematics,Generative model,Estimator |
DocType | Volume | Citations |
Journal | abs/1810.01367 | 11 |
PageRank | References | Authors |
0.53 | 5 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Will Grathwohl | 1 | 11 | 2.90 |
Tian Qi Chen | 2 | 83 | 7.00 |
Jesse Bettencourt | 3 | 30 | 2.14 |
Ilya Sutskever | 4 | 25814 | 1120.24 |
David K. Duvenaud | 5 | 629 | 32.63 |