Title
On Learning Non-Convergent Short-Run MCMC Toward Energy-Based Model.
Abstract
This paper studies a curious phenomenon in learning energy-based model (EBM) using MCMC. In each learning iteration, we generate synthesized examples by running a non-convergent, non-mixing, and non-persistent short-run MCMC toward the current model, always starting from the same initial distribution such as uniform noise distribution, and always running a fixed number of MCMC steps. After generating synthesized examples, we then update the model parameters according to the maximum likelihood learning gradient, as if the synthesized examples are fair samples from the current model. We treat this non-convergent short-run MCMC as a learned generator model or a flow model, with the initial image serving as the latent variables, and discard the learned EBM. We provide arguments for treating the learned non-convergent short-run MCMC as a valid model. We show that the learned short-run MCMC is capable of generating realistic images. Moreover, unlike traditional EBM or MCMC, the learned short-run MCMC is also capable of reconstructing observed images and interpolating different images, like generator model or flow model. The code can be found in the Appendix.
Year
Venue
DocType
2019
arXiv: Machine Learning
Journal
Volume
Citations 
PageRank 
abs/1904.09770
2
0.35
References 
Authors
0
3
Name
Order
Citations
PageRank
Erik Nijkamp160.84
Song-Chun Zhu26580741.75
Ying Nian Wu31652267.72