Title
Video (language) modeling: a baseline for generative models of natural videos.
Abstract
We propose a strong baseline model for unsupervised feature learning using video data. By learning to predict missing frames or extrapolate future frames from an input video sequence, the model discovers both spatial and temporal correlations which are useful to represent complex deformations and motion patterns. The models we propose are largely borrowed from the language modeling literature, and adapted to the vision domain by quantizing the space of image patches into a large dictionary. We demonstrate the approach on both a filling and a generation task. For the first time, we show that, after training on natural videos, such a model can predict non-trivial motions over short video sequences.
Year
Venue
Field
2014
CoRR
Computer science,Motion compensation,Speech recognition,Artificial intelligence,Generative grammar,Quantization (signal processing),Machine learning,Language model,Feature learning
DocType
Volume
Citations 
Journal
abs/1412.6604
79
PageRank 
References 
Authors
3.77
17
6
Name
Order
Citations
PageRank
Marc'Aurelio Ranzato15242470.94
Arthur Szlam212610.64
J. Bruna3169782.95
Michaël Mathieu41915151.59
Ronan Collobert54002308.61
Sumit Chopra62835181.37