Title
Learning Image Representations Tied to Ego-Motion
Abstract
Understanding how images of objects and scenes behave in response to specific ego-motions is a crucial aspect of proper visual development, yet existing visual learning methods are conspicuously disconnected from the physical source of their images. We propose to exploit proprioceptive motor signals to provide unsupervised regularization in convolutional neural networks to learn visual representations from egocentric video. Specifically, we enforce that our learned features exhibit equivariance, i.e, they respond predictably to transformations associated with distinct ego-motions. With three datasets, we show that our unsupervised feature learning approach significantly outperforms previous approaches on visual recognition and next-best-view prediction tasks. In the most challenging test, we show that features learned from video captured on an autonomous driving platform improve large-scale scene recognition in static images from a disjoint domain.
Year
DOI
Venue
2015
10.1109/ICCV.2015.166
ICCV
Field
DocType
Volume
Computer vision,Disjoint sets,Pattern recognition,Computer science,Convolutional neural network,Id, ego and super-ego,Exploit,Unsupervised learning,Regularization (mathematics),Artificial intelligence,Visual learning,Feature learning
Conference
2015
Issue
ISSN
Citations 
1
1550-5499
58
PageRank 
References 
Authors
2.01
29
2
Name
Order
Citations
PageRank
Dinesh Jayaraman131815.69
Kristen Grauman26258326.34