Title
An Analysis of SVD for Deep Rotation Estimation
Abstract
Symmetric orthogonalization via SVD, and closely related procedures, are well-known techniques for projecting matrices onto $O(n)$ or $SO(n)$. These tools have long been used for applications in computer vision, for example optimal 3D alignment problems solved by orthogonal Procrustes, rotation averaging, or Essential matrix decomposition. Despite its utility in different settings, SVD orthogonalization as a procedure for producing rotation matrices is typically overlooked in deep learning models, where the preferences tend toward classic representations like unit quaternions, Euler angles, and axis-angle, or more recently-introduced methods. Despite the importance of 3D rotations in computer vision and robotics, a single universally effective representation is still missing. Here, we explore the viability of SVD orthogonalization for 3D rotations in neural networks. We present a theoretical analysis that shows SVD is the natural choice for projecting onto the rotation group. Our extensive quantitative analysis shows simply replacing existing representations with the SVD orthogonalization procedure obtains state of the art performance in many deep learning applications covering both supervised and unsupervised training.
Year
Venue
DocType
2020
NeurIPS
Conference
Volume
Citations 
PageRank 
33
0
0.34
References 
Authors
0
7
Name
Order
Citations
PageRank
Jake Levinson100.34
Carlos Esteves224.42
Kefan Chen300.34
Noah Snavely44262197.04
Angjoo Kanazawa527210.36
Afshin Rostamizadeh691144.15
Ameesh Makadia770134.69