Title
SEBOOST - Boosting Stochastic Learning Using Subspace Optimization Techniques.
Abstract
We present SEBOOST, a technique for boosting the performance of existing stochastic optimization methods. SEBOOST applies a secondary optimization process in the subspace spanned by the last steps and descent directions. The method was inspired by the SESOP optimization method, and has been adapted for the stochastic learning. It can be applied on top of any existing optimization method with no need to tweak the internal algorithm. We show that the method is able to boost the performance of different algorithms, and make them more robust to changes in their hyper-parameters. As the boosting steps of SEBOOST are applied between large sets of descent steps, the additional subspace optimization hardly increases the overall computational burden. We introduce hyper-parameters that control the balance between the baseline method and the secondary optimization process. The method was evaluated on several deep learning tasks, demonstrating significant improvement in performance. Video presentation is given in [15]
Year
Venue
DocType
2016
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016)
Conference
Volume
ISSN
Citations 
29
1049-5258
1
PageRank 
References 
Authors
0.35
5
4
Name
Order
Citations
PageRank
Richardson, Elad1301.36
Herskovitz, Rom210.35
Ginsburg, Boris3758.77
Michael Zibulevsky41087124.28