Title
One Method to Rule Them All: Variance Reduction for Data, Parameters and Many New Methods.
Abstract
We propose a remarkably general variance-reduced method suitable for solving regularized empirical risk minimization problems with either a large number of training examples, or a large model dimension, or both. In special cases, our method reduces to several known and previously thought to be unrelated methods, such as {\tt SAGA}, {\tt LSVRG}, {\tt JacSketch}, {\tt SEGA} and {\tt ISEGA}, and their arbitrary sampling and proximal generalizations. However, we also highlight a large number of new specific algorithms with interesting properties. We provide a single theorem establishing linear convergence of the method under smoothness and quasi strong convexity assumptions. With this theorem we recover best-known and sometimes improved rates for known methods arising in special cases. As a by-product, we provide the first unified method and theory for stochastic gradient and stochastic coordinate descent type methods.
Year
Venue
DocType
2019
arXiv: Optimization and Control
Journal
Volume
Citations 
PageRank 
abs/1905.11266
0
0.34
References 
Authors
0
2
Name
Order
Citations
PageRank
Filip Hanzely154.80
Peter Richtárik2131484.53