Title
Stochastic Learning under Random Reshuffling.
Abstract
In empirical risk optimization, it has been observed that stochastic gradient implementations that rely on random reshuffling of the data achieve better performance than implementations that rely on sampling the data uniformly. Recent works have pursued justifications for this behavior by examining the convergence rate of the learning process under diminishing step-sizes. This work focuses on the constant step-size case. In this case, convergence is guaranteed to a small neighborhood of the optimizer albeit at a linear rate. The analysis establishes analytically that random reshuffling outperforms uniform sampling by showing explicitly that iterates approach a smaller neighborhood of size $O(mu^2)$ around the minimizer rather than $O(mu)$. Furthermore, we derive an analytical expression for the steady-state mean-square-error performance of the algorithm, which helps clarify in greater detail the differences between sampling with and without replacement. We also explain the periodic behavior that is observed in random reshuffling implementations.
Year
Venue
Field
2018
arXiv: Learning
Convergence (routing),Applied mathematics,Computer science,Sampling (statistics),Rate of convergence,Periodic graph (geometry),Iterated function
DocType
Volume
Citations 
Journal
abs/1803.07964
3
PageRank 
References 
Authors
0.44
0
4
Name
Order
Citations
PageRank
B. Ying1598.60
Kun Yuan232423.18
Stefan Vlaski32311.39
Ali H. Sayed49134667.71