Title
An Accelerated Second-Order Method for Distributed Stochastic Optimization.
Abstract
We consider distributed stochastic optimization problems that are solved with master/workers computation architecture. Statistical arguments allow to exploit statistical similarity and approximate this problem by a finite-sum problem, for which we propose an inexact accelerated cubic-regularized Newton's method that achieves lower communication complexity bound for this setting and improves upon existing upper bound. We further exploit this algorithm to obtain convergence rate bounds for the original stochastic optimization problem and compare our bounds with the existing bounds in several regimes when the goal is to minimize the number of communication rounds and increase the parallelization by increasing the number of workers.
Year
DOI
Venue
2021
10.1109/CDC45484.2021.9683400
CDC
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
0
7