Title
Accelerated gradient sliding for structured convex optimization
Abstract
Our main goal in this paper is to show that one can skip gradient computations for gradient descent type methods applied to certain structured convex programming (CP) problems. To this end, we first present an accelerated gradient sliding (AGS) method for minimizing the summation of two smooth convex functions with different Lipschitz constants. We show that the AGS method can skip the gradient computation for one of these smooth components without slowing down the overall optimal rate of convergence. This result is much sharper than the classic black-box CP complexity results especially when the difference between the two Lipschitz constants associated with these components is large. We then consider an important class of bilinear saddle point problem whose objective function is given by the summation of a smooth component and a nonsmooth one with a bilinear saddle point structure. Using the aforementioned AGS method for smooth composite optimization and Nesterov’s smoothing technique, we show that one only needs $${{\mathcal{O}}}(1/\sqrt{\varepsilon })$$ gradient computations for the smooth component while still preserving the optimal $${{\mathcal{O}}}(1/\varepsilon )$$ overall iteration complexity for solving these saddle point problems. We demonstrate that even more significant savings on gradient computations can be obtained for strongly convex smooth and bilinear saddle point problems.
Year
DOI
Venue
2022
10.1007/s10589-022-00365-z
Computational Optimization and Applications
Keywords
DocType
Volume
Convex programming, Accelerated gradient sliding, Structure, Complexity, Nesterov’s method, 90C25, 90C06, 49M37
Journal
82
Issue
ISSN
Citations 
2
0926-6003
0
PageRank 
References 
Authors
0.34
19
2
Name
Order
Citations
PageRank
Guanghui Lan1121266.26
Yuyuan Ouyang200.34