Title
Accelerated Proximal Alternating Gradient-Descent-Ascent for Nonconvex Minimax Machine Learning
Abstract
Alternating gradient-descent-ascent (AltGDA) is an optimization algorithm that has been widely used for model training in various machine learning applications, which aims to solve a nonconvex minimax optimization problem. However, the existing studies show that it suffers from a high computation complexity in nonconvex minimax optimization. In this paper, we develop a single-loop and fast AltGDA-type algorithm that leverages proximal gradient updates and momentum acceleration to solve regularized nonconvex minimax optimization problems. By leveraging the momentum acceleration technique, we prove that the algorithm converges to a critical point in nonconvex minimax optimization and achieves a computation complexity in the order of $\mathcal{O}\left( {{\kappa ^{\frac{{11}}{6}}}{\varepsilon ^{ - 2}}} \right)$, where ϵ is the desired level of accuracy and κ is the problem’s condition number. Such a computation complexity improves the state-of-the-art complexities of single-loop GDA and AltGDA algorithms (see the summary of comparison in Table I). We demonstrate the effectiveness of our algorithm via an experiment on adversarial deep learning.
Year
DOI
Venue
2022
10.1109/ISIT50566.2022.9834691
2022 IEEE International Symposium on Information Theory (ISIT)
Keywords
DocType
ISSN
Minimiax optimization,alternating gradient descent ascent,proximal gradient,momentum,complexity
Conference
2157-8095
ISBN
Citations 
PageRank 
978-1-6654-2160-7
0
0.34
References 
Authors
8
3
Name
Order
Citations
PageRank
Ziyi Chen100.34
Shaocong Ma201.69
Yi Zhou36517.55