Back to Search Start Over

Accelerated Proximal Alternating Gradient-Descent-Ascent for Nonconvex Minimax Machine Learning

Authors :
Chen, Ziyi
Ma, Shaocong
Zhou, Yi
Source :
2022 IEEE International Symposium on Information Theory (ISIT).
Publication Year :
2022
Publisher :
IEEE, 2022.

Abstract

Alternating gradient-descent-ascent (AltGDA) is an optimization algorithm that has been widely used for model training in various machine learning applications, which aims to solve a nonconvex minimax optimization problem. However, the existing studies show that it suffers from a high computation complexity in nonconvex minimax optimization. In this paper, we develop a single-loop and fast AltGDA-type algorithm that leverages proximal gradient updates and momentum acceleration to solve regularized nonconvex minimax optimization problems. By leveraging the momentum acceleration technique, we prove that the algorithm converges to a critical point in nonconvex minimax optimization and achieves a computation complexity in the order of $\mathcal{O}(\kappa^{\frac{11}{6}}\epsilon^{-2})$, where $\epsilon$ is the desired level of accuracy and $\kappa$ is the problem's condition number. {Such a computation complexity improves the state-of-the-art complexities of single-loop GDA and AltGDA algorithms (see the summary of comparison in \Cref{table1})}. We demonstrate the effectiveness of our algorithm via an experiment on adversarial deep learning.<br />Comment: 12 pages, 1 figure. Added acknowledgement of NSF funding. arXiv admin note: text overlap with arXiv:2102.04653

Details

Database :
OpenAIRE
Journal :
2022 IEEE International Symposium on Information Theory (ISIT)
Accession number :
edsair.doi.dedup.....521f2f67f076b9c3317b006d00e2b4d5