Back to Search Start Over

A New Inexact Gradient Descent Method with Applications to Nonsmooth Convex Optimization

Authors :
Khanh, Pham Duy
Mordukhovich, Boris S.
Tran, Dat Ba
Publication Year :
2023

Abstract

The paper proposes and develops a novel inexact gradient method (IGD) for minimizing C1-smooth functions with Lipschitzian gradients, i.e., for problems of C1,1 optimization. We show that the sequence of gradients generated by IGD converges to zero. The convergence of iterates to stationary points is guaranteed under the Kurdyka- Lojasiewicz (KL) property of the objective function with convergence rates depending on the KL exponent. The newly developed IGD is applied to designing two novel gradient-based methods of nonsmooth convex optimization such as the inexact proximal point methods (GIPPM) and the inexact augmented Lagrangian method (GIALM) for convex programs with linear equality constraints. These two methods inherit global convergence properties from IGD and are confirmed by numerical experiments to have practical advantages over some well-known algorithms of nonsmooth convex optimization.<br />Comment: 23 pages, 8 figures

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2303.08785
Document Type :
Working Paper