Back to Search
Start Over
Convex optimization via inertial algorithms with vanishing Tikhonov regularization: fast convergence to the minimum norm solution.
- Source :
- Mathematical Methods of Operations Research; Jun2024, Vol. 99 Issue 3, p307-347, 41p
- Publication Year :
- 2024
-
Abstract
- In a Hilbertian framework, for the minimization of a general convex differentiable function f, we introduce new inertial dynamics and algorithms that generate trajectories and iterates that converge fastly towards the minimizer of f with minimum norm. Our study is based on the non-autonomous version of the Polyak heavy ball method, which, at time t, is associated with the strongly convex function obtained by adding to f a Tikhonov regularization term with vanishing coefficient ε (t) . In this dynamic, the damping coefficient is proportional to the square root of the Tikhonov regularization parameter ε (t) . By adjusting the speed of convergence of ε (t) towards zero, we will obtain both rapid convergence towards the infimal value of f, and the strong convergence of the trajectories towards the element of minimum norm of the set of minimizers of f. In particular, we obtain an improved version of the dynamic of Su-Boyd-Candès for the accelerated gradient method of Nesterov. This study naturally leads to corresponding first-order algorithms obtained by temporal discretization. In the case of a proper lower semicontinuous and convex function f, we study the proximal algorithms in detail, and show that they benefit from similar properties. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 14322994
- Volume :
- 99
- Issue :
- 3
- Database :
- Complementary Index
- Journal :
- Mathematical Methods of Operations Research
- Publication Type :
- Academic Journal
- Accession number :
- 178231599
- Full Text :
- https://doi.org/10.1007/s00186-024-00867-y