Back to Search Start Over

Superfast Second-Order Methods for Unconstrained Convex Optimization.

Authors :
Nesterov, Yurii
Source :
Journal of Optimization Theory & Applications. Oct2021, Vol. 191 Issue 1, p1-30. 30p.
Publication Year :
2021

Abstract

In this paper, we present new second-order methods with convergence rate O k - 4 , where k is the iteration counter. This is faster than the existing lower bound for this type of schemes (Agarwal and Hazan in Proceedings of the 31st conference on learning theory, PMLR, pp. 774–792, 2018; Arjevani and Shiff in Math Program 178(1–2):327–360, 2019), which is O k - 7 / 2 . Our progress can be explained by a finer specification of the problem class. The main idea of this approach consists in implementation of the third-order scheme from Nesterov (Math Program 186:157–183, 2021) using the second-order oracle. At each iteration of our method, we solve a nontrivial auxiliary problem by a linearly convergent scheme based on the relative non-degeneracy condition (Bauschke et al. in Math Oper Res 42:330–348, 2016; Lu et al. in SIOPT 28(1):333–354, 2018). During this process, the Hessian of the objective function is computed once, and the gradient is computed O ln 1 ϵ times, where ϵ is the desired accuracy of the solution for our problem. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00223239
Volume :
191
Issue :
1
Database :
Academic Search Index
Journal :
Journal of Optimization Theory & Applications
Publication Type :
Academic Journal
Accession number :
152927721
Full Text :
https://doi.org/10.1007/s10957-021-01930-y