scholarly journals An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and Its Implications to Second-Order Methods

2013 ◽  
Vol 23 (2) ◽  
pp. 1092-1125 ◽  
Author(s):  
Renato D. C. Monteiro ◽  
B. F. Svaiter
2019 ◽  
Vol 484 (6) ◽  
pp. 667-671
Author(s):  
A. V. Gasnikov ◽  
E. A. Gorbunov ◽  
D. A. Kovalev ◽  
A. A. M. Mokhammed ◽  
E. A. Chernousova

The Monteiro-Svaiter accelerated hybrid proximal extragradient method (2013) with one step of Newton’s method used at every iteration for the approximate solution of an auxiliary problem is considered. The Monteiro-Svaiter method is optimal (with respect to the number of gradient and Hessian evaluations for the optimized function) for sufficiently smooth convex optimization problems in the class of methods using only the gradient and Hessian of the optimized function. An optimal tensor method involving higher derivatives is proposed by replacing Newton’s step with a step of Yu.E. Nesterov’s recently proposed tensor method (2018) and by using a special generalization of the step size selection condition in the outer accelerated proximal extragradient method. This tensor method with derivatives up to the third order inclusive is found fairly practical, since the complexity of its iteration is comparable with that of Newton’s one. Thus, a constructive solution is obtained for Nesterov’s problem (2018) of closing the gap between tight lower and overstated upper bounds for the convergence rate of existing tensor methods of order p ≥ 3.


Automatica ◽  
2022 ◽  
Vol 137 ◽  
pp. 110118
Author(s):  
Lanlin Yu ◽  
Xiaodong Cheng ◽  
Jacquelien M.A. Scherpen ◽  
Junlin Xiong

2020 ◽  
Vol 65 (2) ◽  
pp. 846-853
Author(s):  
Natasa Krklec Jerinkic ◽  
Dusan Jakovetic ◽  
Natasa Krejic ◽  
Dragana Bajovic

Sign in / Sign up

Export Citation Format

Share Document