first order methods
Recently Published Documents


TOTAL DOCUMENTS

117
(FIVE YEARS 49)

H-INDEX

20
(FIVE YEARS 3)

2021 ◽  
Vol 2021 ◽  
pp. 1-11
Author(s):  
Yan Liu ◽  
Maojun Zhang ◽  
Zhiwei Zhong ◽  
Xiangrong Zeng

In this work, we introduce AdaCN, a novel adaptive cubic Newton method for nonconvex stochastic optimization. AdaCN dynamically captures the curvature of the loss landscape by diagonally approximated Hessian plus the norm of difference between previous two estimates. It only requires at most first order gradients and updates with linear complexity for both time and memory. In order to reduce the variance introduced by the stochastic nature of the problem, AdaCN hires the first and second moment to implement and exponential moving average on iteratively updated stochastic gradients and approximated stochastic Hessians, respectively. We validate AdaCN in extensive experiments, showing that it outperforms other stochastic first order methods (including SGD, Adam, and AdaBound) and stochastic quasi-Newton method (i.e., Apollo), in terms of both convergence speed and generalization performance.


Author(s):  
Zichong Li ◽  
Yangyang Xu

First-order methods (FOMs) have been widely used for solving large-scale problems. A majority of existing works focus on problems without constraint or with simple constraints. Several recent works have studied FOMs for problems with complicated functional constraints. In this paper, we design a novel augmented Lagrangian (AL)–based FOM for solving problems with nonconvex objective and convex constraint functions. The new method follows the framework of the proximal point (PP) method. On approximately solving PP subproblems, it mixes the usage of the inexact AL method (iALM) and the quadratic penalty method, whereas the latter is always fed with estimated multipliers by the iALM. The proposed method achieves the best-known complexity result to produce a near Karush–Kuhn–Tucker (KKT) point. Theoretically, the hybrid method has a lower iteration-complexity requirement than its counterpart that only uses iALM to solve PP subproblems; numerically, it can perform significantly better than a pure-penalty-based method. Numerical experiments are conducted on nonconvex linearly constrained quadratic programs. The numerical results demonstrate the efficiency of the proposed methods over existing ones.


Author(s):  
Hao Luo ◽  
Long Chen

AbstractConvergence analysis of accelerated first-order methods for convex optimization problems are developed from the point of view of ordinary differential equation solvers. A new dynamical system, called Nesterov accelerated gradient (NAG) flow, is derived from the connection between acceleration mechanism and A-stability of ODE solvers, and the exponential decay of a tailored Lyapunov function along with the solution trajectory is proved. Numerical discretizations of NAG flow are then considered and convergence rates are established via a discrete Lyapunov function. The proposed differential equation solver approach can not only cover existing accelerated methods, such as FISTA, Güler’s proximal algorithm and Nesterov’s accelerated gradient method, but also produce new algorithms for composite convex optimization that possess accelerated convergence rates. Both the convex and the strongly convex cases are handled in a unified way in our approach.


Author(s):  
Pavel Dvurechensky ◽  
Shimrit Shtern ◽  
Mathias Staudigl

Author(s):  
Shervin Parvini Ahmadi ◽  
Anders Hansson ◽  
Sina Khoshfetrat Pakazad

AbstractIn this paper, we propose a distributed algorithm for sensor network localization based on a maximum likelihood formulation. It relies on the Levenberg-Marquardt algorithm where the computations are distributed among different computational agents using message passing, or equivalently dynamic programming. The resulting algorithm provides a good localization accuracy, and it converges to the same solution as its centralized counterpart. Moreover, it requires fewer iterations and communications between computational agents as compared to first-order methods. The performance of the algorithm is demonstrated with extensive simulations in Julia in which it is shown that our method outperforms distributed methods that are based on approximate maximum likelihood formulations.


Author(s):  
Vasily I. Repnikov ◽  
Boris V. Faleichik ◽  
Andrew V. Moisa

In this work we present explicit Adams-type multi-step methods with extended stability intervals, which are analogous to the stabilised Chebyshev Runge – Kutta methods. It is proved that for any k ≥ 1 there exists an explicit k-step Adams-type method of order one with stability interval of length 2k. The first order methods have remarkably simple expressions for their coefficients and error constant. A damped modification of these methods is derived. In the general case, to construct a k-step method of order p it is necessary to solve a constrained optimisation problem in which the objective function and p constraints are second degree polynomials in k variables. We calculate higher-order methods up to order six numerically and perform some numerical experiments to confirm the accuracy and stability of the methods.


Sign in / Sign up

Export Citation Format

Share Document