accelerated gradient
Recently Published Documents


TOTAL DOCUMENTS

132
(FIVE YEARS 59)

H-INDEX

14
(FIVE YEARS 3)

Mathematics ◽  
2022 ◽  
Vol 10 (2) ◽  
pp. 259
Author(s):  
Milena J. Petrović ◽  
Dragana Valjarević ◽  
Dejan Ilić ◽  
Aleksandar Valjarević ◽  
Julija Mladenović

We propose an improved variant of the accelerated gradient optimization models for solving unconstrained minimization problems. Merging the positive features of either double direction, as well as double step size accelerated gradient models, we define an iterative method of a simpler form which is generally more effective. Performed convergence analysis shows that the defined iterative method is at least linearly convergent for uniformly convex and strictly convex functions. Numerical test results confirm the efficiency of the developed model regarding the CPU time, the number of iterations and the number of function evaluations metrics.


Mathematics ◽  
2021 ◽  
Vol 9 (23) ◽  
pp. 3021
Author(s):  
Jie Chen ◽  
Joe Suzuki

We consider biclustering that clusters both samples and features and propose efficient convex biclustering procedures. The convex biclustering algorithm (COBRA) procedure solves twice the standard convex clustering problem that contains a non-differentiable function optimization. We instead convert the original optimization problem to a differentiable one and improve another approach based on the augmented Lagrangian method (ALM). Our proposed method combines the basic procedures in the ALM with the accelerated gradient descent method (Nesterov’s accelerated gradient method), which can attain O(1/k2) convergence rate. It only uses first-order gradient information, and the efficiency is not influenced by the tuning parameter λ so much. This advantage allows users to quickly iterate among the various tuning parameters λ and explore the resulting changes in the biclustering solutions. The numerical experiments demonstrate that our proposed method has high accuracy and is much faster than the currently known algorithms, even for large-scale problems.


2021 ◽  
Vol 7 ◽  
pp. 7854-7877
Author(s):  
Yin Fang ◽  
Iman Ahmadianfar ◽  
Arvin Samadi-Koucheksaraee ◽  
Reza Azarsa ◽  
Miklas Scholz ◽  
...  

Author(s):  
Hao Luo ◽  
Long Chen

AbstractConvergence analysis of accelerated first-order methods for convex optimization problems are developed from the point of view of ordinary differential equation solvers. A new dynamical system, called Nesterov accelerated gradient (NAG) flow, is derived from the connection between acceleration mechanism and A-stability of ODE solvers, and the exponential decay of a tailored Lyapunov function along with the solution trajectory is proved. Numerical discretizations of NAG flow are then considered and convergence rates are established via a discrete Lyapunov function. The proposed differential equation solver approach can not only cover existing accelerated methods, such as FISTA, Güler’s proximal algorithm and Nesterov’s accelerated gradient method, but also produce new algorithms for composite convex optimization that possess accelerated convergence rates. Both the convex and the strongly convex cases are handled in a unified way in our approach.


2021 ◽  
Vol 11 (4) ◽  
pp. 287-306
Author(s):  
Jarosław Bilski ◽  
Bartosz Kowalczyk ◽  
Andrzej Marjański ◽  
Michał Gandor ◽  
Jacek Zurada

Abstract In this paper1 a new neural networks training algorithm is presented. The algorithm originates from the Recursive Least Squares (RLS) method commonly used in adaptive filtering. It uses the QR decomposition in conjunction with the Givens rotations for solving a normal equation - resulting from minimization of the loss function. An important parameter in neural networks is training time. Many commonly used algorithms require a big number of iterations in order to achieve a satisfactory outcome while other algorithms are effective only for small neural networks. The proposed solution is characterized by a very short convergence time compared to the well-known backpropagation method and its variants. The paper contains a complete mathematical derivation of the proposed algorithm. There are presented extensive simulation results using various benchmarks including function approximation, classification, encoder, and parity problems. Obtained results show the advantages of the featured algorithm which outperforms commonly used recent state-of-the-art neural networks training algorithms, including the Adam optimizer and the Nesterov’s accelerated gradient.


Sign in / Sign up

Export Citation Format

Share Document