newton direction
Recently Published Documents


TOTAL DOCUMENTS

23
(FIVE YEARS 2)

H-INDEX

4
(FIVE YEARS 0)

Symmetry ◽  
2021 ◽  
Vol 13 (10) ◽  
pp. 1829
Author(s):  
Jamilu Sabi’u ◽  
Kanikar Muangchoo ◽  
Abdullah Shah ◽  
Auwal Bala Abubakar ◽  
Kazeem Olalekan Aremu

This article presents an inexact optimal hybrid conjugate gradient (CG) method for solving symmetric nonlinear systems. The method is a convex combination of the optimal Dai–Liao (DL) and the extended three-term Polak–Ribiére–Polyak (PRP) CG methods. However, two different formulas for selecting the convex parameter are derived by using the conjugacy condition and also by combining the proposed direction with the default Newton direction. The proposed method is again derivative-free, therefore the Jacobian information is not required throughout the iteration process. Furthermore, the global convergence of the proposed method is shown using some appropriate assumptions. Finally, the numerical performance of the method is demonstrated by solving some examples of symmetric nonlinear problems and comparing them with some existing symmetric nonlinear equations CG solvers.


Symmetry ◽  
2021 ◽  
Vol 13 (2) ◽  
pp. 234
Author(s):  
Jamilu Sabi’u ◽  
Kanikar Muangchoo ◽  
Abdullah Shah ◽  
Auwal Bala Abubakar ◽  
Lateef Olakunle Jolaoso

Inspired by the large number of applications for symmetric nonlinear equations, this article will suggest two optimal choices for the modified Polak–Ribiére–Polyak (PRP) conjugate gradient (CG) method by minimizing the measure function of the search direction matrix and combining the proposed direction with the default Newton direction. In addition, the corresponding PRP parameters are incorporated with the Li and Fukushima approximate gradient to propose two robust CG-type algorithms for finding solutions for large-scale systems of symmetric nonlinear equations. We have also demonstrated the global convergence of the suggested algorithms using some classical assumptions. Finally, we demonstrated the numerical advantages of the proposed algorithms compared to some of the existing methods for nonlinear symmetric equations.


2018 ◽  
Vol 30 (6) ◽  
pp. 1673-1724 ◽  
Author(s):  
Chien-Chih Wang ◽  
Kent Loong Tan ◽  
Chun-Ting Chen ◽  
Yu-Hsiang Lin ◽  
S. Sathiya Keerthi ◽  
...  

Deep learning involves a difficult nonconvex optimization problem with a large number of weights between any two adjacent layers of a deep structure. To handle large data sets or complicated networks, distributed training is needed, but the calculation of function, gradient, and Hessian is expensive. In particular, the communication and the synchronization cost may become a bottleneck. In this letter, we focus on situations where the model is distributedly stored and propose a novel distributed Newton method for training deep neural networks. By variable and feature-wise data partitions and some careful designs, we are able to explicitly use the Jacobian matrix for matrix-vector products in the Newton method. Some techniques are incorporated to reduce the running time as well as memory consumption. First, to reduce the communication cost, we propose a diagonalization method such that an approximate Newton direction can be obtained without communication between machines. Second, we consider subsampled Gauss-Newton matrices for reducing the running time as well as the communication cost. Third, to reduce the synchronization cost, we terminate the process of finding an approximate Newton direction even though some nodes have not finished their tasks. Details of some implementation issues in distributed environments are thoroughly investigated. Experiments demonstrate that the proposed method is effective for the distributed training of deep neural networks. Compared with stochastic gradient methods, it is more robust and may give better test accuracy.


2013 ◽  
Vol 347-350 ◽  
pp. 2586-2589
Author(s):  
Wen Jun Wang ◽  
Ju Bo Zhu ◽  
Xiao Jun Duan

When the Hessian matrix is not positive, the Newton direction maybe not the descending direction. A new method named eigenvalue decomposition based modified Newton algorithm is presented, which first takes eigenvalue decomposition on the Hessian matrix, then replaces the negative eigenvalues with their absolutely values, finally reconstruct Hessian matrix and modify searching direction. The new searching direction is always the descending direction, and the convergence of the algorithm is proved and conclusion on convergence rate is presented qualitatively. At last, a numerical experiment is given for comparing the convergence domains of modified algorithm and classical algorithm.


2013 ◽  
Vol 2013 ◽  
pp. 1-4
Author(s):  
Wen-jun Wang ◽  
Ju-bo Zhu ◽  
Xiao-jun Duan

When the Hessian matrix is not positive, the Newton direction may not be the descending direction. A new method named eigenvalue decomposition-based modified Newton algorithm is presented, which first takes the eigenvalue decomposition of the Hessian matrix, then replaces the negative eigenvalues with their absolute values, and finally reconstructs the Hessian matrix and modifies the searching direction. The new searching direction is always the descending direction. The convergence of the algorithm is proven and the conclusion on convergence rate is presented qualitatively. Finally, a numerical experiment is given for comparing the convergence domains of the modified algorithm and the classical algorithm.


Sign in / Sign up

Export Citation Format

Share Document