Signal recovery with convex constrained nonlinear monotone equations through conjugate gradient hybrid approach

Author(s):  
Abubakar Sani Halilu ◽  
Arunava Majumder ◽  
Mohammed Yusuf Waziri ◽  
Kabiru Ahmed
Mathematics ◽  
2019 ◽  
Vol 7 (8) ◽  
pp. 693 ◽  
Author(s):  
Abubakar ◽  
Kumam ◽  
Awwal ◽  
Thounthong

In this article, we propose a modified self-adaptive conjugate gradient algorithm for handling nonlinear monotone equations with the constraints being convex. Under some nice conditions, the global convergence of the method was established. Numerical examples reported show that the method is promising and efficient for solving monotone nonlinear equations. In addition, we applied the proposed algorithm to solve sparse signal reconstruction problems.


Author(s):  
Mompati Koorapetse ◽  
P Kaelo ◽  
S Kooepile-Reikeletseng

In this paper, a new modified Perry-type derivative-free projection method for solving large-scale nonlinear monotone equations is presented. The method is developed by combining a modified Perry's conjugate gradient method with the hyperplane projection technique. Global convergence and numerical results of the proposed method are established. Preliminary numerical results show that the proposed method is promising and efficient compared to some existing methods in the literature.


2020 ◽  
Vol 2020 ◽  
pp. 1-12
Author(s):  
Haishan Feng ◽  
Tingting Li

Combining the three-term conjugate gradient method of Yuan and Zhang and the acceleration step length of Andrei with the hyperplane projection method of Solodov and Svaiter, we propose an accelerated conjugate gradient algorithm for solving nonlinear monotone equations in this paper. The presented algorithm has the following properties: (i) All search directions generated by the algorithm satisfy the sufficient descent and trust region properties independent of the line search technique. (ii) A derivative-free search technique is proposed along the direction to obtain the step length αk. (iii) If ϕk=−αkhk−hwkTdk>0, then an acceleration scheme is used to modify the step length in a multiplicative manner and create a point. (iv) If the point satisfies the given condition, then it is the next point; otherwise, the hyperplane projection technique is used to obtain the next point. (v) The global convergence of the proposed algorithm is established under some suitable conditions. Numerical comparisons with other conjugate gradient algorithms show that the accelerated computing scheme is more competitive. In addition, the presented algorithm can also be applied to image restoration.


2014 ◽  
Vol 2014 ◽  
pp. 1-12 ◽  
Author(s):  
San-Yang Liu ◽  
Yuan-Yuan Huang ◽  
Hong-Wei Jiao

Two unified frameworks of some sufficient descent conjugate gradient methods are considered. Combined with the hyperplane projection method of Solodov and Svaiter, they are extended to solve convex constrained nonlinear monotone equations. Their global convergence is proven under some mild conditions. Numerical results illustrate that these methods are efficient and can be applied to solve large-scale nonsmooth equations.


Sign in / Sign up

Export Citation Format

Share Document