kkt point
Recently Published Documents


TOTAL DOCUMENTS

10
(FIVE YEARS 4)

H-INDEX

1
(FIVE YEARS 0)

2022 ◽  
Vol 12 (1) ◽  
pp. 93
Author(s):  
Jutamas Kerdkaew ◽  
Rabian Wangkeeree ◽  
Rattanaporn Wangkeeree

<p style='text-indent:20px;'>In this paper, a robust optimization problem, which features a maximum function of continuously differentiable functions as its objective function, is investigated. Some new conditions for a robust KKT point, which is a robust feasible solution that satisfies the robust KKT condition, to be a global robust optimal solution of the uncertain optimization problem, which may have many local robust optimal solutions that are not global, are established. The obtained conditions make use of underestimators, which were first introduced by Jayakumar and Srisatkunarajah [<xref ref-type="bibr" rid="b1">1</xref>,<xref ref-type="bibr" rid="b2">2</xref>] of the Lagrangian associated with the problem at the robust KKT point. Furthermore, we also investigate the Wolfe type robust duality between the smooth uncertain optimization problem and its uncertain dual problem by proving the sufficient conditions for a weak duality and a strong duality between the deterministic robust counterpart of the primal model and the optimistic counterpart of its dual problem. The results on robust duality theorems are established in terms of underestimators. Additionally, to illustrate or support this study, some examples are presented.</p>


Author(s):  
Zichong Li ◽  
Yangyang Xu

First-order methods (FOMs) have been widely used for solving large-scale problems. A majority of existing works focus on problems without constraint or with simple constraints. Several recent works have studied FOMs for problems with complicated functional constraints. In this paper, we design a novel augmented Lagrangian (AL)–based FOM for solving problems with nonconvex objective and convex constraint functions. The new method follows the framework of the proximal point (PP) method. On approximately solving PP subproblems, it mixes the usage of the inexact AL method (iALM) and the quadratic penalty method, whereas the latter is always fed with estimated multipliers by the iALM. The proposed method achieves the best-known complexity result to produce a near Karush–Kuhn–Tucker (KKT) point. Theoretically, the hybrid method has a lower iteration-complexity requirement than its counterpart that only uses iALM to solve PP subproblems; numerically, it can perform significantly better than a pure-penalty-based method. Numerical experiments are conducted on nonconvex linearly constrained quadratic programs. The numerical results demonstrate the efficiency of the proposed methods over existing ones.


2021 ◽  
Vol 2021 (1) ◽  
Author(s):  
Jueyu Wang ◽  
Chao Gu ◽  
Guoqiang Wang

AbstractRecent studies show that the filter method has good numerical performance for nonlinear complementary problems (NCPs). Their approach is to reformulate an NCP as a constrained optimization solved by filter algorithms. However, they can only prove that the iterative sequence converges to the KKT point of the constrained optimization. In this paper, we investigate the relation between the KKT point of the constrained optimization and the solution of the NCP. First, we give several sufficient conditions under which the KKT point of the constrained optimization is the solution of the NCP; second, we define regular conditions and regular point which include and generalize the previous results; third, we prove that the level sets of the objective function of the constrained optimization are bounded for a strongly monotone function or a uniform P-function; finally, we present some examples to verify the previous results.


2020 ◽  
Vol 37 (04) ◽  
pp. 2040018
Author(s):  
Rui Shen ◽  
Zhiqing Meng ◽  
Min Jiang

In this paper, a smoothing partial exact penalty function of biconvex programming is studied. First, concepts of partial KKT point, partial optimum point, partial KKT condition, partial Slater constraint qualification and partial exactness are defined for biconvex programming. It is proved that the partial KKT point is equal to the partial optimum point under the condition of partial Slater constraint qualification and the penalty function of biconvex programming is partially exact if partial KKT condition holds. We prove the error bounds properties between smoothing penalty function and penalty function of biconvex programming when the partial KKT condition holds, as well as the error bounds between objective value of a partial optimum point of smoothing penalty function problem and its [Formula: see text]-feasible solution. So, a partial optimum point of the smoothing penalty function optimization problem is an approximately partial optimum point of biconvex programming. Second, based on the smoothing penalty function, two algorithms are presented for finding a partial optimum or approximate [Formula: see text]-feasible solution to an inequality constrained biconvex optimization and their convergence is proved under some conditions. Finally, numerical experiments show that a satisfactory approximate solution can be obtained by the proposed algorithm.


2014 ◽  
Vol 651-653 ◽  
pp. 2107-2111
Author(s):  
Ai Fen Feng ◽  
Li Ming Zhang ◽  
Zhen Xia Xue

This paper, a new class of augmented Lagrange functions with the new NCP function is proposed for the minimization of a smooth function subject to inequality constraints. Under some conditions, we prove of the equivalences of the KKT point and local point and globe point between primal constrained nonlinear programming problem and the new unconstrained problem. By the character of augmented Lagrange function, the algorithm which uses alternating direction method is constructed and proved convergence.


2012 ◽  
Vol 2012 ◽  
pp. 1-12 ◽  
Author(s):  
Bingzhuang Liu ◽  
Changyu Wang ◽  
Wenling Zhao

We consider a smooth penalty algorithm to solve nonconvex optimization problem based on a family of smooth functions that approximate the usual exact penalty function. At each iteration in the algorithm we only need to find a stationary point of the smooth penalty function, so the difficulty of computing the global solution can be avoided. Under a generalized Mangasarian-Fromovitz constraint qualification condition (GMFCQ) that is weaker and more comprehensive than the traditional MFCQ, we prove that the sequence generated by this algorithm will enter the feasible solution set of the primal problem after finite times of iteration, and if the sequence of iteration points has an accumulation point, then it must be a Karush-Kuhn-Tucker (KKT) point. Furthermore, we obtain better convergence for convex optimization problem.


2011 ◽  
Vol 2011 ◽  
pp. 1-14 ◽  
Author(s):  
Jinchuan Zhou ◽  
Xunzhi Zhu ◽  
Lili Pan ◽  
Wenling Zhao

New convergence properties of the proximal augmented Lagrangian method is established for continuous nonconvex optimization problem with both equality and inequality constrains. In particular, the multiplier sequences are not required to be bounded. Different convergence results are discussed dependent on whether the iterative sequence{xk}generated by algorithm is convergent or divergent. Furthermore, under certain convexity assumption, we show that every accumulation point of{xk}is either a degenerate point or a KKT point of the primal problem. Numerical experiments are presented finally.


Sign in / Sign up

Export Citation Format

Share Document