scholarly journals Global optimality conditions and duality theorems for robust optimal solutions of optimization problems with data uncertainty, using underestimators

2022 ◽  
Vol 12 (1) ◽  
pp. 93
Author(s):  
Jutamas Kerdkaew ◽  
Rabian Wangkeeree ◽  
Rattanaporn Wangkeeree

<p style='text-indent:20px;'>In this paper, a robust optimization problem, which features a maximum function of continuously differentiable functions as its objective function, is investigated. Some new conditions for a robust KKT point, which is a robust feasible solution that satisfies the robust KKT condition, to be a global robust optimal solution of the uncertain optimization problem, which may have many local robust optimal solutions that are not global, are established. The obtained conditions make use of underestimators, which were first introduced by Jayakumar and Srisatkunarajah [<xref ref-type="bibr" rid="b1">1</xref>,<xref ref-type="bibr" rid="b2">2</xref>] of the Lagrangian associated with the problem at the robust KKT point. Furthermore, we also investigate the Wolfe type robust duality between the smooth uncertain optimization problem and its uncertain dual problem by proving the sufficient conditions for a weak duality and a strong duality between the deterministic robust counterpart of the primal model and the optimistic counterpart of its dual problem. The results on robust duality theorems are established in terms of underestimators. Additionally, to illustrate or support this study, some examples are presented.</p>

2021 ◽  
Vol 2 (3) ◽  
Author(s):  
Najeeb Abdulaleem

AbstractIn this paper, a class of E-differentiable vector optimization problems with both inequality and equality constraints is considered. The so-called vector mixed E-dual problem is defined for the considered E-differentiable vector optimization problem with both inequality and equality constraints. Then, several mixed E-duality theorems are established under (generalized) V-E-invexity hypotheses.


Mathematics ◽  
2019 ◽  
Vol 7 (4) ◽  
pp. 372
Author(s):  
Liu He ◽  
Qi-Lin Wang ◽  
Ching-Feng Wen ◽  
Xiao-Yan Zhang ◽  
Xiao-Bing Li

In this paper, we introduce the notion of higher-order weak adjacent epiderivative for a set-valued map without lower-order approximating directions and obtain existence theorem and some properties of the epiderivative. Then by virtue of the epiderivative and Benson proper efficiency, we establish the higher-order Mond-Weir type dual problem for a set-valued optimization problem and obtain the corresponding weak duality, strong duality and converse duality theorems, respectively.


Author(s):  
Izhar Ahmad ◽  
Arshpreet Kaur ◽  
Mahesh Kumar Sharma

Robust optimization has come out to be a potent approach to study mathematical problems with data uncertainty. We use robust optimization to study a nonsmooth nonconvex mathematical program over cones with data uncertainty containing generalized convex functions. We study sufficient optimality conditions for the problem. Then we construct its robust dual problem and provide appropriate duality theorems which show the relation between uncertainty problems and their corresponding robust dual problems.


2021 ◽  
Vol 12 (4) ◽  
pp. 81-100
Author(s):  
Yao Peng ◽  
Zepeng Shen ◽  
Shiqi Wang

Multimodal optimization problem exists in multiple global and many local optimal solutions. The difficulty of solving these problems is finding as many local optimal peaks as possible on the premise of ensuring global optimal precision. This article presents adaptive grouping brainstorm optimization (AGBSO) for solving these problems. In this article, adaptive grouping strategy is proposed for achieving adaptive grouping without providing any prior knowledge by users. For enhancing the diversity and accuracy of the optimal algorithm, elite reservation strategy is proposed to put central particles into an elite pool, and peak detection strategy is proposed to delete particles far from optimal peaks in the elite pool. Finally, this article uses testing functions with different dimensions to compare the convergence, accuracy, and diversity of AGBSO with BSO. Experiments verify that AGBSO has great localization ability for local optimal solutions while ensuring the accuracy of the global optimal solutions.


2011 ◽  
Vol 421 ◽  
pp. 559-563
Author(s):  
Yong Chao Gao ◽  
Li Mei Liu ◽  
Heng Qian ◽  
Ding Wang

The scale and complexity of search space are important factors deciding the solving difficulty of an optimization problem. The information of solution space may lead searching to optimal solutions. Based on this, an algorithm for combinatorial optimization is proposed. This algorithm makes use of the good solutions found by intelligent algorithms, contracts the search space and partitions it into one or several optimal regions by backbones of combinatorial optimization solutions. And optimization of small-scale problems is carried out in optimal regions. Statistical analysis is not necessary before or through the solving process in this algorithm, and solution information is used to estimate the landscape of search space, which enhances the speed of solving and solution quality. The algorithm breaks a new path for solving combinatorial optimization problems, and the results of experiments also testify its efficiency.


2019 ◽  
Vol 35 (3) ◽  
pp. 417-426 ◽  
Author(s):  
CHANOKSUDA KHANTREE ◽  
RABIAN WANGKEEREE ◽  
◽  

This paper devotes to the quasi ε-solution for robust semi-infinite optimization problems (RSIP) involving a locally Lipschitz objective function and infinitely many locally Lipschitz constraint functions with data uncertainty. Under the fulfillment of robust type Guignard constraint qualification and robust type Kuhn-Tucker constraint qualification, a necessary condition for a quasi ε-solution to problem (RSIP). After introducing the generalized convexity, we give a sufficient optimality for such a quasi ε-solution to problem (RSIP). Finally, we also establish approximate duality theorems in term of Wolfe type which is formulated in approximate form.


2015 ◽  
Vol 2015 ◽  
pp. 1-11 ◽  
Author(s):  
Tingting Zou

Duality theorem is an attractive approach for solving fuzzy optimization problems. However, the duality gap is generally nonzero for nonconvex problems. So far, most of the studies focus on continuous variables in fuzzy optimization problems. And, in real problems and models, fuzzy optimization problems also involve discrete and mixed variables. To address the above problems, we improve the extended duality theory by adding fuzzy objective functions. In this paper, we first define continuous fuzzy nonlinear programming problems, discrete fuzzy nonlinear programming problems, and mixed fuzzy nonlinear programming problems and then provide the extended dual problems, respectively. Finally we prove the weak and strong extended duality theorems, and the results show no duality gap between the original problem and extended dual problem.


2021 ◽  
Vol 6 (11) ◽  
pp. 12321-12338
Author(s):  
Yanfei Chai ◽  

<abstract><p>This paper deals with the robust strong duality for nonconvex optimization problem with the data uncertainty in constraint. A new weak conjugate function which is abstract convex, is introduced and three kinds of robust dual problems are constructed to the primal optimization problem by employing this weak conjugate function: the robust augmented Lagrange dual, the robust weak Fenchel dual and the robust weak Fenchel-Lagrange dual problem. Characterizations of inequality (1.1) according to robust abstract perturbation weak conjugate duality are established by using the abstract convexity. The results are used to obtain robust strong duality between noncovex uncertain optimization problem and its robust dual problems mentioned above, the optimality conditions for this noncovex uncertain optimization problem are also investigated.</p></abstract>


Author(s):  
Jiantao Liu ◽  
Hae Chang Gea ◽  
Ping An Du

Robust structural design optimization with non-probabilistic uncertainties is often formulated as a two-level optimization problem. The top level optimization problem is simply to minimize a specified objective function while the optimized solution at the second level solution is within bounds. The second level optimization problem is to find the worst case design under non-probabilistic uncertainty. Although the second level optimization problem is a non-convex problem, the global optimal solution must be assured in order to guarantee the solution robustness at the first level. In this paper, a new approach is proposed to solve the robust structural optimization problems with non-probabilistic uncertainties. The WCDO problems at the second level are solved directly by the monotonocity analysis and the global optimality is assured. Then, the robust structural optimization problem is reduced to a single level problem and can be easily solved by any gradient based method. To illustrate the proposed approach, truss examples with non-probabilistic uncertainties on stiffness and loading are presented.


Sign in / Sign up

Export Citation Format

Share Document