An inner-point modification of PSO for constrained optimization

2015 ◽  
Vol 32 (7) ◽  
pp. 2005-2019 ◽  
Author(s):  
Daniele Peri

Purpose – The purpose of this paper is to propose a modification of the original PSO algorithm in order to avoid the evaluation of the objective function outside the feasible set, improving the parallel performances of the algorithm in the view of its application on parallel architectures. Design/methodology/approach – Classical PSO iteration is repeated for each particle until a feasible point is found: the global search is performed by a set of independent sub-iteration, at the particle level, and the evaluation of the objective function is performed only once the full swarm is feasible. After that, the main attractors are updated and a new sub-iteration is initiated. Findings – While the main qualities of PSO are preserved, a great advantage in terms of identification of the feasible region and detection of the best feasible solution is obtained. Furthermore, the parallel structure of the algorithm is preserved, and the load balance improved. The results of the application to real-life optimization problems, where constraint satisfaction sometime represents a problem itself, gives the measure of this advantage: an improvement of about 10 percent of the optimal solution is obtained by using the modified version of the algorithm, with a more precise identification of the optimal design variables. Originality/value – Differently from the standard approach, utilizing a penalty function in order to discharge unfeasible points, here only feasible points are produced, improving the exploration of the feasible region and preserving the parallel structure of the algorithm.

2021 ◽  
Vol 11 (2) ◽  
pp. 227-238
Author(s):  
V.P. Ofitserov ◽  

A new approach to the formulation and solution of optimization problems of linear and nonlinear type is stated in this article. The problem statement under consideration differs from the classical linear programming problem of the opti-mal distribution of limited resources between given processes by the need to choose a limited number of processes from a certain finite set and allocate resources over these processes. The goal is to obtain the optimal value of the objective function in relation to other options for choosing the number of processes from the same set and the distribution of resources between them. The objective function can be either linear or non-linear. A nonlinear function must have cer-tain properties for the correct operation of the proposed algorithm for finding the optimal solution. The described method is based on the development of Bellman's ideas of dynamic programming. The proofs of the optimality of the obtained solutions are provided. The article gives an estimate of the computational complexity of the algorithm and a comparison with classical methods for solving the problems under consideration. The types of applied problems solved using the proposed method are characterized. Computer implementations of the described algorithm can be used in automated decision support systems.


2016 ◽  
Vol 38 (4) ◽  
pp. 307-317
Author(s):  
Pham Hoang Anh

In this paper, the optimal sizing of truss structures is solved using a novel evolutionary-based optimization algorithm. The efficiency of the proposed method lies in the combination of global search and local search, in which the global move is applied for a set of random solutions whereas the local move is performed on the other solutions in the search population. Three truss sizing benchmark problems with discrete variables are used to examine the performance of the proposed algorithm. Objective functions of the optimization problems are minimum weights of the whole truss structures and constraints are stress in members and displacement at nodes. Here, the constraints and objective function are treated separately so that both function and constraint evaluations can be saved. The results show that the new algorithm can find optimal solution effectively and it is competitive with some recent metaheuristic algorithms in terms of number of structural analyses required.


2017 ◽  
Vol 8 (3) ◽  
pp. 1-23 ◽  
Author(s):  
Ghanshyam Tejani ◽  
Vimal Savsani ◽  
Vivek Patel

In this study, a modified heat transfer search (MHTS) algorithm is proposed by incorporating sub-population based simultaneous heat transfer modes viz. conduction, convection, and radiation in the basic HTS algorithm. However, the basic HTS algorithm considers only one of the modes of heat transfer for each generation. The multiple natural frequency constraints in truss optimization problems can improve the dynamic behavior of the structure and prevent undesirable vibrations. However, shape and size variables subjected to frequency constraints are difficult to handle due to the complexity of its feasible region, which is non-linear, non-convex, implicit, and often converging to the local optimal solution. The viability and effectiveness of the HTS and MHTS algorithms are investigated by six standard trusses problems. The solutions illustrate that the MHTS algorithm performs better than the HTS algorithm.


2019 ◽  
Vol 61 (4) ◽  
pp. 177-185
Author(s):  
Moritz Mühlenthaler ◽  
Alexander Raß

Abstract A discrete particle swarm optimization (PSO) algorithm is a randomized search heuristic for discrete optimization problems. A fundamental question about randomized search heuristics is how long it takes, in expectation, until an optimal solution is found. We give an overview of recent developments related to this question for discrete PSO algorithms. In particular, we give a comparison of known upper and lower bounds of expected runtimes and briefly discuss the techniques used to obtain these bounds.


2019 ◽  
Vol 25 (1) ◽  
pp. 54-64 ◽  
Author(s):  
Sudhanshu Aggarwal

PurposeThe purpose of this paper is to present an efficient heuristic algorithm based on the 3-neighborhood approach. In this paper, search is made from sides of both feasible and infeasible regions to find near-optimal solutions.Design/methodology/approachThe algorithm performs a series of selection and exchange operations in 3-neighborhood to see whether this exchange yields still an improved feasible solution or converges to a near-optimal solution in which case the algorithm stops.FindingsThe proposed algorithm has been tested on complex system structures which have been widely used. The results show that this 3-neighborhood approach not only can obtain various known solutions but also is computationally efficient for various complex systems.Research limitations/implicationsIn general, the proposed heuristic is applicable to any coherent system with no restrictions on constraint functions; however, to enforce convergence, inferior solutions might be included only when they are not being too far from the optimum.Practical implicationsIt is observed that the proposed heuristic is reasonably proficient in terms of various measures of performance and computational time.Social implicationsReliability optimization is very important in real life systems such as computer and communication systems, telecommunications, automobile, nuclear, defense systems, etc. It is an important issue prior to real life systems design.Originality/valueThe utilization of 3-neighborhood strategy seems to be encouraging as it efficiently enforces the convergence to a near-optimal solution; indeed, it attains quality solutions in less computational time in comparison to other existing heuristic algorithms.


Author(s):  
Shafiullah Khan ◽  
Shiyou Yang ◽  
Obaid Ur Rehman

Purpose The aim of this paper is to explore the potential of particle swarm optimization (PSO) algorithm to solve an electromagnetic inverse problem. Design/methodology/approach A modified PSO algorithm is designed. Findings The modified PSO algorithm is a more stable, robust and efficient global optimizer for solving the well-known benchmark optimization problems. The new mutation approach preserves the diversity of the population, whereas the proposed dynamic and adaptive parameters maintain a good balance between the exploration and exploitation searches. The numerically experimental results of two case studies demonstrate the merits of the proposed algorithm. Originality/value Some improvements, such as the design of a new global mutation mechanism and introducing a novel strategy for learning and control parameters, are proposed.


2013 ◽  
Vol 816-817 ◽  
pp. 1154-1157
Author(s):  
Xu Yin ◽  
Ai Min Ji

To solve problems that exist in optimal design such as falling into local optimal solution easily and low efficiency in collaborative optimization, a new mix strategy optimization method combined design of experiments (DOE) with gradient optimization (GO) was proposed. In order to reduce the effect on the result of optimization made by the designers decision, DOE for preliminary analysis of the function model was used, and the optimal values obtained in DOE stage was taken as the initial values of design variables in GO stage in the new optimization method. The reducer MDO problem was taken as a example to confirm the global degree, efficiency, and accuracy of the method. The results show the optimization method could not only avoid falling into local solution, but also have an obvious superiority in treating the complex collaborative optimization problems.


Author(s):  
Haopeng Zhang ◽  
Qing Hui

Model predictive control (MPC) is a heuristic control strategy to find a consequence of best controllers during each finite-horizon regarding to certain performance functions of a dynamic system. MPC involves two main operations: estimation and optimization. Due to high complexity of the performance functions, such as, nonlinear, non-convex, large-scale objective functions, the optimization algorithms for MPC must be capable of handling those problems with both computational efficiency and accuracy. Multiagent coordination optimization (MCO) is a recently developed heuristic algorithm by embedding multiagent coordination into swarm intelligence to accelerate the searching process for the optimal solution in the particle swarm optimization (PSO) algorithm. With only some elementary operations, the MCO algorithm can obtain the best solution extremely fast, which is especially necessary to solve the online optimization problems in MPC. Therefore, in this paper, we propose an MCO based MPC strategy to enhance the performance of the MPC controllers when addressing non-convex large-scale nonlinear problems. Moreover, as an application, the network resource balanced allocation problem is numerically illustrated by the MCO based MPC strategy.


2011 ◽  
Vol 383-390 ◽  
pp. 7208-7213
Author(s):  
De Kun Tan

To overcome the shortage of standard Particle Swarm Optimization(SPSO) on premature convergence, Quantum-behaved Particle Swarm Optimization (QPSO) is presented to solve engineering constrained optimization problem. QPSO algorithm is a novel PSO algorithm model in terms of quantum mechanics. The model is based on Delta potential, and we think the particle has the behavior of quanta. Because the particle doesn’t have a certain trajectory, it has more randomicity than the particle which has fixed path in PSO, thus the QPSO more easily escapes from local optima, and has more capability to seek the global optimal solution. In the period of iterative optimization, outside point method is used to deal with those particles that violate the constraints. Furthermore, compared with other intelligent algorithms, the QPSO is verified by two instances of engineering constrained optimization, experimental results indicate that the algorithm performs better in terms of accuracy and robustness.


2015 ◽  
Vol 2015 ◽  
pp. 1-7
Author(s):  
Sakineh Tahmasebzadeh ◽  
Hamidreza Navidi ◽  
Alaeddin Malek

This paper proposes three numerical algorithms based on Karmarkar’s interior point technique for solving nonlinear convex programming problems subject to linear constraints. The first algorithm uses the Karmarkar idea and linearization of the objective function. The second and third algorithms are modification of the first algorithm using the Schrijver and Malek-Naseri approaches, respectively. These three novel schemes are tested against the algorithm of Kebiche-Keraghel-Yassine (KKY). It is shown that these three novel algorithms are more efficient and converge to the correct optimal solution, while the KKY algorithm fails in some cases. Numerical results are given to illustrate the performance of the proposed algorithms.


Sign in / Sign up

Export Citation Format

Share Document