scholarly journals REVIEW OF THE MULTI-OBJECTIVE SWARM INTELLIGENCE OPTIMIZATION ALGORITHMS

2021 ◽  
Vol 20 (Number 2) ◽  
pp. 171-211
Author(s):  
Shaymah Akram Yasear ◽  
Ku Ruhana Ku-Mahamud

Multi-objective swarm intelligence (MOSI) metaheuristics were proposed to solve multi-objective optimization problems (MOPs) that consists of two or more conflict objectives, in which improving an objective leads to the degradation of the other. The MOSI algorithms are based on the integration of single objective algorithms and multi-objective optimization (MOO) approach. The MOO approaches include scalarization, Pareto dominance, decomposition and indicator-based. In this paper, the status of MOO research and state-of-the-art MOSI algorithms namely, multi-objective particle swarm, artificial bee colony, firefly algorithm, bat algorithm, gravitational search algorithm, grey wolf optimizer, bacterial foraging and moth-flame optimization algorithms have been reviewed. These reviewed algorithms were mainly developed to solve continuous MOPs. The review is based on how the algorithms deal with objective functions using MOO approaches, the benchmark MOPs used in the evaluation and performance metrics. Furthermore, it describes the advantages and disadvantages of each MOO approach and provides some possible future research directions in this area. The results show that several MOO approaches have not been used in most of the proposed MOSI algorithms. Integrating other different MOO approaches may help in developing more effective optimization algorithms, especially in solving complex MOPs. Furthermore, most of the MOSI algorithms have been evaluated using MOPs with two objectives, which clarifies open issues in this research area.

Mathematics ◽  
2021 ◽  
Vol 9 (11) ◽  
pp. 1190
Author(s):  
Mohammad Dehghani ◽  
Zeinab Montazeri ◽  
Štěpán Hubálovský

There are many optimization problems in the different disciplines of science that must be solved using the appropriate method. Population-based optimization algorithms are one of the most efficient ways to solve various optimization problems. Population-based optimization algorithms are able to provide appropriate solutions to optimization problems based on a random search of the problem-solving space without the need for gradient and derivative information. In this paper, a new optimization algorithm called the Group Mean-Based Optimizer (GMBO) is presented; it can be applied to solve optimization problems in various fields of science. The main idea in designing the GMBO is to use more effectively the information of different members of the algorithm population based on two selected groups, with the titles of the good group and the bad group. Two new composite members are obtained by averaging each of these groups, which are used to update the population members. The various stages of the GMBO are described and mathematically modeled with the aim of being used to solve optimization problems. The performance of the GMBO in providing a suitable quasi-optimal solution on a set of 23 standard objective functions of different types of unimodal, high-dimensional multimodal, and fixed-dimensional multimodal is evaluated. In addition, the optimization results obtained from the proposed GMBO were compared with eight other widely used optimization algorithms, including the Marine Predators Algorithm (MPA), the Tunicate Swarm Algorithm (TSA), the Whale Optimization Algorithm (WOA), the Grey Wolf Optimizer (GWO), Teaching–Learning-Based Optimization (TLBO), the Gravitational Search Algorithm (GSA), Particle Swarm Optimization (PSO), and the Genetic Algorithm (GA). The optimization results indicated the acceptable performance of the proposed GMBO, and, based on the analysis and comparison of the results, it was determined that the GMBO is superior and much more competitive than the other eight algorithms.


Author(s):  
Xiaohui Yuan ◽  
Zhihuan Chen ◽  
Yanbin Yuan ◽  
Yuehua Huang ◽  
Xiaopan Zhang

A novel strength Pareto gravitational search algorithm (SPGSA) is proposed to solve multi-objective optimization problems. This SPGSA algorithm utilizes the strength Pareto concept to assign the fitness values for agents and uses a fine-grained elitism selection mechanism to keep the population diversity. Furthermore, the recombination operators are modeled in this approach to decrease the possibility of trapping in local optima. Experiments are conducted on a series of benchmark problems that are characterized by difficulties in local optimality, nonuniformity, and nonconvexity. The results show that the proposed SPGSA algorithm performs better in comparison with other related works. On the other hand, the effectiveness of two subtle means added to the GSA are verified, i.e. the fine-grained elitism selection and the use of SBX and PMO operators. Simulation results show that these measures not only improve the convergence ability of original GSA, but also preserve the population diversity adequately, which enables the SPGSA algorithm to have an excellent ability that keeps a desirable balance between the exploitation and exploration so as to accelerate the convergence speed to the true Pareto-optimal front.


Author(s):  
Jesper Kristensen ◽  
You Ling ◽  
Isaac Asher ◽  
Liping Wang

Adaptive sampling methods have been used to build accurate meta-models across large design spaces from which engineers can explore data trends, investigate optimal designs, study the sensitivity of objectives on the modeling design features, etc. For global design optimization applications, adaptive sampling methods need to be extended to sample more efficiently near the optimal domains of the design space (i.e., the Pareto front/frontier in multi-objective optimization). Expected Improvement (EI) methods have been shown to be efficient to solve design optimization problems using meta-models by incorporating prediction uncertainty. In this paper, a set of state-of-the-art methods (hypervolume EI method and centroid EI method) are presented and implemented for selecting sampling points for multi-objective optimizations. The classical hypervolume EI method uses hyperrectangles to represent the Pareto front, which shows undesirable behavior at the tails of the Pareto front. This issue is addressed utilizing the concepts from physical programming to shape the Pareto front. The modified hypervolume EI method can be extended to increase local Pareto front accuracy in any area identified by an engineer, and this method can be applied to Pareto frontiers of any shape. A novel hypervolume EI method is also developed that does not rely on the assumption of hyperrectangles, but instead assumes the Pareto frontier can be represented by a convex hull. The method exploits fast methods for convex hull construction and numerical integration, and results in a Pareto front shape that is desired in many practical applications. Various performance metrics are defined in order to quantitatively compare and discuss all methods applied to a particular 2D optimization problem from the literature. The modified hypervolume EI methods lead to dramatic resource savings while improving the predictive capabilities near the optimal objective values.


2022 ◽  
Vol 54 (8) ◽  
pp. 1-34
Author(s):  
Ye Tian ◽  
Langchun Si ◽  
Xingyi Zhang ◽  
Ran Cheng ◽  
Cheng He ◽  
...  

Multi-objective evolutionary algorithms (MOEAs) have shown promising performance in solving various optimization problems, but their performance may deteriorate drastically when tackling problems containing a large number of decision variables. In recent years, much effort been devoted to addressing the challenges brought by large-scale multi-objective optimization problems. This article presents a comprehensive survey of stat-of-the-art MOEAs for solving large-scale multi-objective optimization problems. We start with a categorization of these MOEAs into decision variable grouping based, decision space reduction based, and novel search strategy based MOEAs, discussing their strengths and weaknesses. Then, we review the benchmark problems for performance assessment and a few important and emerging applications of MOEAs for large-scale multi-objective optimization. Last, we discuss some remaining challenges and future research directions of evolutionary large-scale multi-objective optimization.


2011 ◽  
Vol 20 (01) ◽  
pp. 209-219 ◽  
Author(s):  
MOHAMMAD HAMDAN

Polynomial mutation is widely used in evolutionary optimization algorithms as a variation operator. In previous work on the use of evolutionary algorithms for solving multi-objective problems, two versions of polynomial mutations were introduced. The first is non-highly disruptive that is not prone to local optima and the second is highly disruptive polynomial mutation. This paper looks at the two variants and proposes a dynamic version of polynomial mutation. The experimental results show that the proposed adaptive algorithm is doing well for three evolutionary multiobjective algorithms on well known multiobjective optimization problems in terms of convergence speed, generational distance and hypervolume performance metrics.


2020 ◽  
pp. 1573-1593
Author(s):  
Kamel Zeltni ◽  
Souham Meshoul ◽  
Heyam H. Al-Baity

This article reviews existing constraint-handling techniques then presents a new design for Swarm Intelligence Metaheuristics (SIM) to deal with constrained multi-objective optimization problems (CMOPs). This new design aims to investigate potential effects of leader concepts that characterize the dynamic of SIM in the hope to help the population to reach Pareto optimal solutions in a constrained search space. The new leader-based constraint handling mechanism is incorporated in Constrained Multi-Objective Cuckoo Search (C-MOCS) and Constrained Multi-Objective Particle Swarm Optimization (C-MOPSO) as specific instances of a more general class of SIMs. The experimental results are carried out using a set of six well-known test functions and two performance metrics. The convergence and diversity of C-MOCS and C-MOPSO are analysed and compared to the well-known Multi-Objective Evolutionary Algorithm (MOEA) NSGA-II and discussed based on the obtained results.


2012 ◽  
Vol 3 (3) ◽  
pp. 32-49 ◽  
Author(s):  
Hadi Nobahari ◽  
Mahdi Nikusokhan ◽  
Patrick Siarry

This paper proposes an extension of the Gravitational Search Algorithm (GSA) to multi-objective optimization problems. The new algorithm, called Non-dominated Sorting GSA (NSGSA), utilizes the non-dominated sorting concept to update the gravitational acceleration of the particles. An external archive is also used to store the Pareto optimal solutions and to provide some elitism. It also guides the search toward the non-crowding and the extreme regions of the Pareto front. A new criterion is proposed to update the external archive and two new mutation operators are also proposed to promote the diversity within the swarm. Numerical results show that NSGSA can obtain comparable and even better performances as compared to the previous multi-objective variant of GSA and some other multi-objective optimization algorithms.


Sign in / Sign up

Export Citation Format

Share Document