A Population-Based Hybrid Extremal Optimization Algorithm

Author(s):  
Yu Chen ◽  
Kai Zhang ◽  
Xiufen Zou
2016 ◽  
Vol 330 ◽  
pp. 49-73 ◽  
Author(s):  
Guo-Qiang Zeng ◽  
Jie Chen ◽  
Li-Min Li ◽  
Min-Rong Chen ◽  
Lie Wu ◽  
...  

Mathematics ◽  
2021 ◽  
Vol 9 (11) ◽  
pp. 1190
Author(s):  
Mohammad Dehghani ◽  
Zeinab Montazeri ◽  
Štěpán Hubálovský

There are many optimization problems in the different disciplines of science that must be solved using the appropriate method. Population-based optimization algorithms are one of the most efficient ways to solve various optimization problems. Population-based optimization algorithms are able to provide appropriate solutions to optimization problems based on a random search of the problem-solving space without the need for gradient and derivative information. In this paper, a new optimization algorithm called the Group Mean-Based Optimizer (GMBO) is presented; it can be applied to solve optimization problems in various fields of science. The main idea in designing the GMBO is to use more effectively the information of different members of the algorithm population based on two selected groups, with the titles of the good group and the bad group. Two new composite members are obtained by averaging each of these groups, which are used to update the population members. The various stages of the GMBO are described and mathematically modeled with the aim of being used to solve optimization problems. The performance of the GMBO in providing a suitable quasi-optimal solution on a set of 23 standard objective functions of different types of unimodal, high-dimensional multimodal, and fixed-dimensional multimodal is evaluated. In addition, the optimization results obtained from the proposed GMBO were compared with eight other widely used optimization algorithms, including the Marine Predators Algorithm (MPA), the Tunicate Swarm Algorithm (TSA), the Whale Optimization Algorithm (WOA), the Grey Wolf Optimizer (GWO), Teaching–Learning-Based Optimization (TLBO), the Gravitational Search Algorithm (GSA), Particle Swarm Optimization (PSO), and the Genetic Algorithm (GA). The optimization results indicated the acceptable performance of the proposed GMBO, and, based on the analysis and comparison of the results, it was determined that the GMBO is superior and much more competitive than the other eight algorithms.


2021 ◽  
Vol 31 (1) ◽  
pp. 70-94
Author(s):  
Jeffrey O. Agushaka ◽  
Absalom E. Ezugwu

Abstract Arithmetic optimization algorithm (AOA) is one of the recently proposed population-based metaheuristic algorithms. The algorithmic design concept of the AOA is based on the distributive behavior of arithmetic operators, namely, multiplication (M), division (D), subtraction (S), and addition (A). Being a new metaheuristic algorithm, the need for a performance evaluation of AOA is significant to the global optimization research community and specifically to nature-inspired metaheuristic enthusiasts. This article aims to evaluate the influence of the algorithm control parameters, namely, population size and the number of iterations, on the performance of the newly proposed AOA. In addition, we also investigated and validated the influence of different initialization schemes available in the literature on the performance of the AOA. Experiments were conducted using different initialization scenarios and the first is where the population size is large and the number of iterations is low. The second scenario is when the number of iterations is high, and the population size is small. Finally, when the population size and the number of iterations are similar. The numerical results from the conducted experiments showed that AOA is sensitive to the population size and requires a large population size for optimal performance. Afterward, we initialized AOA with six initialization schemes, and their performances were tested on the classical functions and the functions defined in the CEC 2020 suite. The results were presented, and their implications were discussed. Our results showed that the performance of AOA could be influenced when the solution is initialized with schemes other than default random numbers. The Beta distribution outperformed the random number distribution in all cases for both the classical and CEC 2020 functions. The performance of uniform distribution, Rayleigh distribution, Latin hypercube sampling, and Sobol low discrepancy sequence are relatively competitive with the Random number. On the basis of our experiments’ results, we recommend that a solution size of 6,000, the number of iterations of 100, and initializing the solutions with Beta distribution will lead to AOA performing optimally for scenarios considered in our experiments.


2017 ◽  
Vol 7 (04) ◽  
pp. 1
Author(s):  
Srividya Ravindra Kumar ◽  
Ciji Pearl Kurian ◽  
Marcos Eduardo Gomes-Borges

Sign in / Sign up

Export Citation Format

Share Document