scholarly journals Expansion of Particle Multi-Swarm Optimization

2018 ◽  
Vol 7 (2) ◽  
pp. 74 ◽  
Author(s):  
Hiroshi Sho

For improving the search ability and performance of elementary multiple particle swarm optimizers, we, in this paper, propose a series of multiple particle swarm optimizers with information sharing by introducing a special strategy,called multi-swarm information sharing. The key idea, here, is to add a special confidence term into the updating rule of the particle's velocity by the best solution found out by the particle multi-swarm search. This is a new type approach for the technical development and evolution of particle multi-swarm optimization itself. In order to confirm the effectiveness of the information sharing strategy in the proposed particle multi-swarm search, several computer experiments of dealing with a suite of benchmark problems are carried out. For investigating the performance and efficiency of these proposed methods, we compare their search ability and performance, respectively. The obtained experimental results show that the proposed methods have better search ability and performance than those methods without the strategy. And we still decide the best value of adding the new confidence coefficient to the multi-swarm for dealing with the given optimization problems.

Author(s):  
Jiten Makadia ◽  
C.D. Sankhavara

Swarm Intelligence algorithms like PSO (Particle Swarm Optimization), ACO (Ant Colony Optimization), ABC (Artificial Bee Colony), Glow-worm swarm Optimization, etc. have been utilized by researchers for solving optimization problems. This work presents the application of a novel modified EHO (Elephant Herding Optimization) for cost optimization of shell and tube heat exchanger. A comparison of the results obtained by EHO in two benchmark problems shows that it is superior to those obtained with genetic algorithm and particle swarm optimization. The overall cost reduction is 13.3 % and 9.68% for both the benchmark problem compared to PSO. Results indicate that EHO can be effectively utilized for solving real-life optimization problems.


Author(s):  
T. O. Ting ◽  
H. C. Ting ◽  
T. S. Lee

In this work, a hybrid Taguchi-Particle Swarm Optimization (TPSO) is proposed to solve global numerical optimization problems with continuous and discrete variables. This hybrid algorithm combines the well-known Particle Swarm Optimization Algorithm with the established Taguchi method, which has been an important tool for robust design. This paper presents the improvements obtained despite the simplicity of the hybridization process. The Taguchi method is run only once in every PSO iteration and therefore does not give significant impact in terms of computational cost. The method creates a more diversified population, which also contributes to the success of avoiding premature convergence. The proposed method is effectively applied to solve 13 benchmark problems. This study’s results show drastic improvements in comparison with the standard PSO algorithm involving continuous and discrete variables on high dimensional benchmark functions.


2020 ◽  
Vol 2020 ◽  
pp. 1-26
Author(s):  
Wusi Yang ◽  
Li Chen ◽  
Yi Wang ◽  
Maosheng Zhang

The recently proposed multiobjective particle swarm optimization algorithm based on competition mechanism algorithm cannot effectively deal with many-objective optimization problems, which is characterized by relatively poor convergence and diversity, and long computing runtime. In this paper, a novel multi/many-objective particle swarm optimization algorithm based on competition mechanism is proposed, which maintains population diversity by the maximum and minimum angle between ordinary and extreme individuals. And the recently proposed θ-dominance is adopted to further enhance the performance of the algorithm. The proposed algorithm is evaluated on the standard benchmark problems DTLZ, WFG, and UF1-9 and compared with the four recently proposed multiobjective particle swarm optimization algorithms and four state-of-the-art many-objective evolutionary optimization algorithms. The experimental results indicate that the proposed algorithm has better convergence and diversity, and its performance is superior to other comparative algorithms on most test instances.


2014 ◽  
Vol 2014 ◽  
pp. 1-23 ◽  
Author(s):  
Martins Akugbe Arasomwan ◽  
Aderemi Oluyinka Adewumi

A new local search technique is proposed and used to improve the performance of particle swarm optimization algorithms by addressing the problem of premature convergence. In the proposed local search technique, a potential particle position in the solution search space is collectively constructed by a number of randomly selected particles in the swarm. The number of times the selection is made varies with the dimension of the optimization problem and each selected particle donates the value in the location of its randomly selected dimension from its personal best. After constructing the potential particle position, some local search is done around its neighbourhood in comparison with the current swarm global best position. It is then used to replace the global best particle position if it is found to be better; otherwise no replacement is made. Using some well-studied benchmark problems with low and high dimensions, numerical simulations were used to validate the performance of the improved algorithms. Comparisons were made with four different PSO variants, two of the variants implement different local search technique while the other two do not. Results show that the improved algorithms could obtain better quality solution while demonstrating better convergence velocity and precision, stability, robustness, and global-local search ability than the competing variants.


2018 ◽  
Vol 2018 ◽  
pp. 1-17 ◽  
Author(s):  
Xueying Lv ◽  
Yitian Wang ◽  
Junyi Deng ◽  
Guanyu Zhang ◽  
Liu Zhang

In this study, an improved eliminate particle swarm optimization (IEPSO) is proposed on the basis of the last-eliminated principle to solve optimization problems in engineering design. During optimization, the IEPSO enhances information communication among populations and maintains population diversity to overcome the limitations of classical optimization algorithms in solving multiparameter, strong coupling, and nonlinear engineering optimization problems. These limitations include advanced convergence and the tendency to easily fall into local optimization. The parameters involved in the imported “local-global information sharing” term are analyzed, and the principle of parameter selection for performance is determined. The performances of the IEPSO and classical optimization algorithms are then tested by using multiple sets of classical functions to verify the global search performance of the IEPSO. The simulation test results and those of the improved classical optimization algorithms are compared and analyzed to verify the advanced performance of the IEPSO algorithm.


2020 ◽  
Vol 9 (1) ◽  
pp. 54
Author(s):  
Hiroshi Sho

The purpose of this study is to clarify the search performance of differential evolution (DE) and particle swarm optimization (PSO) technologies for instinctively understanding the specificity of the used search methods. Due to achieve the task, here, the several search methods of both, i.e. DE/rand/1, DE/rand/2, DE/best/1, DE/best/2, the PSO, PSOIW, and CPSO, which are implemented in this paper. Therefore, many computer experiments are carried out for handling the given four benchmark problems. Through the analysis of the obtained experimental data, the detail search performance and characteristics of them are observed and compared, respectively. From the obtained results, it is found that the search methods of DE/best/1 and the PSO relatively have better search performance. Based on the findings and know-how, they can provide some important reference and key hint for encouraging development and improvement of both DE and PSO technologies in the near future. And as the applicative examples, the PSO is used to handle typical 2-bit and 3-bit parity problems for pattern classification.


2014 ◽  
Vol 4 (3) ◽  
pp. 189-204 ◽  
Author(s):  
Simone A. Ludwig

Abstract Adaptive Particle Swarm Optimization (PSO) variants have become popular in recent years. The main idea of these adaptive PSO variants is that they adaptively change their search behavior during the optimization process based on information gathered during the run. Adaptive PSO variants have shown to be able to solve a wide range of difficult optimization problems efficiently and effectively. In this paper we propose a Repulsive Self-adaptive Acceleration PSO (RSAPSO) variant that adaptively optimizes the velocity weights of every particle at every iteration. The velocity weights include the acceleration constants as well as the inertia weight that are responsible for the balance between exploration and exploitation. Our proposed RSAPSO variant optimizes the velocity weights that are then used to search for the optimal solution of the problem (e.g., benchmark function). We compare RSAPSO to four known adaptive PSO variants (decreasing weight PSO, time-varying acceleration coefficients PSO, guaranteed convergence PSO, and attractive and repulsive PSO) on twenty benchmark problems. The results show that RSAPSO achives better results compared to the known PSO variants on difficult optimization problems that require large numbers of function evaluations.


2016 ◽  
Vol 2016 ◽  
pp. 1-10 ◽  
Author(s):  
Alan Díaz-Manríquez ◽  
Gregorio Toscano ◽  
Jose Hugo Barron-Zambrano ◽  
Edgar Tello-Leal

We propose to couple theR2performance measure and Particle Swarm Optimization in order to handle multi/many-objective problems. Our proposal shows that through a well-designed interaction process we could maintain the metaheuristic almost inalterable and through theR2performance measure we did not use neither an external archive nor Pareto dominance to guide the search. The proposed approach is validated using several test problems and performance measures commonly adopted in the specialized literature. Results indicate that the proposed algorithm produces results that are competitive with respect to those obtained by four well-known MOEAs. Additionally, we validate our proposal in many-objective optimization problems. In these problems, our approach showed its main strength, since it could outperform another well-known indicator-based MOEA.


2010 ◽  
Vol 1 (2) ◽  
pp. 18-33
Author(s):  
T. O. Ting ◽  
H. C. Ting ◽  
T. S. Lee

In this work, a hybrid Taguchi-Particle Swarm Optimization (TPSO) is proposed to solve global numerical optimization problems with continuous and discrete variables. This hybrid algorithm combines the well-known Particle Swarm Optimization Algorithm with the established Taguchi method, which has been an important tool for robust design. This paper presents the improvements obtained despite the simplicity of the hybridization process. The Taguchi method is run only once in every PSO iteration and therefore does not give significant impact in terms of computational cost. The method creates a more diversified population, which also contributes to the success of avoiding premature convergence. The proposed method is effectively applied to solve 13 benchmark problems. This study’s results show drastic improvements in comparison with the standard PSO algorithm involving continuous and discrete variables on high dimensional benchmark functions.


Sign in / Sign up

Export Citation Format

Share Document