Efficient Particle Swarm Optimization Algorithm Based on Affinity Propagation

2014 ◽  
Vol 571-572 ◽  
pp. 191-195
Author(s):  
Lin Ping Su ◽  
Zhao Wang ◽  
Zheng Guan Huang ◽  
Hao Li

Since the 1950s, with the great development of computer technology and bionics, particle swarm optimization (PSO) was raised. The particle swarm optimization mimics the nature biological group behaviors, and has the following advantages compared to classic optimization algorithms: it is a global optimization process and doesn’t depend on the initial state; it can be applied widely without prior knowledge on the optimization problems; the ideas and the implements of PSO are quite simple, the steps are standardization, and it’s very convenient to integrate it with other algorithms; PSO is based on the swarm intelligence theory, and it has very good potential parallelism. Particle swarm optimization has a feature that fitness value is used to exchange information in the population, and guides the population to close the optimal solution. Therefore, a mount of fitness should be calculated in swarm intelligence optimization algorithms in order to find the optimal solution or an approximate one. However, when the calculation of the fitness is quite complex, the time cost of this kind of algorithms will be too large. What’s more, the fitness of optimization problems in the real world is often difficult to calculate. Addressing this problem,Efficient Particle Swarm Optimization Algorithm Based on Affinity Propagation (EAPSO) is proposed in this paper.

2016 ◽  
Vol 11 (1) ◽  
pp. 58-67 ◽  
Author(s):  
S Sarathambekai ◽  
K Umamaheswari

Discrete particle swarm optimization is one of the most recently developed population-based meta-heuristic optimization algorithm in swarm intelligence that can be used in any discrete optimization problems. This article presents a discrete particle swarm optimization algorithm to efficiently schedule the tasks in the heterogeneous multiprocessor systems. All the optimization algorithms share a common algorithmic step, namely population initialization. It plays a significant role because it can affect the convergence speed and also the quality of the final solution. The random initialization is the most commonly used method in majority of the evolutionary algorithms to generate solutions in the initial population. The initial good quality solutions can facilitate the algorithm to locate the optimal solution or else it may prevent the algorithm from finding the optimal solution. Intelligence should be incorporated to generate the initial population in order to avoid the premature convergence. This article presents a discrete particle swarm optimization algorithm, which incorporates opposition-based technique to generate initial population and greedy algorithm to balance the load of the processors. Make span, flow time, and reliability cost are three different measures used to evaluate the efficiency of the proposed discrete particle swarm optimization algorithm for scheduling independent tasks in distributed systems. Computational simulations are done based on a set of benchmark instances to assess the performance of the proposed algorithm.


2020 ◽  
Vol 2020 ◽  
pp. 1-26
Author(s):  
Wusi Yang ◽  
Li Chen ◽  
Yi Wang ◽  
Maosheng Zhang

The recently proposed multiobjective particle swarm optimization algorithm based on competition mechanism algorithm cannot effectively deal with many-objective optimization problems, which is characterized by relatively poor convergence and diversity, and long computing runtime. In this paper, a novel multi/many-objective particle swarm optimization algorithm based on competition mechanism is proposed, which maintains population diversity by the maximum and minimum angle between ordinary and extreme individuals. And the recently proposed θ-dominance is adopted to further enhance the performance of the algorithm. The proposed algorithm is evaluated on the standard benchmark problems DTLZ, WFG, and UF1-9 and compared with the four recently proposed multiobjective particle swarm optimization algorithms and four state-of-the-art many-objective evolutionary optimization algorithms. The experimental results indicate that the proposed algorithm has better convergence and diversity, and its performance is superior to other comparative algorithms on most test instances.


2013 ◽  
Vol 427-429 ◽  
pp. 1934-1938
Author(s):  
Zhong Rong Zhang ◽  
Jin Peng Liu ◽  
Ke De Fei ◽  
Zhao Shan Niu

The aim is to improve the convergence of the algorithm, and increase the population diversity. Adaptively particles of groups fallen into local optimum is adjusted in order to realize global optimal. by judging groups spatial location of concentration and fitness variance. At the same time, the global factors are adjusted dynamically with the action of the current particle fitness. Four typical function optimization problems are drawn into simulation experiment. The results show that the improved particle swarm optimization algorithm is convergent, robust and accurate.


Author(s):  
Megha Vora ◽  
T. T. Mirnalinee

In the past two decades, Swarm Intelligence (SI)-based optimization techniques have drawn the attention of many researchers for finding an efficient solution to optimization problems. Swarm intelligence techniques are characterized by their decentralized way of working that mimics the behavior of colony of ants, swarm of bees, flock of birds, or school of fishes. Algorithmic simplicity and effectiveness of swarm intelligence techniques have made it a powerful tool for solving global optimization problems. Simulation studies of the graceful, but unpredictable, choreography of bird flocks led to the design of the particle swarm optimization algorithm. Studies of the foraging behavior of ants resulted in the development of ant colony optimization algorithm. This chapter provides insight into swarm intelligence techniques, specifically particle swarm optimization and its variants. The objective of this chapter is twofold: First, it describes how swarm intelligence techniques are employed to solve various optimization problems. Second, it describes how swarm intelligence techniques are efficiently applied for clustering, by imposing clustering as an optimization problem.


2016 ◽  
pp. 1519-1544 ◽  
Author(s):  
Megha Vora ◽  
T. T. Mirnalinee

In the past two decades, Swarm Intelligence (SI)-based optimization techniques have drawn the attention of many researchers for finding an efficient solution to optimization problems. Swarm intelligence techniques are characterized by their decentralized way of working that mimics the behavior of colony of ants, swarm of bees, flock of birds, or school of fishes. Algorithmic simplicity and effectiveness of swarm intelligence techniques have made it a powerful tool for solving global optimization problems. Simulation studies of the graceful, but unpredictable, choreography of bird flocks led to the design of the particle swarm optimization algorithm. Studies of the foraging behavior of ants resulted in the development of ant colony optimization algorithm. This chapter provides insight into swarm intelligence techniques, specifically particle swarm optimization and its variants. The objective of this chapter is twofold: First, it describes how swarm intelligence techniques are employed to solve various optimization problems. Second, it describes how swarm intelligence techniques are efficiently applied for clustering, by imposing clustering as an optimization problem.


Complexity ◽  
2020 ◽  
Vol 2020 ◽  
pp. 1-18
Author(s):  
Feng Qian ◽  
Mohammad Reza Mahmoudi ◽  
Hamïd Parvïn ◽  
Kim-Hung Pho ◽  
Bui Anh Tuan

Conventional optimization methods are not efficient enough to solve many of the naturally complicated optimization problems. Thus, inspired by nature, metaheuristic algorithms can be utilized as a new kind of problem solvers in solution to these types of optimization problems. In this paper, an optimization algorithm is proposed which is capable of finding the expected quality of different locations and also tuning its exploration-exploitation dilemma to the location of an individual. A novel particle swarm optimization algorithm is presented which implements the conditioning learning behavior so that the particles are led to perform a natural conditioning behavior on an unconditioned motive. In the problem space, particles are classified into several categories so that if a particle lies within a low diversity category, it would have a tendency to move towards its best personal experience. But, if the particle’s category is with high diversity, it would have the tendency to move towards the global optimum of that category. The idea of the birds’ sensitivity to its flying space is also utilized to increase the particles’ speed in undesired spaces in order to leave those spaces as soon as possible. However, in desirable spaces, the particles’ velocity is reduced to provide a situation in which the particles have more time to explore their environment. In the proposed algorithm, the birds’ instinctive behavior is implemented to construct an initial population randomly or chaotically. Experiments provided to compare the proposed algorithm with the state-of-the-art methods show that our optimization algorithm is one of the most efficient and appropriate ones to solve the static optimization problems.


Sign in / Sign up

Export Citation Format

Share Document