An Improved Adaboost Method for Face Detection

2014 ◽  
Vol 599-601 ◽  
pp. 807-813
Author(s):  
Zhi Xun Zhang ◽  
Juan Wang ◽  
Yong Dong Wang

Existing Adaboost methods for face detection based on particle swarm optimization (PSO) do not consider that PSO suffers from easily trapping in local optimum and slow convergence speed. This paper presents an improved Adaboost method for face detection to solve this problem. In this method, self-adaptive escape PSO (AEPSO) is introduced into conventional Adaboost face detection, meanwhile, Haar-Like rectangular features are represented by particles, so that features selection and classifiers construction could be resolved by using AEPSO. Results of simulation based on Matlab indicate the improved method obtains better detection performance.

2013 ◽  
Vol 760-762 ◽  
pp. 2194-2198 ◽  
Author(s):  
Xue Mei Wang ◽  
Yi Zhuo Guo ◽  
Gui Jun Liu

Adaptive Particle Swarm Optimization algorithm with mutation operation based on K-means is proposed in this paper, this algorithm Combined the local searching optimization ability of K-means with the gobal searching optimization ability of Particle Swarm Optimization, the algorithm self-adaptively adjusted inertia weight according to fitness variance of population. Mutation operation was peocessed for the poor performative particle in population. The results showed that the algorithm had solved the poblems of slow convergence speed of traditional Particle Swarm Optimization algorithm and easy falling into the local optimum of K-Means, and more effectively improved clustering quality.


2013 ◽  
Vol 427-429 ◽  
pp. 1934-1938
Author(s):  
Zhong Rong Zhang ◽  
Jin Peng Liu ◽  
Ke De Fei ◽  
Zhao Shan Niu

The aim is to improve the convergence of the algorithm, and increase the population diversity. Adaptively particles of groups fallen into local optimum is adjusted in order to realize global optimal. by judging groups spatial location of concentration and fitness variance. At the same time, the global factors are adjusted dynamically with the action of the current particle fitness. Four typical function optimization problems are drawn into simulation experiment. The results show that the improved particle swarm optimization algorithm is convergent, robust and accurate.


2017 ◽  
Vol 2017 ◽  
pp. 1-22
Author(s):  
Danping Wang ◽  
Kunyuan Hu ◽  
Lianbo Ma ◽  
Maowei He ◽  
Hanning Chen

A hybrid coevolution particle swarm optimization algorithm with dynamic multispecies strategy based on K-means clustering and nonrevisit strategy based on Binary Space Partitioning fitness tree (called MCPSO-PSH) is proposed. Previous search history memorized into the Binary Space Partitioning fitness tree can effectively restrain the individuals’ revisit phenomenon. The whole population is partitioned into several subspecies and cooperative coevolution is realized by an information communication mechanism between subspecies, which can enhance the global search ability of particles and avoid premature convergence to local optimum. To demonstrate the power of the method, comparisons between the proposed algorithm and state-of-the-art algorithms are grouped into two categories: 10 basic benchmark functions (10-dimensional and 30-dimensional), 10 CEC2005 benchmark functions (30-dimensional), and a real-world problem (multilevel image segmentation problems). Experimental results show that MCPSO-PSH displays a competitive performance compared to the other swarm-based or evolutionary algorithms in terms of solution accuracy and statistical tests.


IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 184656-184663
Author(s):  
Xiaoqiang Tian ◽  
Lingfu Kong ◽  
Deming Kong ◽  
Li Yuan ◽  
Dehan Kong

Author(s):  
Jiarui Zhou ◽  
Junshan Yang ◽  
Ling Lin ◽  
Zexuan Zhu ◽  
Zhen Ji

Particle swarm optimization (PSO) is a swarm intelligence algorithm well known for its simplicity and high efficiency on various problems. Conventional PSO suffers from premature convergence due to the rapid convergence speed and lack of population diversity. It is easy to get trapped in local optima. For this reason, improvements are made to detect stagnation during the optimization and reactivate the swarm to search towards the global optimum. This chapter imposes the reflecting bound-handling scheme and von Neumann topology on PSO to increase the population diversity. A novel crown jewel defense (CJD) strategy is introduced to restart the swarm when it is trapped in a local optimum region. The resultant algorithm named LCJDPSO-rfl is tested on a group of unimodal and multimodal benchmark functions with rotation and shifting. Experimental results suggest that the LCJDPSO-rfl outperforms state-of-the-art PSO variants on most of the functions.


2018 ◽  
Vol 173 ◽  
pp. 02016
Author(s):  
Jin Liang ◽  
Wang Yongzhi ◽  
Bao Xiaodong

The common method of power load forecasting is the least squares support vector machine, but this method is very dependent on the selection of parameters. Particle swarm optimization algorithm is an algorithm suitable for optimizing the selection of support vector parameters, but it is easy to fall into the local optimum. In this paper, we propose a new particle swarm optimization algorithm, it uses non-linear inertial factor change that is used to optimize the algorithm least squares support vector machine to avoid falling into the local optimum. It aims to make the prediction accuracy of the algorithm reach the highest. The experimental results show this method is correct and effective.


2020 ◽  
Vol 14 (4) ◽  
pp. 285-311
Author(s):  
Bernd Bassimir ◽  
Manuel Schmitt ◽  
Rolf Wanka

Abstract We study the variant of Particle Swarm Optimization that applies random velocities in a dimension instead of the regular velocity update equations as soon as the so-called potential of the swarm falls below a certain small bound in this dimension, arbitrarily set by the user. In this case, the swarm performs a forced move. In this paper, we are interested in how, by counting the forced moves, the swarm can decide for itself to stop its movement because it is improbable to find better candidate solutions than the already-found best solution. We formally prove that when the swarm is close to a (local) optimum, it behaves like a blind-searching cloud and that the frequency of forced moves exceeds a certain, objective function-independent value. Based on this observation, we define stopping criteria and evaluate them experimentally showing that good candidate solutions can be found much faster than setting upper bounds on the iterations and better solutions compared to applying other solutions from the literature.


Sign in / Sign up

Export Citation Format

Share Document