scholarly journals Improved ensemble of differential evolution variants

PLoS ONE ◽  
2021 ◽  
Vol 16 (8) ◽  
pp. e0256206
Author(s):  
Juan Yao ◽  
Zhe Chen ◽  
Zhenling Liu

In the field of Differential Evolution (DE), a number of measures have been used to enhance algorithm. However, most of the measures need revision for fitting ensemble of different combinations of DE operators—ensemble DE algorithm. Meanwhile, although ensemble DE algorithm may show better performance than each of its constituent algorithms, there still exists the possibility of further improvement on performance with the help of revised measures. In this paper, we manage to implement measures into Ensemble of Differential Evolution Variants (EDEV). Firstly, we extend the collecting range of optional external archive of JADE—one of the constituent algorithm in EDEV. Then, we revise and implement the Event-Triggered Impulsive (ETI) control. Finally, Linear Population Size Reduction (LPSR) is used by us. Then, we obtain Improved Ensemble of Differential Evolution Variants (IEDEV). In our experiments, good performers in the CEC competitions on real parameter single objective optimization among population-based metaheuristics, state-of-the-art DE algorithms, or up-to-date DE algorithms are involved. Experiments show that our IEDEV is very competitive.

2015 ◽  
Vol 2015 ◽  
pp. 1-10 ◽  
Author(s):  
V. Gonuguntla ◽  
R. Mallipeddi ◽  
Kalyana C. Veluvolu

Differential evolution (DE) is simple and effective in solving numerous real-world global optimization problems. However, its effectiveness critically depends on the appropriate setting of population size and strategy parameters. Therefore, to obtain optimal performance the time-consuming preliminary tuning of parameters is needed. Recently, different strategy parameter adaptation techniques, which can automatically update the parameters to appropriate values to suit the characteristics of optimization problems, have been proposed. However, most of the works do not control the adaptation of the population size. In addition, they try to adapt each strategy parameters individually but do not take into account the interaction between the parameters that are being adapted. In this paper, we introduce a DE algorithm where both strategy parameters are self-adapted taking into account the parameter dependencies by means of a multivariate probabilistic technique based on Gaussian Adaptation working on the parameter space. In addition, the proposed DE algorithm starts by sampling a huge number of sample solutions in the search space and in each generation a constant number of individuals from huge sample set are adaptively selected to form the population that evolves. The proposed algorithm is evaluated on 14 benchmark problems of CEC 2005 with different dimensionality.


2018 ◽  
Vol 8 (10) ◽  
pp. 1945 ◽  
Author(s):  
Tarik Eltaeib ◽  
Ausif Mahmood

Differential evolution (DE) has been extensively used in optimization studies since its development in 1995 because of its reputation as an effective global optimizer. DE is a population-based metaheuristic technique that develops numerical vectors to solve optimization problems. DE strategies have a significant impact on DE performance and play a vital role in achieving stochastic global optimization. However, DE is highly dependent on the control parameters involved. In practice, the fine-tuning of these parameters is not always easy. Here, we discuss the improvements and developments that have been made to DE algorithms. In particular, we present a state-of-the-art survey of the literature on DE and its recent advances, such as the development of adaptive, self-adaptive and hybrid techniques.


Author(s):  
Sukanta Nama ◽  
Apu Kumar Saha

The population-based efficient iterative evolutionary algorithm (EA) is differential evolution (DE). It has fewer control parameters but is useful when dealing with complex problems of optimization in the real world. A great deal of progress has already been made and implemented in various fields of engineering and science. Nevertheless, DE is prone to the setting of control parameters in its performance evaluation. Therefore, the appropriate adjustment of the time-consuming control parameters is necessary to achieve optimal DE efficiency. This research proposes a new version of the DE algorithm control parameters and mutation operator. For the justifiability of the suggested method, several benchmark functions are taken from the literature. The test results are contrasted with other literary algorithms.


2013 ◽  
Vol 415 ◽  
pp. 349-352
Author(s):  
Hong Wei Zhao ◽  
Hong Gang Xia

Differential evolution (DE) is a population-based stochastic function minimizer (or maximizer), whose simple yet powerful and straightforward features make it very attractive for numerical optimization. However, DE is easy to trapped into local optima. In this paper, an improved differential evolution algorithm (IDE) proposed to speed the convergence rate of DE and enhance the global search of DE. The IDE employed a new mutation operation and modified crossover operation. The former can rapidly enhance the convergence of the MDE, and the latter can prevent the MDE from being trapped into the local optimum effectively. Besides, we dynamic adjust the scaling factor (F) and the crossover rate (CR), which is aimed at further improving algorithm performance. Based on several benchmark experiment simulations, the IDE has demonstrated stronger convergence and stability than original differential (DE) algorithm and other algorithms (PSO and JADE) that reported in recent literature.


2020 ◽  
Vol 13 (6) ◽  
pp. 168-178
Author(s):  
Pyae Cho ◽  
◽  
Thi Nyunt ◽  

Differential Evolution (DE) has become an advanced, robust, and proficient alternative technique for clustering on account of their population-based stochastic and heuristic search manners. Balancing better the exploitation and exploration power of the DE algorithm is important because this ability influences the performance of the algorithm. Besides, keeping superior solutions for the initial population raises the probability of finding better solutions and the rate of convergence. In this paper, an enhanced DE algorithm is introduced for clustering to offer better cluster solutions with faster convergence. The proposed algorithm performs a modified mutation strategy to improve the DE’s search behavior and exploits Quasi-Opposition-based Learning (QBL) to choose fitter initial solutions. This mutation strategy that uses the best solution as a target solution and applies three differentials contributes to avoiding local optima trap and slow convergence. The QBL based initialization method also contributes to increasing the quality of the clustering results and convergence rate. The experimental analysis was conducted on seven real datasets from the UCI repository to evaluate the performance of the proposed clustering algorithm. The obtained results showed that the proposed algorithm achieves more compact clusters and stable solutions than the competing conventional DE variants. Moreover, the performance of the proposed algorithm was compared with the existing state of the art clustering techniques based on DE. The corresponding results also pointed out that the proposed algorithm is comparable to other DE based clustering approaches in terms of the value of the objective functions. Therefore, the proposed algorithm can be regarded as an efficient clustering tool.


Sign in / Sign up

Export Citation Format

Share Document