A Novel Approach to Designing Surrogate-assisted Genetic Algorithms by Combining Efficient Learning of Walsh Coefficients and Dependencies

2021 ◽  
Vol 1 (2) ◽  
pp. 1-23
Author(s):  
Arkadiy Dushatskiy ◽  
Tanja Alderliesten ◽  
Peter A. N. Bosman

Surrogate-assisted evolutionary algorithms have the potential to be of high value for real-world optimization problems when fitness evaluations are expensive, limiting the number of evaluations that can be performed. In this article, we consider the domain of pseudo-Boolean functions in a black-box setting. Moreover, instead of using a surrogate model as an approximation of a fitness function, we propose to precisely learn the coefficients of the Walsh decomposition of a fitness function and use the Walsh decomposition as a surrogate. If the coefficients are learned correctly, then the Walsh decomposition values perfectly match with the fitness function, and, thus, the optimal solution to the problem can be found by optimizing the surrogate without any additional evaluations of the original fitness function. It is known that the Walsh coefficients can be efficiently learned for pseudo-Boolean functions with k -bounded epistasis and known problem structure. We propose to learn dependencies between variables first and, therefore, substantially reduce the number of Walsh coefficients to be calculated. After the accurate Walsh decomposition is obtained, the surrogate model is optimized using GOMEA, which is considered to be a state-of-the-art binary optimization algorithm. We compare the proposed approach with standard GOMEA and two other Walsh decomposition-based algorithms. The benchmark functions in the experiments are well-known trap functions, NK-landscapes, MaxCut, and MAX3SAT problems. The experimental results demonstrate that the proposed approach is scalable at the supposed complexity of O (ℓ log ℓ) function evaluations when the number of subfunctions is O (ℓ) and all subfunctions are k -bounded, outperforming all considered algorithms.

2013 ◽  
Vol 2013 ◽  
pp. 1-10
Author(s):  
Hamid Reza Erfanian ◽  
M. H. Noori Skandari ◽  
A. V. Kamyad

We present a new approach for solving nonsmooth optimization problems and a system of nonsmooth equations which is based on generalized derivative. For this purpose, we introduce the first order of generalized Taylor expansion of nonsmooth functions and replace it with smooth functions. In other words, nonsmooth function is approximated by a piecewise linear function based on generalized derivative. In the next step, we solve smooth linear optimization problem whose optimal solution is an approximate solution of main problem. Then, we apply the results for solving system of nonsmooth equations. Finally, for efficiency of our approach some numerical examples have been presented.


Author(s):  
Laurens Bliek ◽  
Sicco Verwer ◽  
Mathijs de Weerdt

Abstract When a black-box optimization objective can only be evaluated with costly or noisy measurements, most standard optimization algorithms are unsuited to find the optimal solution. Specialized algorithms that deal with exactly this situation make use of surrogate models. These models are usually continuous and smooth, which is beneficial for continuous optimization problems, but not necessarily for combinatorial problems. However, by choosing the basis functions of the surrogate model in a certain way, we show that it can be guaranteed that the optimal solution of the surrogate model is integer. This approach outperforms random search, simulated annealing and a Bayesian optimization algorithm on the problem of finding robust routes for a noise-perturbed traveling salesman benchmark problem, with similar performance as another Bayesian optimization algorithm, and outperforms all compared algorithms on a convex binary optimization problem with a large number of variables.


1998 ◽  
Vol 9 ◽  
pp. 317-365 ◽  
Author(s):  
G. Di Caro ◽  
M. Dorigo

This paper introduces AntNet, a novel approach to the adaptive learning of routing tables in communications networks. AntNet is a distributed, mobile agents based Monte Carlo system that was inspired by recent work on the ant colony metaphor for solving optimization problems. AntNet's agents concurrently explore the network and exchange collected information. The communication among the agents is indirect and asynchronous, mediated by the network itself. This form of communication is typical of social insects and is called stigmergy. We compare our algorithm with six state-of-the-art routing algorithms coming from the telecommunications and machine learning fields. The algorithms' performance is evaluated over a set of realistic testbeds. We run many experiments over real and artificial IP datagram networks with increasing number of nodes and under several paradigmatic spatial and temporal traffic distributions. Results are very encouraging. AntNet showed superior performance under all the experimental conditions with respect to its competitors. We analyze the main characteristics of the algorithm and try to explain the reasons for its superiority.


2018 ◽  
Vol 35 (4) ◽  
pp. 1805-1828 ◽  
Author(s):  
Kimia Bazargan Lari ◽  
Ali Hamzeh

Purpose Recently, many-objective optimization evolutionary algorithms have been the main issue for researchers in the multi-objective optimization community. To deal with many-objective problems (typically for four or more objectives) some modern frameworks are proposed which have the potential of achieving the finest non-dominated solutions in many-objective spaces. The effectiveness of these algorithms deteriorates greatly as the problem’s dimension increases. Diversity reduction in the objective space is the main reason of this phenomenon. Design/methodology/approach To properly deal with this undesirable situation, this work introduces an indicator-based evolutionary framework that can preserve the population diversity by producing a set of discriminated solutions in high-dimensional objective space. This work attempts to diversify the objective space by proposing a fitness function capable of discriminating the chromosomes in high-dimensional space. The numerical results prove the potential of the proposed method, which had superior performance in most of test problems in comparison with state-of-the-art algorithms. Findings The achieved numerical results empirically prove the superiority of the proposed method to state-of-the-art counterparts in the most test problems of a known artificial benchmark. Originality/value This paper provides a new interpretation and important insights into the many-objective optimization realm by emphasizing on preserving the population diversity.


Author(s):  
Miguel Terra-Neves ◽  
Inês Lynce ◽  
Vasco Manquinho

A Minimal Correction Subset (MCS) of an unsatisfiable constraint set is a minimal subset of constraints that, if removed, makes the constraint set satisfiable. MCSs enjoy a wide range of applications, such as finding approximate solutions to constrained optimization problems. However, existing work on applying MCS enumeration to optimization problems focuses on the single-objective case. In this work, Pareto Minimal Correction Subsets (Pareto-MCSs) are proposed for approximating the Pareto-optimal solution set of multi-objective constrained optimization problems. We formalize and prove an equivalence relationship between Pareto-optimal solutions and Pareto-MCSs. Moreover, Pareto-MCSs and MCSs can be connected in such a way that existing state-of-the-art MCS enumeration algorithms can be used to enumerate Pareto-MCSs. Finally, experimental results on the multi-objective virtual machine consolidation problem show that the Pareto-MCS approach is competitive with state-of-the-art algorithms.


2014 ◽  
Vol 2014 ◽  
pp. 1-12 ◽  
Author(s):  
Ming-Der Yang ◽  
Yeh-Fen Yang ◽  
Tung-Ching Su ◽  
Kai-Siang Huang

Genetic algorithm (GA) is designed to search the optimal solution via weeding out the worse gene strings based on a fitness function. GA had demonstrated effectiveness in solving the problems of unsupervised image classification, one of the optimization problems in a large domain. Many indices or hybrid algorithms as a fitness function in a GA classifier are built to improve the classification accuracy. This paper proposes a new index, DBFCMI, by integrating two common indices, DBI and FCMI, in a GA classifier to improve the accuracy and robustness of classification. For the purpose of testing and verifying DBFCMI, well-known indices such as DBI, FCMI, and PASI are employed as well for comparison. A SPOT-5 satellite image in a partial watershed of Shihmen reservoir is adopted as the examined material for landuse classification. As a result, DBFCMI acquires higher overall accuracy and robustness than the rest indices in unsupervised classification.


2014 ◽  
Vol 26 (1) ◽  
pp. 84-131 ◽  
Author(s):  
Masashi Sugiyama ◽  
Gang Niu ◽  
Makoto Yamada ◽  
Manabu Kimura ◽  
Hirotaka Hachiya

Information-maximization clustering learns a probabilistic classifier in an unsupervised manner so that mutual information between feature vectors and cluster assignments is maximized. A notable advantage of this approach is that it involves only continuous optimization of model parameters, which is substantially simpler than discrete optimization of cluster assignments. However, existing methods still involve nonconvex optimization problems, and therefore finding a good local optimal solution is not straightforward in practice. In this letter, we propose an alternative information-maximization clustering method based on a squared-loss variant of mutual information. This novel approach gives a clustering solution analytically in a computationally efficient way via kernel eigenvalue decomposition. Furthermore, we provide a practical model selection procedure that allows us to objectively optimize tuning parameters included in the kernel function. Through experiments, we demonstrate the usefulness of the proposed approach.


Author(s):  
Thang Trung Nguyen ◽  
Dieu Ngoc Vo

This chapter proposes a Cuckoo Search Algorithm (CSA) and a Modified Cuckoo Search Algorithm (MCSA) for solving short-term hydrothermal scheduling (ST-HTS) problem. The CSA method is a new meta-heuristic algorithm inspired from the obligate brood parasitism of some cuckoo species by laying their eggs in the nests of other host birds of other species for solving optimization problems. In the MCSA method, the eggs are first classified into two groups in which ones with low fitness function are put in top group whereas others with higher fitness function are put in abandoned group. In addition, an updated step size in the MCSA changes and tends to decrease as the iteration increases leading to near global optimal solution. The robustness and effectiveness of the CSA and MCSA are tested on several systems with different objective functions of thermal units. The results obtained by the CSA and MCSA are analyzed and compared have shown that the two methods are favorable for solving short-term hydrothermal scheduling problems.


Author(s):  
Gilles Lebrun ◽  
Olivier Lezoray ◽  
Christopher Charrier ◽  
Hubert Cardot

Evolutionary algorithms (EA) (Rechenberg, 1965) belong to a family of stochastic search algorithms inspired by natural evolution. In the last years, EA were used successfully to produce efficient solutions for a great number of hard optimization problems (Beasley, 1997). These algorithms operate on a population of potential solutions and apply a survival principle according to a fitness measure associated to each solution to produce better approximations of the optimal solution. At each iteration, a new set of solutions is created by selecting individuals according to their level of fitness and by applying to them several operators. These operators model natural processes, such as selection, recombination, mutation, migration, locality and neighborhood. Although the basic idea of EA is straightforward, solutions coding, size of population, fitness function and operators must be defined in compliance with the kind of problem to optimize. Multi-class problems with binary SVM (Support Vector Machine) classifiers are commonly treated as a decomposition in several binary sub-problems. An open question is how to properly choose all models for these sub-problems in order to have the lowest error rate for a specific SVM multi-class scheme. In this paper, we propose a new approach to optimize the generalization capacity of such SVM multi-class schemes. This approach consists in a global selection of models for sub-problems altogether and is denoted as multi-model selection. A multi-model selection can outperform the classical individual model selection used until now in the literature, but this type of selection defines a hard optimisation problem, because it corresponds to a search a efficient solution into a huge space. Therefore, we propose an adapted EA to achieve that multi-model selection by defining specific fitness function and recombination operator.


2019 ◽  
Vol 11 (8) ◽  
pp. 942 ◽  
Author(s):  
Jia ◽  
Peng ◽  
Song ◽  
Oliva ◽  
Lang ◽  
...  

A novel multilevel threshold segmentation method for color satellite images based on Masi entropy is proposed in this paper. Lévy multiverse optimization algorithm (LMVO) has a strong advantage over the traditional multiverse optimization algorithm (MVO) in finding the optimal solution for the segmentation in the three channels of an RGB image. As the work advancement introduces a Lévy multiverse optimization algorithm which uses tournament selection instead of roulette wheel selection, and updates some formulas in the algorithm with mutation factor. Then, the proposal is called TLMVO, and another advantage is that the population diversity of the algorithm in the latest iterations is maintained. The Masi entropy is used as an application and combined with the improved TLMVO algorithm for satellite color image segmentation. Masi entropy combines the additivity of Renyi entropy and the non-extensibility of Tsallis entropy. By increasing the number of thesholds, the quality of segmenttion becomes better, then the dimensionality of the problem also increases. Fitness function value, average CPU running time, Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM) and Feature Similarity Index (FSIM) were used to evaluate the segmentation results. Further statistical evaluation was given by Wilcoxon's rank sum test and Friedman test. The experimental results show that the TLMVO algorithm has wide adaptability to high-dimensional optimization problems, and has obvious advantages in objective function value, image quality detection, convergence performance and robustness.


Sign in / Sign up

Export Citation Format

Share Document