Hookes-Jeeves-Based Variant of Memetic Algorithm

Author(s):  
Dipti Singh ◽  
Kusum Deep

Due to their wide applicability and easy implementation, Genetic algorithms (GAs) are preferred to solve many optimization problems over other techniques. When a local search (LS) has been included in Genetic algorithms, it is known as Memetic algorithms. In this chapter, a new variant of single-meme Memetic Algorithm is proposed to improve the efficiency of GA. Though GAs are efficient at finding the global optimum solution of nonlinear optimization problems but usually converge slow and sometimes arrive at premature convergence. On the other hand, LS algorithms are fast but are poor global searchers. To exploit the good qualities of both techniques, they are combined in a way that maximum benefits of both the approaches are reaped. It lets the population of individuals evolve using GA and then applies LS to get the optimal solution. To validate our claims, it is tested on five benchmark problems of dimension 10, 30 and 50 and a comparison between GA and MA has been made.

2016 ◽  
pp. 450-475
Author(s):  
Dipti Singh ◽  
Kusum Deep

Due to their wide applicability and easy implementation, Genetic algorithms (GAs) are preferred to solve many optimization problems over other techniques. When a local search (LS) has been included in Genetic algorithms, it is known as Memetic algorithms. In this chapter, a new variant of single-meme Memetic Algorithm is proposed to improve the efficiency of GA. Though GAs are efficient at finding the global optimum solution of nonlinear optimization problems but usually converge slow and sometimes arrive at premature convergence. On the other hand, LS algorithms are fast but are poor global searchers. To exploit the good qualities of both techniques, they are combined in a way that maximum benefits of both the approaches are reaped. It lets the population of individuals evolve using GA and then applies LS to get the optimal solution. To validate our claims, it is tested on five benchmark problems of dimension 10, 30 and 50 and a comparison between GA and MA has been made.


2006 ◽  
Vol 12 (1) ◽  
pp. 89-115 ◽  
Author(s):  
Hideaki Suzuki ◽  
Hidefumi Sawai ◽  
Wojciech Piaseczny

A chemical genetic algorithm (CGA) in which several types of molecules (information units) react with each other in a cell is proposed. Not only the information in DNA, but also smaller molecules responsible for the transcription and translation of DNA into amino acids, are adaptively changed during evolution, which optimizes the fundamental mapping from binary substrings in DNA (genotype) to real values for a parameter set (phenotype). Through the struggle between cells containing a DNA unit and small molecular units, the codes (DNA) and the interpreter (the small molecular units) coevolve, and a specific output function, from which a cell's fitness is evaluated, is optimized. To demonstrate the effectiveness of the CGA, it is applied to a set of variable-separable and variable-inseparable problems, and it is shown that the CGA can robustly solve a wide range of optimization problems regardless of their fitness characteristics. To ascertain the optimization of the genotypeto-phenotype mapping by the CGA, we also conduct analytical experiments for some problems while observing the basin size of a global optimum solution in the binary genotype space. The results show that the CGA effectively augments the basin size, makes it easier for evolution to find a path to the global optimum solution, and enhances the GA's evolvability during evolution.


Author(s):  
Masao Arakawa ◽  
Tomoyuki Miyashita ◽  
Hiroshi Ishikawa

In some cases of developing a new product, response surface of an objective function is not always single peaked function, and it is often multi-peaked function. In that case, designers would like to have not oniy global optimum solution but also as many local optimum solutions and/or quasi-optimum solutions as possible, so that he or she can select one out of them considering the other conditions that are not taken into account priori to optimization. Although this information is quite useful, it is not that easy to obtain with a single trial of optimization. In this study, we will propose a screening of fitness function in genetic algorithms (GA). Which change fitness function during searching. Therefore, GA needs to have higher flexibility in searching. Genetic Range Genetic Algorithms include a number of searching range in a single generation. Just like there are a number of species in wild life. Therefore, it can arrange to have both global searching range and also local searching range with different fitness function. In this paper, we demonstrate the effectiveness of the proposed method through a simple benchmark test problems.


Author(s):  
Foo Fong Yeng ◽  
Soo Kum Yoke ◽  
Azrina Suhaimi

Genetic Algorithm is an algorithm imitating the natural evolution process in solving optimization problems. All feasible (candidate) solutions would be encoded into chromosomes and undergo the execution of genetic operators in evolution. The evolution itself is a process searching for optimum solution. The searching would stop when a stopping criterion is met. Then, the fittest chromosome of last generation is declared as the optimum solution. However, this optimum solution might be a local optimum or a global optimum solution. Hence, an appropriate stopping criterion is important such that the search is not ended before a global optimum solution is found. In this paper, saturation of population fitness is proposed as a stopping criterion for ending the search. The proposed stopping criteria was compared with conventional stopping criterion, fittest chromosomes repetition, under various parameters setting. The results show that the performance of proposed stopping criterion is superior as compared to the conventional stopping criterion.


Author(s):  
Masao Arakawa ◽  
Hiroshi Yamakawa

Abstract In the design process, it would be much better to give designers more and more acceptable and reasonable design candidates to derive their preferences to meet requirements of multiple high performances. In this paper, we will propose a new method, to give such acceptable design candidates based on genetic algorithms (GAs) with considering something like strategy for adaptation. GAs are search algorithms based on the mechanics of natural selection and natural genetics. Yet, they are no simple random walk but they efficiently exploit historical information to speculate on new search points with expected improved performance. Thus, we expect in GAs to give multiple acceptable and near optimum design candidates just like so many species in natural living things. However, in simple GAs, after specific number of generations, their populations become near one or two specific local optimum solutions. (Hoping for including global optimum solution.) In order to obtain multiple acceptable solutions, we need to perform GAs processes by keeping variations in characters of individuals. The proposed method tries to find multiple acceptable solutions by considering strategy for adaptation into GAs processes, which are food chain, strategy of foraging, death strategy and strategy of reproduction. As a numerical example, we apply the proposed method to simple multi-objective optimization and demonstrate its efficiencies.


2016 ◽  
Vol 38 (4) ◽  
pp. 307-317
Author(s):  
Pham Hoang Anh

In this paper, the optimal sizing of truss structures is solved using a novel evolutionary-based optimization algorithm. The efficiency of the proposed method lies in the combination of global search and local search, in which the global move is applied for a set of random solutions whereas the local move is performed on the other solutions in the search population. Three truss sizing benchmark problems with discrete variables are used to examine the performance of the proposed algorithm. Objective functions of the optimization problems are minimum weights of the whole truss structures and constraints are stress in members and displacement at nodes. Here, the constraints and objective function are treated separately so that both function and constraint evaluations can be saved. The results show that the new algorithm can find optimal solution effectively and it is competitive with some recent metaheuristic algorithms in terms of number of structural analyses required.


2021 ◽  
Author(s):  
Zuanjia Xie ◽  
Chunliang Zhang ◽  
Haibin Ouyang ◽  
Steven Li ◽  
Liqun Gao

Abstract Jaya algorithm is an advanced optimization algorithm, which has been applied to many real-world optimization problems. Jaya algorithm has better performance in some optimization field. However, Jaya algorithm exploration capability is not better. In order to enhance exploration capability of the Jaya algorithm, a self-adaptively commensal learning-based Jaya algorithm with multi-populations (Jaya-SCLMP) is presented in this paper. In Jaya-SCLMP, a commensal learning strategy is used to increase the probability of finding the global optimum, in which the person history best and worst information is used to explore new solution area. Moreover, a multi-populations strategy based on Gaussian distribution scheme and learning dictionary is utilized to enhance the exploration capability, meanwhile every sub-population employed three Gaussian distributions at each generation, roulette wheel selection is employed to choose a scheme based on learning dictionary. The performance of Jaya-SCLMP is evaluated based on 28 CEC 2013 unconstrained benchmark problems. In addition, three reliability problems, i.e. complex (bridge) system, series system and series-parallel system are selected. Compared with several Jaya variants and several state-of-the-art other algorithms, the experimental results reveal that Jaya-SCLMP is effective.


2021 ◽  
pp. 1-21
Author(s):  
Xin Li ◽  
Xiaoli Li ◽  
Kang Wang

The key characteristic of multi-objective evolutionary algorithm is that it can find a good approximate multi-objective optimal solution set when solving multi-objective optimization problems(MOPs). However, most multi-objective evolutionary algorithms perform well on regular multi-objective optimization problems, but their performance on irregular fronts deteriorates. In order to remedy this issue, this paper studies the existing algorithms and proposes a multi-objective evolutionary based on niche selection to deal with irregular Pareto fronts. In this paper, the crowding degree is calculated by the niche method in the process of selecting parents when the non-dominated solutions converge to the first front, which improves the the quality of offspring solutions and which is beneficial to local search. In addition, niche selection is adopted into the process of environmental selection through considering the number and the location of the individuals in its niche radius, which improve the diversity of population. Finally, experimental results on 23 benchmark problems including MaF and IMOP show that the proposed algorithm exhibits better performance than the compared MOEAs.


2015 ◽  
pp. 1246-1276
Author(s):  
Wen Fung Leong ◽  
Yali Wu ◽  
Gary G. Yen

Generally, constraint-handling techniques are designed for evolutionary algorithms to solve Constrained Multiobjective Optimization Problems (CMOPs). Most Multiojective Particle Swarm Optimization (MOPSO) designs adopt these existing constraint-handling techniques to deal with CMOPs. In this chapter, the authors present a constrained MOPSO in which the information related to particles' infeasibility and feasibility status is utilized effectively to guide the particles to search for feasible solutions and to improve the quality of the optimal solution found. The updating of personal best archive is based on the particles' Pareto ranks and their constraint violations. The infeasible global best archive is adopted to store infeasible nondominated solutions. The acceleration constants are adjusted depending on the personal bests' and selected global bests' infeasibility and feasibility statuses. The personal bests' feasibility statuses are integrated to estimate the mutation rate in the mutation procedure. The simulation results indicate that the proposed constrained MOPSO is highly competitive in solving selected benchmark problems.


2011 ◽  
Vol 19 (3) ◽  
pp. 345-371 ◽  
Author(s):  
Daniel Karapetyan ◽  
Gregory Gutin

Memetic algorithms are known to be a powerful technique in solving hard optimization problems. To design a memetic algorithm, one needs to make a host of decisions. Selecting the population size is one of the most important among them. Most of the algorithms in the literature fix the population size to a certain constant value. This reduces the algorithm's quality since the optimal population size varies for different instances, local search procedures, and runtimes. In this paper we propose an adjustable population size. It is calculated as a function of the runtime of the whole algorithm and the average runtime of the local search for the given instance. Note that in many applications the runtime of a heuristic should be limited and, therefore, we use this bound as a parameter of the algorithm. The average runtime of the local search procedure is measured during the algorithm's run. Some coefficients which are independent of the instance and the local search are to be tuned at the design time; we provide a procedure to find these coefficients. The proposed approach was used to develop a memetic algorithm for the multidimensional assignment problem (MAP). We show that our adjustable population size makes the algorithm flexible to perform efficiently for a wide range of running times and local searches and this does not require any additional tuning of the algorithm.


Sign in / Sign up

Export Citation Format

Share Document