Solving metameric variable-length optimization problems using genetic algorithms

2016 ◽  
Vol 18 (2) ◽  
pp. 247-277 ◽  
Author(s):  
Matthew L. Ryerkerk ◽  
Ronald C. Averill ◽  
Kalyanmoy Deb ◽  
Erik D. Goodman
1995 ◽  
Vol 29 (4) ◽  
pp. 39-56 ◽  
Author(s):  
S. Hurley ◽  
L. Moutinho ◽  
N.M. Stephens

2013 ◽  
Vol 310 ◽  
pp. 609-613
Author(s):  
Ioana D. Balea ◽  
Radu Hulea ◽  
Georgios E. Stavroulakis

This paper presents an implementation of Eurocode load cases for discrete global optimization algorithm for planar structures based on the principles of finite element methods and genetic algorithms. The final optimal design is obtained using IPE sections chosen as feasible by the algorithm, from the available steel sections from industry. The algorithm is tested on an asymmetric planar steel frame with promising results.


Author(s):  
Hamidreza Salmani mojaveri

One of the discussed topics in scheduling problems is Dynamic Flexible Job Shop with Parallel Machines (FDJSPM). Surveys show that this problem because of its concave and nonlinear nature usually has several local optimums. Some of the scheduling problems researchers think that genetic algorithms (GA) are appropriate approach to solve optimization problems of this kind. But researches show that one of the disadvantages of classical genetic algorithms is premature convergence and the probability of trap into the local optimum. Considering these facts, in present research, represented a developed genetic algorithm that its controlling parameters change during algorithm implementation and optimization process. This approach decreases the probability of premature convergence and trap into the local optimum. The several experiments were done show that the priority of proposed procedure of solving in field of the quality of obtained solution and convergence speed toward other present procedure.


2008 ◽  
Vol 16 (3) ◽  
pp. 385-416 ◽  
Author(s):  
Shengxiang Yang

In recent years the genetic algorithm community has shown a growing interest in studying dynamic optimization problems. Several approaches have been devised. The random immigrants and memory schemes are two major ones. The random immigrants scheme addresses dynamic environments by maintaining the population diversity while the memory scheme aims to adapt genetic algorithms quickly to new environments by reusing historical information. This paper investigates a hybrid memory and random immigrants scheme, called memory-based immigrants, and a hybrid elitism and random immigrants scheme, called elitism-based immigrants, for genetic algorithms in dynamic environments. In these schemes, the best individual from memory or the elite from the previous generation is retrieved as the base to create immigrants into the population by mutation. This way, not only can diversity be maintained but it is done more efficiently to adapt genetic algorithms to the current environment. Based on a series of systematically constructed dynamic problems, experiments are carried out to compare genetic algorithms with the memory-based and elitism-based immigrants schemes against genetic algorithms with traditional memory and random immigrants schemes and a hybrid memory and multi-population scheme. The sensitivity analysis regarding some key parameters is also carried out. Experimental results show that the memory-based and elitism-based immigrants schemes efficiently improve the performance of genetic algorithms in dynamic environments.


2017 ◽  
Vol 187 ◽  
pp. 77-87 ◽  
Author(s):  
Rafael de Paula Garcia ◽  
Beatriz Souza Leite Pires de Lima ◽  
Afonso Celso de Castro Lemonge ◽  
Breno Pinheiro Jacob

2016 ◽  
pp. 450-475
Author(s):  
Dipti Singh ◽  
Kusum Deep

Due to their wide applicability and easy implementation, Genetic algorithms (GAs) are preferred to solve many optimization problems over other techniques. When a local search (LS) has been included in Genetic algorithms, it is known as Memetic algorithms. In this chapter, a new variant of single-meme Memetic Algorithm is proposed to improve the efficiency of GA. Though GAs are efficient at finding the global optimum solution of nonlinear optimization problems but usually converge slow and sometimes arrive at premature convergence. On the other hand, LS algorithms are fast but are poor global searchers. To exploit the good qualities of both techniques, they are combined in a way that maximum benefits of both the approaches are reaped. It lets the population of individuals evolve using GA and then applies LS to get the optimal solution. To validate our claims, it is tested on five benchmark problems of dimension 10, 30 and 50 and a comparison between GA and MA has been made.


Sign in / Sign up

Export Citation Format

Share Document