scholarly journals Generate random rectangular maps

Author(s):  
I. V. Kozin ◽  
S. E. Batovskiy

It is known that a large number of applied optimization problems can’t be exactly solved nowadays, because their computational complexity is related to the NP-hard class. In many cases metaheuristics of various types are used to search for approximate solutions, but the choice of the concrete metaheuristic has open question of the quality of the chosen method. There are several possible solutions to this problem, one of which is the verification of metaheuristic algorithms using examples from known test libraries with known records. Another approach to solving the problem of evaluating the quality of algorithms is to compare the "new" algorithm with other algorithms, the work of which has already been investigated. The construction a generator of random problems with a known optimal solution can solve the problem of obtaining "average" estimates of the accuracy for used algorithm in comparison with other methods. The article considers the construction of generators of random non-waste maps of rec-tangular cutting with restrictions on the rectangles of limited sizes. The existence of sets of such cards forms the basis of test problems for checking the quality of approximate algorithms for searching for optimal solution. Rectangular cutting, which is considered in the article, is also the basis for building cuts using more complex shapes. As the simplest method of generating random rectangular non-waste maps, considered a method that uses guillotine cutting. Also, a more complex algorithm for generating a random rectangular cut is given, whose job is to generate a random dot grid and remove some random points from this grid. Much attention is paid to the implementation of the above methods, since the main purpose of the article is to simplify using of generators in practice. All the above algorithms are already used in the software system for testing evolution-aryfragmentary algorithms for various classes of optimization problems on the graphs

2021 ◽  
pp. 1-21
Author(s):  
Xin Li ◽  
Xiaoli Li ◽  
Kang Wang

The key characteristic of multi-objective evolutionary algorithm is that it can find a good approximate multi-objective optimal solution set when solving multi-objective optimization problems(MOPs). However, most multi-objective evolutionary algorithms perform well on regular multi-objective optimization problems, but their performance on irregular fronts deteriorates. In order to remedy this issue, this paper studies the existing algorithms and proposes a multi-objective evolutionary based on niche selection to deal with irregular Pareto fronts. In this paper, the crowding degree is calculated by the niche method in the process of selecting parents when the non-dominated solutions converge to the first front, which improves the the quality of offspring solutions and which is beneficial to local search. In addition, niche selection is adopted into the process of environmental selection through considering the number and the location of the individuals in its niche radius, which improve the diversity of population. Finally, experimental results on 23 benchmark problems including MaF and IMOP show that the proposed algorithm exhibits better performance than the compared MOEAs.


2014 ◽  
Vol 591 ◽  
pp. 172-175
Author(s):  
M. Chandrasekaran ◽  
P. Sriramya ◽  
B. Parvathavarthini ◽  
M. Saravanamanikandan

In modern years, there has been growing importance in the design, analysis and to resolve extremely complex problems. Because of the complexity of problem variants and the difficult nature of the problems they deal with, it is arguably impracticable in the majority time to build appropriate guarantees about the number of fitness evaluations needed for an algorithm to and an optimal solution. In such situations, heuristic algorithms can solve approximate solutions; however suitable time and space complication take part an important role. In present, all recognized algorithms for NP-complete problems are requiring time that's exponential within the problem size. The acknowledged NP-hardness results imply that for several combinatorial optimization problems there are no efficient algorithms that realize a best resolution, or maybe a close to best resolution, on each instance. The study Computational Complexity Analysis of Selective Breeding algorithm involves both an algorithmic issue and a theoretical challenge and the excellence of a heuristic.


2021 ◽  
Vol 11 (16) ◽  
pp. 7591
Author(s):  
Waqas Haider Bangyal ◽  
Kashif Nisar ◽  
Ag. Asri Bin Ag. Ibrahim ◽  
Muhammad Reazul Haque ◽  
Joel J. P. C. Rodrigues ◽  
...  

Metaheuristic algorithms have been widely used to solve diverse kinds of optimization problems. For an optimization problem, population initialization plays a significant role in metaheuristic algorithms. These algorithms can influence the convergence to find an efficient optimal solution. Mainly, for recognizing the importance of diversity, several researchers have worked on the performance for the improvement of metaheuristic algorithms. Population initialization is a vital factor in metaheuristic algorithms such as PSO and DE. Instead of applying the random distribution for the initialization of the population, quasirandom sequences are more useful for the improvement the diversity and convergence factors. This study presents three new low-discrepancy sequences named WELL sequence, Knuth sequence, and Torus sequence to initialize the population in the search space. This paper also gives a comprehensive survey of the various PSO and DE initialization approaches based on the family of quasirandom sequences such as Sobol sequence, Halton sequence, and uniform random distribution. The proposed methods for PSO (TO-PSO, KN-PSO, and WE-PSO) and DE (DE-TO, DE-WE, and DE-KN) have been examined for well-known benchmark test problems and training of the artificial neural network. The finding of our techniques shows promising performance using the family of low-discrepancy sequences over uniform random numbers. For a fair comparison, the approaches using low-discrepancy sequences for PSO and DE are compared with the other family of low-discrepancy sequences and uniform random number and depict the superior results. The experimental results show that the low-discrepancy sequences-based initialization performed exceptionally better than a uniform random number. Moreover, the outcome of our work presents a foresight on how the proposed technique profoundly impacts convergence and diversity. It is anticipated that this low-discrepancy sequence comparative simulation survey would be helpful for studying the metaheuristic algorithm in detail for the researcher.


2018 ◽  
Vol 61 (1) ◽  
pp. 76-98 ◽  
Author(s):  
TING LI ◽  
ZHONG WAN

We propose a new adaptive and composite Barzilai–Borwein (BB) step size by integrating the advantages of such existing step sizes. Particularly, the proposed step size is an optimal weighted mean of two classical BB step sizes and the weights are updated at each iteration in accordance with the quality of the classical BB step sizes. Combined with the steepest descent direction, the adaptive and composite BB step size is incorporated into the development of an algorithm such that it is efficient to solve large-scale optimization problems. We prove that the developed algorithm is globally convergent and it R-linearly converges when applied to solve strictly convex quadratic minimization problems. Compared with the state-of-the-art algorithms available in the literature, the proposed step size is more efficient in solving ill-posed or large-scale benchmark test problems.


Author(s):  
O Schütze ◽  
A Lara ◽  
C A Coello Coello ◽  
M Vasile

When making decisions, having multiple options available for a possible realization of the same project can be advantageous. One way to increase the number of interesting choices is to consider, in addition to the optimal solution x*, also nearly optimal or approximate solutions; these alternative solutions differ from x* and can be in different regions – in the design space – but fulfil certain proximity to its function value f( x*). The scope of this article is the efficient computation and discretization of the set E of ϵ–approximate solutions for scalar optimization problems. To accomplish this task, two strategies to archive and update the data of the search procedure will be suggested and investigated. To make emphasis on data storage efficiency, a way to manage significant and insignificant parameters is also presented. Further on, differential evolution will be used together with the new archivers for the computation of E. Finally, the behaviour of the archiver, as well as the efficiency of the resulting search procedure, will be demonstrated on some academic functions as well as on three models related to space mission design.


2011 ◽  
Vol 2011 ◽  
pp. 1-11
Author(s):  
S. Raja Balachandar ◽  
K. Kannan

This paper presents a1-optheuristic approach to solve resource allocation/reallocation problem which is known as 0/1 multichoice multidimensional knapsack problem (MMKP). The intercept matrix of the constraints is employed to find optimal or near-optimal solution of the MMKP. This heuristic approach is tested for 33 benchmark problems taken from OR library of sizes upto 7000, and the results have been compared with optimum solutions. Computational complexity is proved to be of solving heuristically MMKP using this approach. The performance of our heuristic is compared with the best state-of-art heuristic algorithms with respect to the quality of the solutions found. The encouraging results especially for relatively large-size test problems indicate that this heuristic approach can successfully be used for finding good solutions for highly constrained NP-hard problems.


Author(s):  
Miguel Terra-Neves ◽  
Inês Lynce ◽  
Vasco Manquinho

A Minimal Correction Subset (MCS) of an unsatisfiable constraint set is a minimal subset of constraints that, if removed, makes the constraint set satisfiable. MCSs enjoy a wide range of applications, such as finding approximate solutions to constrained optimization problems. However, existing work on applying MCS enumeration to optimization problems focuses on the single-objective case. In this work, Pareto Minimal Correction Subsets (Pareto-MCSs) are proposed for approximating the Pareto-optimal solution set of multi-objective constrained optimization problems. We formalize and prove an equivalence relationship between Pareto-optimal solutions and Pareto-MCSs. Moreover, Pareto-MCSs and MCSs can be connected in such a way that existing state-of-the-art MCS enumeration algorithms can be used to enumerate Pareto-MCSs. Finally, experimental results on the multi-objective virtual machine consolidation problem show that the Pareto-MCS approach is competitive with state-of-the-art algorithms.


2019 ◽  
Vol 142 (5) ◽  
Author(s):  
Eliot Rudnick-Cohen ◽  
Jeffrey W. Herrmann ◽  
Shapour Azarm

Abstract Feasibility robust optimization techniques solve optimization problems with uncertain parameters that appear only in their constraint functions. Solving such problems requires finding an optimal solution that is feasible for all realizations of the uncertain parameters. This paper presents a new feasibility robust optimization approach involving uncertain parameters defined on continuous domains. The proposed approach is based on an integration of two techniques: (i) a sampling-based scenario generation scheme and (ii) a local robust optimization approach. An analysis of the computational cost of this integrated approach is performed to provide worst-case bounds on its computational cost. The proposed approach is applied to several non-convex engineering test problems and compared against two existing robust optimization approaches. The results show that the proposed approach can efficiently find a robust optimal solution across the test problems, even when existing methods for non-convex robust optimization are unable to find a robust optimal solution. A scalable test problem is solved by the approach, demonstrating that its computational cost scales with problem size as predicted by an analysis of the worst-case computational cost bounds.


2003 ◽  
Vol 13 (2) ◽  
pp. 139-151 ◽  
Author(s):  
Edmund Burke ◽  
Yuri Bykov ◽  
James Newall ◽  
Sanja Petrovic

A common weakness of local search metaheuristics, such as Simulated Annealing, in solving combinatorial optimization problems, is the necessity of setting a certain number of parameters. This tends to generate a significant increase in the total amount of time required to solve the problem and often requires a high level of experience from the user. This paper is motivated by the goal of overcoming this drawback by employing "parameter-free" techniques in the context of automatically solving course timetabling problems. We employ local search techniques with "straightforward" parameters, i.e. ones that an inexperienced user can easily understand. In particular, we present an extended variant of the "Great Deluge" algorithm, which requires only two parameters (which can be interpreted as search time and an estimation of the required level of solution quality). These parameters affect the performance of the algorithm so that a longer search provides a better result - as long as we can intelligently stop the approach from converging too early. Hence, a user can choose a balance between processing time and the quality of the solution. The proposed method has been tested on a range of university course timetabling problems and the results were evaluated within an International Timetabling Competition. The effectiveness of the proposed technique has been confirmed by a high level of quality of results. These results represented the third overall average rating among 21 participants and the best solutions on 8 of the 23 test problems. .


Author(s):  
Gary G. Yen ◽  
Wen-Fung Leong

Constraint handling techniques are mainly designed for evolutionary algorithms to solve constrained multiobjective optimization problems (CMOPs). Most multiojective particle swarm optimization (MOPSO) designs adopt these existing constraint handling techniques to deal with CMOPs. In the proposed constrained MOPSO, information related to particles’ infeasibility and feasibility status is utilized effectively to guide the particles to search for feasible solutions and improve the quality of the optimal solution. This information is incorporated into the four main procedures of a standard MOPSO algorithm. The involved procedures include the updating of personal best archive based on the particles’ Pareto ranks and their constraint violation values; the adoption of infeasible global best archives to store infeasible nondominated solutions; the adjustment of acceleration constants that depend on the personal bests’ and selected global best’s infeasibility and feasibility status; and the integration of personal bests’ feasibility status to estimate the mutation rate in the mutation procedure. Simulation to investigate the proposed constrained MOPSO in solving the selected benchmark problems is conducted. The simulation results indicate that the proposed constrained MOPSO is highly competitive in solving most of the selected benchmark problems.


Sign in / Sign up

Export Citation Format

Share Document