scholarly journals A New Multiobjective Time-Cost Trade-Off for Scheduling Maintenance Problem in a Series-Parallel System

2021 ◽  
Vol 2021 ◽  
pp. 1-13
Author(s):  
Leyla Sadat Tavassoli ◽  
Reza Massah ◽  
Arsalan Montazeri ◽  
Mirpouya Mirmozaffari ◽  
Guang-Jun Jiang ◽  
...  

In this paper, a modified model of Nondominated Sorting Genetic Algorithm 2 (NSGA-II), which is one of the Multiobjective Evolutionary Algorithms, is proposed. This algorithm is a new model designed to make a trade-off between minimizing the cost of preventive maintenance (PM) and minimizing the time taken to perform this maintenance for a series-parallel system. In this model, the limitations of labor and equipment of the maintenance team and the effects of maintenance issues on manufacturing problems are also considered. In the mathematical model, finding the appropriate objective functions for the maintenance scheduling problem requires all maintenance costs and failure rates to be integrated. Additionally, the effects of production interruption during preventive maintenance are added to objective functions. Furthermore, to make a better performance compared with a regular NSGA-II algorithm, we proposed a modified algorithm with a repository to keep more unacceptable solutions. These solutions can be modified and changed with the proposed mutation algorithm to acceptable solutions. In this algorithm, modified operators, such as simulated binary crossover and polynomial mutation, will improve the algorithm to generate convergence and uniformly distributed solutions with more diverse solutions. Finally, by comparing the experimental solutions with the solutions of two Strength Pareto Evolutionary Algorithm 2 (SPEA2) and regular NSGA-II, MNSGA-II generates more efficient and uniform solutions than the other two algorithms.

2018 ◽  
Vol 2018 ◽  
pp. 1-11 ◽  
Author(s):  
Herian A. Leyva ◽  
Edén Bojórquez ◽  
Juan Bojórquez ◽  
Alfredo Reyes-Salazar ◽  
José H. Castorena ◽  
...  

In the present study, the optimal seismic design of reinforced concrete (RC) buildings is obtained. For this purpose, genetic algorithms (GAs) are used through the technique NSGA-II (Nondominated Sorting Genetic Algorithm), thus a multiobjective procedure with two objective functions is established. The first objective function is the control of maximum interstory drift which is the most common parameter used in seismic design codes, while the second is to minimize the cost of the structure. For this aim, several RC buildings are designed in accordance with the Mexico City Building Code (MCBC). It is assumed that the structures are constituted by rectangular and square concrete sections for the beams, columns, and slabs which are represented by a binary codification. In conclusion, this study provides complete designed RC buildings which also can be used directly in the structural and civil engineering practice by means of genetic algorithms. Moreover, genetic algorithms are able to find the most adequate structures in terms of seismic performance and economy.


2008 ◽  
Vol 16 (3) ◽  
pp. 355-384 ◽  
Author(s):  
Hongbing Fang ◽  
Qian Wang ◽  
Yi-Cheng Tu ◽  
Mark F. Horstemeyer

We present a new non-dominated sorting algorithm to generate the non-dominated fronts in multi-objective optimization with evolutionary algorithms, particularly the NSGA-II. The non-dominated sorting algorithm used by NSGA-II has a time complexity of O(MN2) in generating non-dominated fronts in one generation (iteration) for a population size N and M objective functions. Since generating non-dominated fronts takes the majority of total computational time (excluding the cost of fitness evaluations) of NSGA-II, making this algorithm faster will significantly improve the overall efficiency of NSGA-II and other genetic algorithms using non-dominated sorting. The new non-dominated sorting algorithm proposed in this study reduces the number of redundant comparisons existing in the algorithm of NSGA-II by recording the dominance information among solutions from their first comparisons. By utilizing a new data structure called the dominance tree and the divide-and-conquer mechanism, the new algorithm is faster than NSGA-II for different numbers of objective functions. Although the number of solution comparisons by the proposed algorithm is close to that of NSGA-II when the number of objectives becomes large, the total computational time shows that the proposed algorithm still has better efficiency because of the adoption of the dominance tree structure and the divide-and-conquer mechanism.


2020 ◽  
Vol 39 (3) ◽  
pp. 3259-3273
Author(s):  
Nasser Shahsavari-Pour ◽  
Najmeh Bahram-Pour ◽  
Mojde Kazemi

The location-routing problem is a research area that simultaneously solves location-allocation and vehicle routing issues. It is critical to delivering emergency goods to customers with high reliability. In this paper, reliability in location and routing problems was considered as the probability of failure in depots, vehicles, and routs. The problem has two objectives, minimizing the cost and maximizing the reliability, the latter expressed by minimizing the expected cost of failure. First, a mathematical model of the problem was presented and due to its NP-hard nature, it was solved by a meta-heuristic approach using a NSGA-II algorithm and a discrete multi-objective firefly algorithm. The efficiency of these algorithms was studied through a complete set of examples and it was found that the multi-objective discrete firefly algorithm has a better Diversification Metric (DM) index; the Mean Ideal Distance (MID) and Spacing Metric (SM) indexes are only suitable for small to medium problems, losing their effectiveness for big problems.


2020 ◽  
Vol 4 (02) ◽  
pp. 34-45
Author(s):  
Naufal Dzikri Afifi ◽  
Ika Arum Puspita ◽  
Mohammad Deni Akbar

Shift to The Front II Komplek Sukamukti Banjaran Project is one of the projects implemented by one of the companies engaged in telecommunications. In its implementation, each project including Shift to The Front II Komplek Sukamukti Banjaran has a time limit specified in the contract. Project scheduling is an important role in predicting both the cost and time in a project. Every project should be able to complete the project before or just in the time specified in the contract. Delay in a project can be anticipated by accelerating the duration of completion by using the crashing method with the application of linear programming. Linear programming will help iteration in the calculation of crashing because if linear programming not used, iteration will be repeated. The objective function in this scheduling is to minimize the cost. This study aims to find a trade-off between the costs and the minimum time expected to complete this project. The acceleration of the duration of this study was carried out using the addition of 4 hours of overtime work, 3 hours of overtime work, 2 hours of overtime work, and 1 hour of overtime work. The normal time for this project is 35 days with a service fee of Rp. 52,335,690. From the results of the crashing analysis, the alternative chosen is to add 1 hour of overtime to 34 days with a total service cost of Rp. 52,375,492. This acceleration will affect the entire project because there are 33 different locations worked on Shift to The Front II and if all these locations can be accelerated then the duration of completion of the entire project will be effective


2020 ◽  
Vol 8 (4) ◽  
pp. 276-286
Author(s):  
Vu Duc Quyen ◽  
Andrey Ronzhin

Three posterior algorithms NSGA-II, MOGWO and MOPSO to solve the problem of multicriteria optimization of the robotic gripper design are considered. The description of the kinematic model of the developed prototype of the four-fingered gripper for picking tomatoes, its limitations and objective functions used in the optimization of the design are given. The main advantage of the developed prototype is the use of one actuator for the control of the fingers and the suction nozzle. The results of optimization of the kinematic model and the dimensions of the elements of robotic gripper using the considered posterior algorithms are presented.


2020 ◽  
Vol 12 (7) ◽  
pp. 2767 ◽  
Author(s):  
Víctor Yepes ◽  
José V. Martí ◽  
José García

The optimization of the cost and CO 2 emissions in earth-retaining walls is of relevance, since these structures are often used in civil engineering. The optimization of costs is essential for the competitiveness of the construction company, and the optimization of emissions is relevant in the environmental impact of construction. To address the optimization, black hole metaheuristics were used, along with a discretization mechanism based on min–max normalization. The stability of the algorithm was evaluated with respect to the solutions obtained; the steel and concrete values obtained in both optimizations were analyzed. Additionally, the geometric variables of the structure were compared. Finally, the results obtained were compared with another algorithm that solved the problem. The results show that there is a trade-off between the use of steel and concrete. The solutions that minimize CO 2 emissions prefer the use of concrete instead of those that optimize the cost. On the other hand, when comparing the geometric variables, it is seen that most remain similar in both optimizations except for the distance between buttresses. When comparing with another algorithm, the results show a good performance in optimization using the black hole algorithm.


Author(s):  
Vincent E. Castillo ◽  
John E. Bell ◽  
Diane A. Mollenkopf ◽  
Theodore P. Stank

2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Jeonghyuk Park ◽  
Yul Ri Chung ◽  
Seo Taek Kong ◽  
Yeong Won Kim ◽  
Hyunho Park ◽  
...  

AbstractThere have been substantial efforts in using deep learning (DL) to diagnose cancer from digital images of pathology slides. Existing algorithms typically operate by training deep neural networks either specialized in specific cohorts or an aggregate of all cohorts when there are only a few images available for the target cohort. A trade-off between decreasing the number of models and their cancer detection performance was evident in our experiments with The Cancer Genomic Atlas dataset, with the former approach achieving higher performance at the cost of having to acquire large datasets from the cohort of interest. Constructing annotated datasets for individual cohorts is extremely time-consuming, with the acquisition cost of such datasets growing linearly with the number of cohorts. Another issue associated with developing cohort-specific models is the difficulty of maintenance: all cohort-specific models may need to be adjusted when a new DL algorithm is to be used, where training even a single model may require a non-negligible amount of computation, or when more data is added to some cohorts. In resolving the sub-optimal behavior of a universal cancer detection model trained on an aggregate of cohorts, we investigated how cohorts can be grouped to augment a dataset without increasing the number of models linearly with the number of cohorts. This study introduces several metrics which measure the morphological similarities between cohort pairs and demonstrates how the metrics can be used to control the trade-off between performance and the number of models.


2020 ◽  
Vol 15 (1) ◽  
pp. 4-17
Author(s):  
Jean-François Biasse ◽  
Xavier Bonnetain ◽  
Benjamin Pring ◽  
André Schrottenloher ◽  
William Youmans

AbstractWe propose a heuristic algorithm to solve the underlying hard problem of the CSIDH cryptosystem (and other isogeny-based cryptosystems using elliptic curves with endomorphism ring isomorphic to an imaginary quadratic order 𝒪). Let Δ = Disc(𝒪) (in CSIDH, Δ = −4p for p the security parameter). Let 0 < α < 1/2, our algorithm requires:A classical circuit of size $2^{\tilde{O}\left(\log(|\Delta|)^{1-\alpha}\right)}.$A quantum circuit of size $2^{\tilde{O}\left(\log(|\Delta|)^{\alpha}\right)}.$Polynomial classical and quantum memory.Essentially, we propose to reduce the size of the quantum circuit below the state-of-the-art complexity $2^{\tilde{O}\left(\log(|\Delta|)^{1/2}\right)}$ at the cost of increasing the classical circuit-size required. The required classical circuit remains subexponential, which is a superpolynomial improvement over the classical state-of-the-art exponential solutions to these problems. Our method requires polynomial memory, both classical and quantum.


2019 ◽  
Vol 11 (9) ◽  
pp. 2571
Author(s):  
Xujing Zhang ◽  
Lichuan Wang ◽  
Yan Chen

Low-carbon production has become one of the top management objectives for every industry. In garment manufacturing, the material distribution process always generates high carbon emissions. In order to reduce carbon emissions and the number of operators to meet enterprises’ requirements to control the cost of production and protect the environment, the paths of material distribution were analyzed to find the optimal solution. In this paper, the model of material distribution to obtain minimum carbon emissions and vehicles (operators) was established to optimize the multi-target management in three different production lines (multi-line, U-shape two-line, and U-shape three-line), while the workstations were organized in three ways: in the order of processes, in the type of machines, and in the components of garment. The NSGA-II algorithm (non-dominated sorting genetic algorithm-II) was applied to obtain the results of this model. The feasibility of the model and algorithm was verified by the practice of men’s shirts manufacture. It could be found that material distribution of multi-line layout produced the least carbon emissions when the machines were arranged in the group of type.


Sign in / Sign up

Export Citation Format

Share Document