scholarly journals An Expensive Multi-Objective Optimization Algorithm Based on Decision Space Compression

Author(s):  
Haosen Liu ◽  
Fangqing Gu ◽  
Yiu-Ming Cheung

Numerous surrogate-assisted expensive multi-objective optimization algorithms were proposed to deal with expensive multi-objective optimization problems in the past few years. The accuracy of the surrogate models degrades as the number of decision variables increases. In this paper, we propose a surrogate-assisted expensive multi-objective optimization algorithm based on decision space compression. Several surrogate models are built in the lower dimensional compressed space. The promising points are generated and selected in the lower compressed decision space and decoded to the original decision space for evaluation. Experimental studies show that the proposed algorithm achieves a good performance in handling expensive multi-objective optimization problems with high-dimensional decision space.

2015 ◽  
Vol 23 (1) ◽  
pp. 69-100 ◽  
Author(s):  
Handing Wang ◽  
Licheng Jiao ◽  
Ronghua Shang ◽  
Shan He ◽  
Fang Liu

There can be a complicated mapping relation between decision variables and objective functions in multi-objective optimization problems (MOPs). It is uncommon that decision variables influence objective functions equally. Decision variables act differently in different objective functions. Hence, often, the mapping relation is unbalanced, which causes some redundancy during the search in a decision space. In response to this scenario, we propose a novel memetic (multi-objective) optimization strategy based on dimension reduction in decision space (DRMOS). DRMOS firstly analyzes the mapping relation between decision variables and objective functions. Then, it reduces the dimension of the search space by dividing the decision space into several subspaces according to the obtained relation. Finally, it improves the population by the memetic local search strategies in these decision subspaces separately. Further, DRMOS has good portability to other multi-objective evolutionary algorithms (MOEAs); that is, it is easily compatible with existing MOEAs. In order to evaluate its performance, we embed DRMOS in several state of the art MOEAs to facilitate our experiments. The results show that DRMOS has the advantage in terms of convergence speed, diversity maintenance, and portability when solving MOPs with an unbalanced mapping relation between decision variables and objective functions.


Author(s):  
Amarjeet Prajapati

AbstractOver the past 2 decades, several multi-objective optimizers (MOOs) have been proposed to address the different aspects of multi-objective optimization problems (MOPs). Unfortunately, it has been observed that many of MOOs experiences performance degradation when applied over MOPs having a large number of decision variables and objective functions. Specially, the performance of MOOs rapidly decreases when the number of decision variables and objective functions increases by more than a hundred and three, respectively. To address the challenges caused by such special case of MOPs, some large-scale multi-objective optimization optimizers (L-MuOOs) and large-scale many-objective optimization optimizers (L-MaOOs) have been developed in the literature. Even after vast development in the direction of L-MuOOs and L-MaOOs, the supremacy of these optimizers has not been tested on real-world optimization problems containing a large number of decision variables and objectives such as large-scale many-objective software clustering problems (L-MaSCPs). In this study, the performance of nine L-MuOOs and L-MaOOs (i.e., S3-CMA-ES, LMOSCO, LSMOF, LMEA, IDMOPSO, ADC-MaOO, NSGA-III, H-RVEA, and DREA) is evaluated and compared over five L-MaSCPs in terms of IGD, Hypervolume, and MQ metrics. The experimentation results show that the S3-CMA-ES and LMOSCO perform better compared to the LSMOF, LMEA, IDMOPSO, ADC-MaOO, NSGA-III, H-RVEA, and DREA in most of the cases. The LSMOF, LMEA, IDMOPSO, ADC-MaOO, NSGA-III, and DREA, are the average performer, and H-RVEA is the worst performer.


Author(s):  
Rizk M. Rizk-Allah ◽  
Aboul Ella Hassanien

This chapter presents a hybrid optimization algorithm namely FOA-FA for solving single and multi-objective optimization problems. The proposed algorithm integrates the benefits of the fruit fly optimization algorithm (FOA) and the firefly algorithm (FA) to avoid the entrapment in the local optima and the premature convergence of the population. FOA operates in the direction of seeking the optimum solution while the firefly algorithm (FA) has been used to accelerate the optimum seeking process and speed up the convergence performance to the global solution. Further, the multi-objective optimization problem is scalarized to a single objective problem by weighting method, where the proposed algorithm is implemented to derive the non-inferior solutions that are in contrast to the optimal solution. Finally, the proposed FOA-FA algorithm is tested on different benchmark problems whether single or multi-objective aspects and two engineering applications. The numerical comparisons reveal the robustness and effectiveness of the proposed algorithm.


Sign in / Sign up

Export Citation Format

Share Document