Cooperative Coevolution with Formula-Based Variable Grouping for Large-Scale Global Optimization

2018 ◽  
Vol 26 (4) ◽  
pp. 569-596 ◽  
Author(s):  
Yuping Wang ◽  
Haiyan Liu ◽  
Fei Wei ◽  
Tingting Zong ◽  
Xiaodong Li

For a large-scale global optimization (LSGO) problem, divide-and-conquer is usually considered an effective strategy to decompose the problem into smaller subproblems, each of which can then be solved individually. Among these decomposition methods, variable grouping is shown to be promising in recent years. Existing variable grouping methods usually assume the problem to be black-box (i.e., assuming that an analytical model of the objective function is unknown), and they attempt to learn appropriate variable grouping that would allow for a better decomposition of the problem. In such cases, these variable grouping methods do not make a direct use of the formula of the objective function. However, it can be argued that many real-world problems are white-box problems, that is, the formulas of objective functions are often known a priori. These formulas of the objective functions provide rich information which can then be used to design an effective variable group method. In this article, a formula-based grouping strategy (FBG) for white-box problems is first proposed. It groups variables directly via the formula of an objective function which usually consists of a finite number of operations (i.e., four arithmetic operations “[Formula: see text]”, “[Formula: see text]”, “[Formula: see text]”, “[Formula: see text]” and composite operations of basic elementary functions). In FBG, the operations are classified into two classes: one resulting in nonseparable variables, and the other resulting in separable variables. In FBG, variables can be automatically grouped into a suitable number of non-interacting subcomponents, with variables in each subcomponent being interdependent. FBG can easily be applied to any white-box problem and can be integrated into a cooperative coevolution framework. Based on FBG, a novel cooperative coevolution algorithm with formula-based variable grouping (so-called CCF) is proposed in this article for decomposing a large-scale white-box problem into several smaller subproblems and optimizing them respectively. To further enhance the efficiency of CCF, a new local search scheme is designed to improve the solution quality. To verify the efficiency of CCF, experiments are conducted on the standard LSGO benchmark suites of CEC'2008, CEC'2010, CEC'2013, and a real-world problem. Our results suggest that the performance of CCF is very competitive when compared with those of the state-of-the-art LSGO algorithms.

Complexity ◽  
2018 ◽  
Vol 2018 ◽  
pp. 1-15 ◽  
Author(s):  
Xingguang Peng ◽  
Yapei Wu

The cooperative coevolution (CC) algorithm features a “divide-and-conquer” problem-solving process. This feature has great potential for large-scale global optimization (LSGO) while inducing some inherent problems of CC if a problem is improperly decomposed. In this work, a novel CC named selective multiple population- (SMP-) based CC (CC-SMP) is proposed to enhance the cooperation of subproblems by addressing two challenges: finding informative collaborators whose fitness and diversity are qualified and adapting to the dynamic landscape. In particular, a CMA-ES-based multipopulation procedure is employed to identify local optima which are then shared as potential informative collaborators. A restart-after-stagnation procedure is incorporated to help the child populations adapt to the dynamic landscape. A biobjective selection is also incorporated to select qualified child populations according to the criteria of informative individuals (fitness and diversity). Only selected child populations are active in the next evolutionary cycle while the others are frozen to save computing resource. In the experimental study, the proposed CC-SMP is compared to 7 state-of-the-art CC algorithms on 20 benchmark functions with 1000 dimensionality. Statistical comparison results figure out significant superiority of the CC-SMP. In addition, behavior of the SMP scheme and sensitivity to the cooperation frequency are also analyzed.


Algorithms ◽  
2021 ◽  
Vol 14 (5) ◽  
pp. 146
Author(s):  
Aleksei Vakhnin ◽  
Evgenii Sopov

Modern real-valued optimization problems are complex and high-dimensional, and they are known as “large-scale global optimization (LSGO)” problems. Classic evolutionary algorithms (EAs) perform poorly on this class of problems because of the curse of dimensionality. Cooperative Coevolution (CC) is a high-performed framework for performing the decomposition of large-scale problems into smaller and easier subproblems by grouping objective variables. The efficiency of CC strongly depends on the size of groups and the grouping approach. In this study, an improved CC (iCC) approach for solving LSGO problems has been proposed and investigated. iCC changes the number of variables in subcomponents dynamically during the optimization process. The SHADE algorithm is used as a subcomponent optimizer. We have investigated the performance of iCC-SHADE and CC-SHADE on fifteen problems from the LSGO CEC’13 benchmark set provided by the IEEE Congress of Evolutionary Computation. The results of numerical experiments have shown that iCC-SHADE outperforms, on average, CC-SHADE with a fixed number of subcomponents. Also, we have compared iCC-SHADE with some state-of-the-art LSGO metaheuristics. The experimental results have shown that the proposed algorithm is competitive with other efficient metaheuristics.


2019 ◽  
Vol 482 ◽  
pp. 1-26 ◽  
Author(s):  
Ivanoe De Falco ◽  
Antonio Della Cioppa ◽  
Giuseppe A. Trunfio

2016 ◽  
Vol 22 (6) ◽  
pp. 2045-2064 ◽  
Author(s):  
Sedigheh Mahdavi ◽  
Shahryar Rahnamayan ◽  
Mohammad Ebrahim Shiri

Author(s):  
Chao Qian ◽  
Yang Yu ◽  
Ke Tang

Subset selection is a fundamental problem in many areas, which aims to select the best subset of size at most $k$ from a universe. Greedy algorithms are widely used for subset selection, and have shown good approximation performances in deterministic situations. However, their behaviors are stochastic in many realistic situations (e.g., large-scale and noisy). For general stochastic greedy algorithms, bounded approximation guarantees were obtained only for subset selection with monotone submodular objective functions, while real-world applications often involve non-monotone or non-submodular objective functions and can be subject to a more general constraint than a size constraint. This work proves their approximation guarantees in these cases, and thus largely extends the applicability of stochastic greedy algorithms.


Sign in / Sign up

Export Citation Format

Share Document