1 Introduction

Over the past decade, the dimensional space of optimization problems has grown significantly in real-world applications [1, 2]. Consequently, large-scale global optimization problems (LSGOPs), involving at least thousands of decision variables [3, 4], have emerged as a vibrant area of research. The dimensionality of the decision variables is a major factor in the complexity of optimization problems. The exponential growth in the size of the search space with respect to the number of decision variables affects the number of candidate solutions in the search space. Generally, the dimensionality of the decision variables also has a direct impact on the computational cost of the optimization and the computational feasibility of detecting correlation between pairs of variables [5]. Therefore, LSGOPs are valuable but challenging to solve. Recently, some attempts have been made for LSGOPs [6].

Evolutionary algorithms (EAs) have demonstrated significant success in solving complex optimization problems [7,8,9,10,11,12,13]. However, due to the ”curse of dimensionality” [14], it is still difficult for traditional EAs to address LSGOPs [15, 16]. Consequently, numerous researchers have undertaken efforts to develop EAs tailored for LSGOPs [15, 17]. These algorithms are commonly known as large-scale evolutionary algorithms (LSEAs). The existing LSEAs can be roughly categorized into metaheuristics and divide-and-conquer methods [18].

The metaheuristics LSEAs mainly focus on enhancing the searchability of algorithms. In other words, various schemes are designed to maintain population diversity in the large-scale space. For instance, Hansen and Ostermeier [19] developed a covariance matrix adaptation evolution strategy (CMA-ES) algorithm utilizing two evolution paths to preserve diversity. Ros and Hansen [20] proposed a variant of CMA-EA, called sep-CMA-ES, to alleviate the time cost associated with covariance matrix calculation in CMA-ES. Additionally, Molina et al. [21] introduced a memetic algorithm, named MA-SW-Chains, which assigns each search intensity to each individual through chaining different local search operators. LaTorre et al. [22] presented a multi-offspring generation framework (MOS) that combines a Genetic Algorithm (GA) with two local searches. Moreover, several algorithms based on particle swarm optimization (PSO) have been proposed. Specifically, Cheng and Jin [23] proposed a competitive particle swarm algorithm (CSO) employing a paired competition mechanism to generate the next population. They also presented a social learning particle swarm optimization algorithm (SL-PSO) [24], which updates the positions of particles by learning from those with better fitness values. Jian et al. [25] proposed a local search strategy, ARS, based on a novel region-based encoding scheme to enhance SL-PSO, forming the SLPSO-ARS. Wang et al. [26, 27] proposed two variants of distributed PSO (DPSO), named dynamic group learning DPSO (DGLDPSO) and adaptive granularity learning DPSO (AGLDPSO). These algorithms increase population diversity by dynamically changing the structure of the subpopulations. Furthermore, Ge et al. [28] developed a novel distributed differential evolution (DE) algorithm with an adaptive population model, named DDE-AMS. Hadi et al. [29] proposed a hybrid algorithm, MLSHADE-SPA, based on three DE strategies and a modified Multiple Trajectory Search (MTS). Later, Koçer and Uymaz [30] proposed an improved MLSHADE-SPA (IMLSHADE-SPA) with a novel local search method. Zhang and Lan [31] introduced a memetic algorithm, MPCE & SSALS, utilizing a multiparent crossover evolution strategy and a step-size adaptive local search method for local exploitation. Recently, Li et al. [32] proposed a dynamic sine cosine algorithm, named DSCA, which includes a nonlinear random convergence parameter to update the equation dynamically, balancing the exploration and exploitation of SCA. Sanjoy et al proposed an enhanced whale optimization algorithm (eWOA) [33]. In eWOA, a selection parameter is introduced, and the co-efficient vectors in the whale optimization algorithm (WOA) are modified.

In contrast to the first category, the divide-and-conquer LSEAs, mainly cooperative co-evolutionary EAs (CCEAs), decompose original LSGOPs into several sub-components and optimize them separately. CCEAs can effectively reduce the search space, which makes it more conducive to solving complex LSGOPs. The primary challenge in employing CC for solving LSGOPs lies in the choice of the problem decomposition method. Consequently, numerous attempts have been made to design effective decomposition methods. For instance, Van den Bergh and Engelbrecht [34] proposed a decision grouping method where a D-dimensional problem is evenly divided into k small-dimensional sub-problems. To consider the variable interdependencies, Yang et al. [35] proposed a random grouping (RG), which randomly decomposes variables into a fixed number of sub-components at each generation. Subsequently, they developed a method to determine the sub-component sizes from a provided set based on the probability of historical performance measures [36]. Omidvar et al. [37] proposed another competitive decomposition method, called delta grouping, which identifies the interacting variables by measuring the averaged difference in a certain variable across the entire population. It was also extended to an improved version that can adaptively determine the size of sub-components, as in [36]. However, these decomposition methods do not detect the underlying structure of variable interactions. To address this, Omidvar et al. [14] proposed a differential grouping (DG) decomposition method. In DG, the interactions between each variable are identified by detecting the fitness changes when perturbing the variables. Sun et al. [38] developed an extended differential grouping (XDG) method to detect the indirect interactions between variables. Additionally, Mei et al. [39] proposed a global differential grouping (GDG) method. Later, Omidvar et al. [40] presented an improved version of DG, named DG2, which exhibits better efficiency and grouping accuracy than DG. Recently, Sun et al. [41] developed a recursive differential grouping (RDG) method based on the bisection method, which significantly improved the efficiency of problem decomposition in terms of time complexity. Chen et al. [42] proposed an efficient adaptive differential grouping algorithm (EDGA) for large-scale black-box optimization problems. In addition, some algorithms [43,44,45] are proposed to allocate different computing resources based on the contribution of the sub-problems, and others [46, 47] based on modified CC technique are also developed.

The CCEAs have proven successful in addressing LSGOPs, which can effectively reduce problem difficulty through a divide-and-conquer approach. The full potential of CCEAs is realized by meticulously designing efficient decomposition and evolutionary search methods. As mentioned above, the decomposition methods can be classified into two different approaches, namely, the manual decomposition method [34,35,36,37] and the automatic decomposition method [14, 38,39,40,41,42]. In manual decomposition methods, the number of sub-components is manually determined. These methods work well on fully separable problems while exhibiting shortcomings in solving non-separable problems due to they do not identify the underlying structure of variable interactions. Conversely, automatic decomposition methods automatically assign decision variables to different sub-components based on their underlying structure, in which the interacting variables can be placed into the same sub-component. However, a drawback is that the decomposed sub-components will remain unchanged during the evolution, with only the best-so-far solutions exchanged between different of them. This makes the automatic decomposition methods insufficient to compensate for information. In particular, there are often challenges in accurately decomposing real-world problems. Moreover, these methods are also unsuitable for both fully separable and fully non-separable problems, as they will place only one variable in each sub-component or group all variables into one big sub-component. Therefore, an effective decomposition method that strikes a balance between the two approaches is needed. Compared with the decomposition method, the evolutionary search method is equally crucial for enhancing CCEA performance, but limited attention has been devoted to improving these search methods.

To address the above issues, this paper proposes an evolutionary dynamic grouping based cooperative co-evolution algorithm, named EDGCC. The proposed EDGCC mainly aims to enhance the informative collaborators among sub-components and improve the searchability of the algorithm in solving LSGOPs. The major contributions of this paper are summarized as follows:

  1. (1)

    An evolutionary dynamic grouping (EDG) method is proposed. In EDG, the sub-components are dynamically generated based on the designed selection, crossover, mutation operations, and an adaptive frequency scheme. It considers both the underlying structure of variable interactions and the transfer of information between different sub-components.

  2. (2)

    An evolutionary search method based on the fireworks search strategy is proposed. In this strategy, the operations of generating offspring populations in the fireworks algorithm are first time introduced to the CCEAs for LSGOPs. It is performed after each cycle of sub-component optimization, which works with the sub-component optimizer to enhance the diversity of the population.

  3. (3)

    An evolutionary dynamic grouping-based cooperative co-evolution (EDGCC) algorithm is proposed for LSGOPs. To demonstrate the effectiveness of EDGCC, a comprehensive empirical study has been conducted. The experimental results indicate that the proposed algorithm achieves promising performance compared with four state-of-the-art methods.

The rest of this paper is organized as follows. In Section 2, a review of related work on the proposed algorithm is provided. In Section 3, the details of the proposed algorithm are described. Experiments are conducted in Section 4. Finally, the conclusions and future work are given in Section 5.

2 Related work

In this section, first, a brief introduction to the LSGOP is given. Then, the idea of the CC technique is elaborated. Finally, the firework algorithm is described.

2.1 Large-scale global optimization problems

Without loss of generality, an LSGOP can be formulated as follows:

$$\begin{aligned} minf\left( \textrm{x} \right) , \textrm{x}=\left[ x_{1},x_{2},\cdots ,x_D\right] \in \mathcal {X}. \end{aligned}$$
(1)

where \(f\left( \textrm{x} \right) \) is the objective function, \(\textrm{x}=\left[ x_1,x_2,\cdots ,x_D\right] \) denotes a D-dimensional decision vector, \(\mathcal {X} \in \mathbb {R} ^{D} \) indicates the feasible solution set. In the LSGOP, \(D \ge 1000\) [48].

There are three categories for LSGOPs according to the interaction relationship among variables, namely, fully separable LSGOP, partially separable LSGOP, and fully non-separable LSGOP. The separable LSGOP is defined as follows:

$$\begin{aligned} \begin{aligned}&\textrm{arg} \underset{\textrm{S} }{min} f\left( \textrm{x} \right) = \\&\left( \textrm{arg} \underset{\mathrm {S_1}}{min} f\left( x_1 ,\cdots \right) ,\cdots , \textrm{arg} \underset{\mathrm {S_m}}{min} f\left( \cdots , x_m \right) \right) . \end{aligned} \end{aligned}$$
(2)

where \(\left\{ \mathrm {S_1}, \mathrm {S_2} \cdots , \mathrm {S_m} \right\} \) represents the m disjoint sub-compo-nents of x. If \(m=D\), the LSGOP defined by (2) is a fully separable LSGOP, and if \(m=1\), the LSGOP defined by (2) is a fully non-separable LSGOP. However, LSGOP defined by (2) is a partially separable LSGOP when \(m \ne D\) and \(m \ne 1\).

2.2 Cooperative co-evolution technique

The cooperative co-evolution(CC) technique is a popular approach for tackling large-scale optimization problems [49,50,51]. It is attributed to the divide-and-conquer strategy, which decomposes a large-scale problem into numbers of smaller sub-problems and optimizes them separately. The performance of CCEAs for solving LSGOPs is sensitive to the choice of decomposition method. As reviewed in Section 1, a number of decomposition methods have been proposed. In this subsection, we detail the recently developed recursive differential grouping (RDG) method [41], which can efficiently and accurately decompose the problems based on the interaction of decision variables. The detailed description of RDG is as follows:

Notation: Let X be the set of decision variables \(\left\{ x_1,\cdots ,\right. \) \(\left. x_D \right\} \) and \(U_X\) be the set of unit vectors in the decision space \(\mathbb {R} ^D\). Let \(X_1\) be a subset of variables X and \(U_{X_1}\) be a subset of unit vectors \(U_X\). For any unit vector \(\textbf{u} = \left( u_1,\cdots ,u_D \right) \in U_{X_1} \), we have

$$\begin{aligned} u_i=0,\textrm{if} x_i\notin X_1. \end{aligned}$$
(3)

Theorem: Let \(f:\mathbb {R} ^D\rightarrow \mathbb {R} \) be an objective function; \(X_1\subset X\) and \(X_2\subset X\) be two mutually exclusive subsets of decision variables: \(X_1\cap X_2=\phi \). There is some interaction between decision variables in \(X_1\) and \(X_2\), if there exist two unit vectors \(\textbf{u} _1 \in U_{X_1} \) and \(\textbf{u} _2 \in U_{X_2} \), two real numbers \(l_1,l_2> 0\), and a candidate solution \(\textbf{x}^*\) in the decision space, that satisfied the following:

$$\begin{aligned} \begin{aligned}&f\left( \textbf{x}^* +l_1 \textbf{u}_1+l_2 \textbf{u}_2\right) - f\left( \textbf{x}^* +l_2 \textbf{u}_2\right) \ne \\&f\left( \textbf{x}^* +l_1 \textbf{u}_1\right) -f\left( \textbf{x}^* \right) . \end{aligned} \end{aligned}$$
(4)

If there is an interaction between \(X_1\) and \(X_2\), \(X_2\) will be further divided into two equal-sized and mutually exclusive subsets. Then, the interactions between \(X_1\) and these subsets are examined. The above process is repeated until RDG finds all the variables that interact with \(X_1\). However, if the formula is not held, \(X_1\) and \(X_2\) will be determined as mutually separable sets.

In RDG, the interrelationship is examined between a pair of sets of variables but not a pair of variables. During this binary search procedure, if there is not an interaction between \(X_1\) and a subset, RDG does not further examine the interrelationship between \(X_1\) and the variables in this subset. Therefore, RDG greatly improved the efficiency of the time complexity compared with DG [14] and DG2 [40]. According to [41], the computational complexity of RDG is \(O\left( D\textrm{log}_2 D \right) \) for decomposing a D-dimensional problem, while DG is \(D^2 + D\), and DG2 is \(\left[ \left( D^2+D+2 \right) /2 \right] \).

2.3 Fireworks algorithm

The fireworks algorithm (FWA) [52] is a swarm intelligence algorithm inspired by the phenomenon of fireworks explosion in the night sky. Liu et al. [53] have presented theoretical analysis and proved that FWA is an absorbing Markov stochastic process. In recent years, there are many pieces of research for FWA in solving practical problems [54, 55].

FWA mainly follows the general framework of EAs. However, it proposes a new search manner that imitates the process of fireworks explosion to search the potential space by a stochastic explosion process within the local space. In FWA, there are three main operations, namely, explosion, Gaussian mutation, and selection. At first, several fireworks are initialized randomly. Then, explosion sparks of these fireworks are generated by the explosion operation. To maintain the diversity of the population and balance the global and local search, the explosion amplitude and the population of the generated sparks are different between fireworks. Specifically, a firework with lower fitness has a higher explosion amplitude, which generates a smaller population within a more extensive range. On the contrary, a firework with better fitness will have a lower explosion amplitude and a larger population. In other words, exploration is achieved by those fireworks with a large explosion amplitude to escape from local minima, while exploitation is achieved by those fireworks with a small explosion amplitude to reinforce the local search ability in promising areas. Moreover, the mutation operation mutates the locations of the fireworks to generate mutation sparks, which are used to further improve the diversity of the swarm. Finally, the selection operation selects among the set of the generated two types of sparks and the original fireworks for the next generation.

3 The proposed algorithm

In this section, the main framework of the proposed algorithm EDGCC is first elaborated. Then, the evolutionary dynamic grouping (EDG) method and the fireworks search strategy as two critical components of EDGCC are introduced in detail.

3.1 Main framework of EDGCC

The main framework of the proposed EDGCC is given in Algorithm 1. It starts by randomly initializing a population, denoted as P, with size N, in which each solution consists of D decision variables. Then, the RDG [41] method is employed to detect the interaction of all decision variables. The RDG method obtains several non-separable variable groups and a separable variable group, which are denoted as nonseps and seps, respectively.

In the main loop, to begin with, the sub-components are generated by the EDG method. Subsequently, all sub-components are optimized by the sub-component optimizer one by one. Meanwhile, the function value improvement \(\varDelta F\) of each sub-component is calculated by subtracting the function values of the best solutions in the original population and the optimized new population. This is utilized for the selection operation of the EDG method. Finally, the fireworks search strategy is executed to further enhance the exploration and exploitation of the sub-component optimizer. It is essential to note that the meaning of parameters in the fireworks search strategy is shown in Table 1. The above optimization process is iterated until the termination condition is satisfied. The subsequent sections elaborate on the EDG method and the fireworks search strategy.

Algorithm 1
figure g

Main Framework of the Proposed EDGCC.

Table 1 The parameters of fireworks search strategy

3.2 Evolutionary dynamic grouping method

In this subsection, the designed evolutionary dynamic grouping (EDG) method is elaborated. Its main idea is to generate sub-components dynamically, just like the implementation process of the genetic algorithm. Algorithm 2 presents the details of the designed EDG method, which includes two parts, i.e., sub-component initialization and sub-component evolution.

At the beginning of EDG, the decision variables are decomposed into several sub-components, which are used for the initial optimization. During this process, the sub-components are decomposed based on the variables preprocessing by the RDG method. Depending on the detection results of the RDG method, the decision variables are divided into three situations corresponding to different sub-component decomposition methods: 1) for fully separable or fully non-separable problems, all variables are randomly and uniformly divided into several sub-components, where the dimension of each sub-component is s; 2) for a partially separable problem and the number of separable variables \(\left| seps \right| \) is half or more of D, each non-separable group is regarded as a sub-component, and the separable variables are randomly and uniformly divided into \(\left| nonseps \right| \) sub-components; 3) otherwise, each non-separable group is regarded as a sub-component, and all separable variables are placed in one sub-component. It is worth noteworthy that the first two situations in the above variable decomposition differ from the RDG method, while the third situation is the same as the RDG method. This distinction is intentional. Specifically, the sub-components in the first situation are advantageous for the crossover and selection operators, whereas the sub-components decomposed by the RDG method make crossover and selection operations impractical. The second situation aims to prevent an imbalance in the number of variables between the separable variable sub-component and other sub-components, which easily appears in the RDG method. Thus, the sub-component of separable variables is further divided in the second situation. Following sub-component decomposition, the variables within each sub-component are further randomly divided into m groups, which serve as the genes for the crossover operation.

After the sub-component initialization, sub-component evolution is performed to dynamically generate the sub-components that are conducive to improving the searchability, as the optimization process proceeds. To visually present the sub-component evolution, Fig. 1 visually illustrates the sub-component evolution, which contains the selection, crossover, and mutation operations. Specifically, two sub-components (i and j) are first selected based on having the smallest function value improvement \(\varDelta F\). The motivation for such selection is the contribution of these two sub-components to the optimization is poor, and thus the variables of these two sub-components should be decomposed again. Subsequently, the crossover operation is performed to generate new sub-components by exchanging variables in a randomly chosen group k from the selected sub-components i and j. Finally, a mutation operation is proposed to enhance the variable diversity in the groups. In this operation, a group g of the sub-component u is randomly selected, and a random variable \(x_d\) is added to this group.

Fig. 1
figure 1

The diagram of sub-components evolution, which contains the selection, crossover and mutation operations

Fig. 2
figure 2

The illustration of the function used in the adaptive frequency scheme. (a) The function image. (b) The function values

In general, the early phase of an algorithm should emphasize the exploration ability, while the later stages should emphasize the exploitation ability. In EDG, the sub-components are dynamically generated to enhance the transfer of information between different sub-components as optimization progresses. A higher frequency of information transfer implies that more information is utilized to guide the generation of offspring populations, resulting in improved exploration ability. Conversely, a lower frequency of information transfer is more conducive to enhancing the exploitation ability of the algorithm. Therefore, an adaptive frequency scheme of the sub-component evolution is designed to adjust the search focus. In this scheme, the frequency of sub-component evolution is higher in the early stages of the algorithm and decreases as the algorithm progresses. Specifically, if the cycle number of sub-component optimization belongs to an assigned set \(\hat{S} \), the sub-component evolution is performed. The cycle number set \(\hat{S} \) is assigned in the following manner:

$$\begin{aligned} \hat{S} = \left\{ \hat{S} ;(cyc^2+cyc+2)/2 \right\} ,Wang2022 \end{aligned}$$
(5)

where cyc is the number of sub-component optimization cycles. In other words, \(cyc = \left\{ 1, 2, \cdots , cyc_{max}\right\} \), where \(cyc_{max}\) represents the maximum value of the cycle number when the termination condition is reached. To have a clear understanding of it, Fig. 2 gives an illustration. Figure 2(a) shows a part of the function \( f(x) = (x^2+x+2)/2 \) and Fig. 2(b) gives the values of f(x) when x is set to \(x= \left\{ 0, 1, 2, 3, 4, 5 \right\} \). It is clear that the interval between the function values of two adjacent variables gradually increases. Therefore, the probability of sub-component evolution is greater at the early stage of the optimization process, while the possibility gets smaller as the optimization proceeds.

Algorithm 2
figure h

The Evolutionary Dynamic Grouping Method.

3.3 Fireworks search strategy

After the EDG method, the sub-components are optimized by the sub-components optimizer. To further enhance the performance of the algorithm, an evolutionary search method based on the fireworks search strategy is performed. Algorithm 3 provides the details of the fireworks search strategy. At first, snum solutions in the current population P are randomly selected as the initial fireworks. Their quality (i.e., fitness) is evaluated to determine the number of explosion sparks and the explosion amplitudes. Then, the fireworks explode to generate several ”explosion sparks” within their local space. Later, the ”Gaussian sparks” are generated by conducting the search in a local Gaussian space around gnum randomly selected fireworks. It should be noted that if the generated spark exceeds the search range, it will be mapped to another location in the search space by a mapping strategy. Finally, to retain the information and pass it to the next iteration, the new solutions are selected from the set of the initial fireworks and generated sparks to replace the original solutions of P.

In this paper, the operations of an enhanced version of FWA (EFWA) [56] is introduced to the fireworks search strategy. Compared to the traditional FWA, EFWA incorporates five modifications. Specifically, for each dimension k, the explosion amplitude \(A_{i}^{k}\) is bound as follows:

$$\begin{aligned} A_{i}^{k} =\left\{ \begin{matrix} A_{min}^{k} &{} if A_{i}^{k}< A_{min}^{k},\\ A_{i}^{k} &{} otherwise, \end{matrix}\right. \end{aligned}$$
(6)

where,

$$\begin{aligned} A_{min}^{k}\left( t \right) =A_{init}-\frac{A_{init}-A_{final}}{evals_{max}} \sqrt{\left( 2*evals_{max}-t \right) t }, \end{aligned}$$
(7)

and

$$\begin{aligned} A_i=\hat{A} \cdot \frac{f\left( X_i \right) -y_{min}+\varepsilon }{ {\textstyle \sum _{i=1}^{N}\left( f\left( X_i \right) -y_{min} \right) +\varepsilon } }, \end{aligned}$$
(8)

t represents the number of function evaluations, \(evals_{max}\) denotes the maximum number of evaluations, \(A_{init}\) and \(A_{final}\) are the initial and final minimum explosion amplitude, \(\hat{A}\) represents the constant to control the explosion amplitude, \(y_{max}=max\left( f\left( X_i \right) \right) \), \(y_{min}=min\left( f\left( X_i \right) \right) \), \(\varepsilon \) is the machine epsilon. The lower bound \(A_{min}\) of the explosion amplitude is proposed to prevent the explosion amplitude from being too small that the locations of the explosion sparks are almost the same as the firework themselves. This paper uses the non-linear decreasing function as \(A_{min}\). In the early stages of the search, \(A_{min}\) is set to a high value to facilitate exploration. As the number of evaluations increases, it is lowered to allow for better exploitation capabilities around good locations. Moreover, the number of explosion sparks s for each firework \(X_i\) is calculated as follows:

$$\begin{aligned} s_i=M_e \cdot \frac{y_{max}-f\left( X_i \right) +\varepsilon }{ {\textstyle \sum _{i=1}^{N}\left( y_{max} -f\left( X_i \right) \right) +\varepsilon } }, \end{aligned}$$
(9)

where \(M_e\) represents the constant to control the number of explosion sparks. To avoid the decisive influence of the fireworks in good positions, the number of sparks is determined by the following:

$$\begin{aligned} s_i=\left\{ \begin{matrix} round\left( aM_e \right) &{} \textrm{if} s_i<aM_e, \\ round\left( bM_e \right) &{} \textrm{if} s_i>bM_e, \\ round\left( s_i \right) &{} otherwise. \end{matrix}\right. \end{aligned}$$
(10)

where a and b are constant parameters that confine the range of the ”explosion sparks” size.

Furthermore, the operation that generates the ”explosion sparks” is to add different offset displacements \(\varDelta X^k\) in each dimension. And the Gaussian mutation operation uses the \(\tilde{X}_{i}^{k} = \tilde{X}_{i}^{k} + (\tilde{X}_{B}^{i} - \tilde{X}_{i}^{k}) *e\) to generate "Gaussian sparks", where \(X_B\) is the location of the currently best firework or explosion spark found so far, and \(e=\mathcal {N} \left( 0,1 \right) \). This mutation operation will expand along the direction between the current firework position and the optimal firework position, which ensures diversity in the search as well as the global movement to find the best location found so far. If the generated sparks are out of bounds, the uniform random mapping strategy \(\bar{X_{i}^{k} } = X_{min}^{k}+ rand *\left( X_{max}^{k} -X_{min}^{k} \right) \) is used. This mapping strategy can avoid focusing the locations of the mapping on the origin.

Fig. 3
figure 3

The mean function values of different parameter m on \(f_{3}\), \(f_{4}\), \(f_{15}\) of CEC’2010. Statistical results are obtained by 25 independent runs. (a) The function values on \(f_{3}\). (b) The function values on \(f_{4}\). (c) The function values on \(f_{15}\)

In addition, the \(Elitism-Random selection\) (ERP) [57] operation is introduced to select the solution. In this operation, the best candidate solution is selected first. Then, the other individuals are selected randomly.

Algorithm 3
figure i

The Fireworks Search Strategy.

4 Experimental study

In this section, a comprehensive experimental study is conducted to examine the performance of the proposed EDGCC. First, the experimental settings and the parameter sensitivity analysis are presented. Subsequently, a set of experiments are conducted to evaluate the performance of the proposed EDGCC. Finally, a discussion is given.

Table 2 Comparisons among EDGCC and compared algorithms on the CEC’2010 benchmark functions
Table 3 Comparisons among EDGCC and compared algorithms on the CEC’2013 benchmark functions

4.1 Experimental settings

  1. 1)

    Compared Algorithms: Four recently proposed state-of-the-art LSEAs, namely, DSCA [32], eWOA [33], DECC-RDG [41] and EDGA [42] are used for comparison. The DSCA and eWOA are the metaheuristics LSEAs. The EDGA and DECC-RDG are the divide-and-conquer LSEAs.

  2. 2)

    Benchmark Suites: Two widely used large-scale global optimization benchmark suites are used. The first one is IEEE CEC’2010 [58], which contains 20 test functions. These functions can be classified into fully separable LSGOPs \(\left( f_1-f_3 \right) \), partially separable LSGOPs \(\left( f_4-f_{18} \right) \), and fully non-separable LSGOPs \(\left( f_{19}\!-\!f_{20} \right) \). The other one is IEEE CEC’2013 [59], which contains 15 test functions. These functions can be classified into fully separable LSGOPs \(\left( f_1-f_3 \right) \), partially separable LSGOPs \(\left( f_4-f_{11} \right) \), and fully non-separable LSGOPs \(\left( f_{12}-f_{15} \right) \). The number of decision variables of each test functions is 1000 in both benchmark suites except for the functions \(f_{13}\) and \(f_{14}\) in CEC’2013, have 905 decision variables.

  3. 3)

    Parameters: In the EDG method, as recommended by [35], s is set to 100. Moreover, the group number of each sub-component m is set to 2. In the fireworks search strategy, all parameters are set the same as recommended in [56], where \(snum=5\), \(gnum=5\), \(M_e=50\), \(\hat{A}=40\), \(a=0.04\), \(b=0.8\), \(A_{init}= \left( X_{max}^{k} - X_{min}^{k} \right) \times 0.02\), and \(A_{final}= \left( X_{max}^{k} - X_{min}^{k} \right) \times 0.001\).

  4. 4)

    Stopping Condition and Population Size: The population size N of the algorithms is set to 50. In each run, the maximum number of function evolutions for each algorithm is set to \(3\times 10^{6}\).

  5. 5)

    Performance Metrics: The mean and standard deviation of the function value obtained from 25 independent runs on each problem. To test the difference for statistical significance, the Wilcoxon rank-sum test with a significance level of 0.05 is employed to assess whether one algorithm is better than another in terms of the function value. The symbols \("+"\), \("-"\), \("\approx "\) are used to indicate that the compared algorithm is significantly better than, worse than, and similar to EDGCC, respectively.

The sub-component optimizer in this paper is implemented by SaNSDE [60], which is the same as the compared DECC-RDG. It should be noted that only m is the parameter introduced by the proposed algorithm. The sensitivity analysis of m is given in the following subsection.

4.2 Parameter sensitivity analysis

In the proposed EDGCC, there is a major parameter, i.e., the group number of each sub-component m, which is used to determine the number of exchange variables in the crossover operation of the EDG method. In this study, some comparisons are designed to study the influence of the above parameter on the performance of EDGCC. Specifically, four values for m are chosen to test the performance of these different parameter settings.

Figure 3 plots the mean function values of each parameter, where m is set to 2,4,6,8 on \(f_3\), \(f_4\), and \(f_{15}\) of CEC’2010 benchmark suit, respectively. From Fig. 3, it can be seen that the different parameters can result in various performances on test benchmark problems. Generally, the function value increases as the increase of m. This is attributed to the difference between sub-components generated by the crossover operator and the original sub-components gradually decreases as the increase of m. Consequently, it can be recommended that \(m=2\) can be used as a general parameter setting for the proposed EDGCC.

4.3 Results on benchmark suits

In this subsection, the comparison experiments on benchmark suits are conducted. Tables 2 and 3 present the mean and standard deviation of the function value obtained from 25 independent runs on the CEC’2010 and CEC’2013 benchmark suite, respectively. The best result on each function is shown in the bold typeface.

As shown in Tables 2 and 3, EDGCC demonstrates superior performance, which achieves the best results on 19 out of 35 functions across two benchmark suites. Followed by EADG, which achieves the best on 10 functions. Moreover, DECC-RDG achieves the best on 4 functions, eWOA achieves the best on 2 functions, and DSCA on none. Specifically, for CEC’2010, EDGCC outperforms others in 13 functions, while EADG and DECC-RDG lead in 4 and 3 functions, respectively. For CEC’2013, both EDGCC and EADG excel in 6 functions, while eWOA and DECC-RDG demonstrate superiority in 2 and 1 functions, respectively. In general, the effectiveness of EDGCC and the other two CCEAs surpasses that of the compared metaheuristic algorithms. This can be attributed to the fact that the CCEAs use variable decomposition methods to reduce the search space.

Table 4 Comparisons among EDGCC and two variants on the CEC’2010 benchmark functions
Fig. 4
figure 4

The convergence curves of EDGCC, and two variants FCC, CC-EDG for functions \(f_3\) and \(f_{10}\). (a) The convergence curves on \(f_3\). (b) The convergence curves on \(f_{10}\)

Detailed view of each algorithm, the proposed EDGCC outperforms them in most functions. Specifically, in comparison to other metaheuristic algorithms, EDGCC is significantly better than eWOA only except for \(f_6\) and \(f_{10}\) in CEC’2013. Similarly, EDGCC outperforms DSCA across all functions except for \(f_6\) in CEC’2013. In comparison to the other CCEAs, EDGCC is better than EADG on 18 out of 35 functions and is statistically similar to EADG on 6 functions. Moreover, EDGCC outperforms DECC-RDG on 26 out of 35 functions and is defeated on only 6 functions.

The performance of EDGCC in solving fully separable LSGOPs (i.e., \(f_1-f_3\) in CEC’2010 and CEC’2013) and fully non-separable LSGOPs (i.e., \(f_{19}-f_{20}\) in CEC’2010 and \(f_{12}-f_{15}\) in CEC’2013) is better than all the compared algorithms. Additionally, for partially separable LSGOPs (i.e., \(f_4-f_{18}\) in CEC’2010 and \(f_4-f_{11}\) CEC’2013), the proposed EDGCC also performs well on the majority of them.

The above statistical comparisons confirm that the proposed EDGCC is more effective in solving the LSGOPs than the compared state-of-the-art LSEAs. This can be attributed to the fact that adequate information exchange between different sub-components and an effective search strategy is necessary.

4.4 The effectiveness analysis of two important components

The proposed EDGCC has higher competitiveness in solving LSGOPs, which can contribute to the cooperation of the main components, i.e., EGD method and fireworks search strategy. To verify their respective effectiveness, two variants of EDGCC are designed and compared with the original EDGCC on the CEC’2010 benchmark suite, where two variants are respectively called CC-EGD, and FCC. In CC-EDG, the EGD method is preserved, while the fireworks search strategy is removed. Its purpose is to investigate the effectiveness of the fireworks search strategy. In FCC, the EGD method is abandoned. Instead, the fireworks search strategy is preserved, which aims to investigate the effectiveness of the EGD method. Note that all parameters in these three algorithms are consistent for a fair comparison.

Table 4 presents the comparison results among the original EDGCC and its two variants on the CEC’2010 benchmark suit. According to the table, EDGCC achieves the best results on 10 out of 20 test problems, while CC-EGD and FCC outperform in 6 and 4 test problems, respectively. This indicates that the two main components in EDGCC are important to the performance of EDGCC, which cooperate to make the EDGCC perform efficiently. Additionally, convergence curves obtained from EDGCC and its two variants on functions \(f_3\) and \(f_{10}\) are depicted in Fig. 4, which can visually illustrate the distinctions. From the figure, it is evident that EDGCC outperforms its variants throughout the evolutionary process. In summary, the synergy of the main components makes the proposed algorithm perform noticeably.

4.5 Computational complexity and time of the EDG method

The EDG method consists of two main components, i.e., sub-component initialization and sub-component evolution. The initial sub-components of decision variables are generated based on the RDG method, which is a recently proposed variable grouping method with low computational complexity. As shown in [41], the computational complexity of the initialization for a D-dimensional problem is O(Dlog(D)).

Following the sub-component initialization, the initial sub-components are dynamically evolving as the optimization process proceeds. This is an additional part compared with static decomposition methods. Table 5 gives the computational time of the sub-component evolution on CEC’2013 benchmark functions. It can be seen that the computational time of this part is very small. In essence, compared with static decomposition methods, the EDG method does not incur significant additional computational time.

Table 5 Computational time of the sub-component dynamic evolution on CEC’2013 benchmark functions
Table 6 Comparisons among EDGCC and compared algorithms on the KP with 100, 200, 300, 500, and 1000 decision variables

4.6 Application in real-world problem

To further validate the effectiveness of the proposed EDGCC, the experimental study of the real-world optimization problem, i.e., the knapsack problem (KP), is conducted. In KP, there is a set of goods, which is associated with weight and profit values, and a backpack, which is associated with maximum capacity. The objective of the KP is to maximize the total value of a subset of goods within the constraints of the capacity of the knapsack. Therefore, it is a constrained optimization problem. The specific settings and Matlab code for the KP are available on PlatEMO [61]. The number of decision variables of KP D is set to 100, 200, 300, 500, and 1000. The maximum number of function evaluations for each problem is set to 3000D.

The result obtained by each algorithm from 25 independent runs on KP is presented in Table 6. According to the results, it can be seen that the proposed EDGCC achieves the best performance of all KPs with different variables. The EADG and DECC-RDG obtain similar results. However, eWOA cannot converge to the optimal solution at all, and the DSCA cannot find a solution that satisfies the constraint conditions in any runs. Generally, the effect of the CCEAs is better than the two metaheuristics algorithms. This observation is the same as the results that obtained on the benchmark suits.

In summary, the proposed EDGCC demonstrates a clear advantage compared with the peer competitors on the real-world KP.

4.7 Discussion

The above statistical comparisons confirm the effectiveness of the proposed EDGCC for solving LSGOPs. This can be attributed to the following reasons. First, with practical information compensation by the EDG method, information is transferred between different sub-components. Sub-components that are more favorable to the current optimization phase are generated. Additionally, the fireworks search strategy further improves the diversity of the population.

Although the proposed EDGCC has shown encouraging results in most test instances, the current study still has some limitations as well. First, many of the parameters in this paper are set as other relevant papers, while not analyzing them in greater depth. For example, for fully separable and fully non-separable problems, the decision variables are randomly divided into some initial sub-components. The size of each sub-component, which has a significant impact on the performance of the algorithm, is set as recommended by [35]. Therefore, the proposed EDGCC only achieves the best on 3 out of 6 fully non-separable problems, which fails to achieve the desired result. Additionally, it should be noted that the fireworks search strategy will consume the function evolution and potentially hamper the convergence speed of the algorithm. Although the limitation of function evolution frequency was not explicitly considered when tackling large-scale optimization problems, where the maximum number of function evaluations is set to be very large, it is necessary to design strategies to enhance population convergence.

5 Conclusion

The performance of the CC technique for solving LSGOPs will be influenced by the decomposition methods. However, most existing decomposition methods divide decision variables into fixed sub-components, which may not be conducive to the information transfer between the variables in different sub-components. Therefore, this paper proposes an evolutionary dynamic grouping based cooperative co-evolution algorithm, named EDGCC.

In EDGCC, a novel variables decomposition method EDG is proposed to generate new sub-components dynamically in the optimization process. The idea of the EDG method is motivated by the evolution process of the genetic algorithm, while the selection, crossover, and mutation operations are redefined. To adjust the search focus in different stages of the algorithm, an adaptive frequency scheme of EDG is also proposed. Additionally, the fireworks search strategy is introduced to work with the sub-components optimizer, which can further improve the diversity of the population.

To examine the performance of the proposed EDGCC, a comprehensive experimental study is conducted. First, the sensitivity analysis of the unique parameter, i.e., the group number of each sub-component m. From the result, it can be seen that the performance of EDGCC will deteriorate as m increases. Therefore, \(m=2\) is used as a general parameter setting in this paper. Then, the comparisons among EDGCC and four recently developed LSEAs on two widely used CEC’2010 and CEC’2010 benchmarks are tested. Experimental results show the effectiveness of EDGCC, in which EDGCC achieves the best on 26 out of 35 functions. In order to verify the effectiveness of the EDG method and fireworks search strategy, EDGCC has been compared with two variants on CEC’2010. From the result, it can be seen that EDGCC obtains better convergence than the compared two variants. Moreover, the computational time of the sub-component dynamic evolution is tested. Finally, EDGCC is applied to solving the real-world KP with 100, 200, 300, 500, and 1000 decision variables. The experimental results show that the EDGCC algorithm has excellent performance in solving KP.

In the future work, we would like to focus on investigating if the EDG method can be extended for large-scale multi-objective optimization problems and developing the fireworks algorithm for large-scale problems.