1 Introduction

Since its introduction in 1995 by Storn and Price [22], the differential evolution (DE) algorithm has been used as a simple yet powerful search technique for solving complex nonlinear and non-differentiable continuous functions. It belongs to the class of stochastic optimization algorithms, which are used to find the best-suited solution to a problem by minimizing an objective function, which is a mapping from a parameter vector \({\overset {\rightharpoonup }{X}}\in {\Re ^{D}}\) to R, within the given constraints and flexibilities.

Starting with a population initialized randomly, the DE algorithm uses simple mutation and crossover operators to generate new candidate solutions, and adopts a one-to-one competitive scheme to determine whether the offspring should replace their parents in the next generation [19]. Owing to its ease of implementation and simplicity, DE has attracted much attention from researchers all over the world, resulting in many variations of the basic algorithm with improved performance. The strategies of the variants often represent varying search capabilities in different search phases of the evolution process.

Most of the approaches to improve the standard DE algorithm mainly concentrate on four aspects: the structure and size of the population, associated control parameter setting, trial vector generation, and hybrid strategies. Of these, control parameter setting and trial vector generation directly affect the search accuracy and convergence speed of the DE algorithm, which is why associated control parameter setting is generally considered together with the trial vector generation strategy. Qin et al. [21] developed a self-adaptive DE (SaDE) algorithm for constrained real-parameter optimization, in which both the trial vector generation strategy and the associated control parameter values are gradually self-adapted according to the learning experience. This algorithm performs much better than both the traditional DE algorithm and several state-of-the-art adaptive parameter DE variants, such as the ADE [32], SDE [18], and JDE [2] algorithms. In [19], a self-adaptive DE algorithm, called SspDE, is proposed with each target individual having its own trial vector generation strategy, scaling factor F, and crossover rate CR, which gradually self-adapt from previous experience in generating promising solutions. Mallipeddi et al. [17] employed an ensemble of mutation strategies and control parameters in their DE algorithm (called EPSDE), in which a pool of distinct mutation strategies coexists with a pool of values for each control parameter throughout the evolution process competing to produce offspring. In [7], the authors designed a heterogeneous distributed algorithm (HdDe) by proposing a new mutation strategy, GPBX- α, and parallel execution in two separate islands using the classic DE/rand/1/bin algorithm. Wang and Zhao [27] presented a DE algorithm with a self-adaptive population resizing mechanism based on JADE [34], called SapsDE. This algorithm gradually self-adapts NP according to previous experience in generating promising solutions and enhances the performance of DE by dynamically choosing one of two mutation strategies and tuning control parameters in a self-adaptive manner. Many hybrid strategies exist that improve the efficiency of the DE algorithms. For the unconstrained global optimization problems, a novel optimization model is proposed, called clustering-based differential evolution with 2 multi-parent crossovers (2-MPCs-CDE) [16], hybridizing DE with the one-step k-means clustering and 2 multi-parent crossovers. Yildiz [31] developed a novel hybrid optimization algorithm called hybrid robust differential evolution (HRDE) by adding positive properties of Taguchis method to the DE algorithm to minimize the production cost associated with multi-pass turning problems.

Although these prominent variants of the DE algorithm largely optimize the DE process, almost all of them still suffer from problems such as difficult parameter setting, slow convergence rate, and premature convergence. Thus, this paper presents a novel adaptive DE algorithm that adopts a dynamic adjustment strategy including a directional strategy and cloud model, which we call ADEwDC. In the ADEwDC algorithm, the whole population is divided into three subgroups with each group dynamically selecting its trial vector generation from a constructed mutation and crossover pool according to its own convergence degree. The proposed algorithm realizes parameter control adaptively by referring to current fitness information. By introducing evolutionary direction [37], the convergence speed of the DE algorithm is further improved. The directional strategy is based on a greedy strategy, which means that individuals with good performance can evolve rapidly in the optimal evolution direction. These good individuals are selected according to their fitness, and the optimal direction is chosen according to the disparity between the target and trial vectors in the fitness space. However, the possibility of the algorithm falling into a local optimum is increased because of the rapid decline in population diversity. Thus, after further analysis, the ADEwDC algorithm improved the diversity of the population by employing the cloud model [13] in each generation. The characteristics of the cloud model, including randomness and stability, are included in the entire cloud droplet group, where randomness maintains the diversity of the population, while stability preserves the performance of excellent cloud droplets [14]. In other words, the proposed algorithm improves the convergence speed of the population, while at the same time, increasing the diversity and stability of the group. Computational experiments and comparisons show that ADEwDC overcomes the shortcomings of slow convergence rate and low efficiency in conventional DE algorithms, effectively avoids falling into a local optimum, and overall performs better than many state-of-the-art DE variants [4], such as JDE and JaDE, when applied to the optimization of benchmark global optimization problems.

The rest of the paper is arranged as follows. Section 2 introduces the traditional DE algorithm. In Section 3, the proposed ADEwDC algorithm is described in detail. The experimental design and results are presented and discussed in Section 4. Finally, the paper is concluded in Section 5.

2 The DE algorithm

Scientists and engineers from many disciplines often have to deal with the classic problems of search and optimization [4]. The DE algorithm is a simple population-based, stochastic parallel search evolutionary algorithm for global optimization and is capable of handling non-differentiable, nonlinear, and multimodal objective functions [9, 26]. In the DE algorithm, the population consists of real-valued vectors with dimension R D, which equals the number of design parameters. The size of the population is adjusted by parameter NP. The initial population is uniformly distributed in the search space \(\left [{\overset {\rightharpoonup }{X}}_{\min },{\overset {\rightharpoonup }{X}}_{\max }\right ]\), where \({\overset {\rightharpoonup }{X}}_{\min } = \left ({x_{\textit {min},1}},{x_{\textit {min},2}},{x_{\textit {min},3}},\ldots ,{x_{\textit {min},D}}\right )\) and \({\overset {\rightharpoonup }{X}}_{\max } = \left ({x_{\max ,1}},{x_{\max ,2}},{x_{\max ,3}},...,{x_{\max ,D}}\right )\). Each component is determined as follows:

$$ x_{i,j}^{g} = {x_{\min ,j}} + rand \cdot \left({x_{\max ,j}} - {x_{\min ,j}}\right) $$
(1)

where x i,j denotes the j th component of the i th individual, 0 denotes the initialized subsequent generation, and rand is a uniformly distributed random number between 0 and 1, which is instantiated independently for each component of the i th vector.

The traditional DE algorithm works through a simple cycle of stages, including mutation, crossover, and selection. Mutation and crossover are applied to each individual to produce the new population, followed by the selection phase, where each individual of the new population is compared with the corresponding individual of the old population, and the better of the two is selected as a member of the population in the next generation. A brief description of each of the evolutionary operators is given below.

Mutation

In the DE literature, a parent vector from the current generation is called the target vector, a mutant vector obtained through the differential mutation operation is known as the donor vector, and finally an offspring formed by combining the donor with the target vector is called a trial vector. There are many mutation strategies to generate donor vector \({\overset {\rightharpoonup }{V}}_{i}^{g}\), of which the most commonly used operator and one with the simplest form is D E/r a n d/1/b i n , which is expressed as:

$$ {\overset{\rightharpoonup}{V}}_{i}^{g} = {\overset{\rightharpoonup}{X}}_{r_{1}^{i}}^{g} + F \cdot \left({\overset{\rightharpoonup}{X}}_{r_{2}^{i}}^{g} - {\overset{\rightharpoonup}{X}}_{r_{3}^{i}}^{g}\right) $$
(2)

where \({\overset {\rightharpoonup }{X}}_{r_{1}^{i}}^{g}, {\overset {\rightharpoonup }{X}}_{r_{2}^{i}}^{g}, {\overset {\rightharpoonup }{X}}_{r_{3}^{i}}^{g}\) are sampled randomly from the current population in the g th generation. Indices r1i, r2i, and r3i are mutually exclusive integers randomly chosen from the range [1,N P], and which also differ from the index of the i th target vector, meaning r1ir2ir3ii∈{1,2,3,...,N P}. \({\overset {\rightharpoonup }{X}}_{r_{2}^{i}}^{g} - {\overset {\rightharpoonup }{X}}_{r_{3}^{i}}^{g}\) is a differential vector, and F is a real-valued mutation scaling factor that controls the amplification of the differential variation.

Crossover

After mutation, a binary crossover operation is applied to form the trial vector, \({\overset {\rightharpoonup }{U}}_{i}^{g} = \left (u_{i,1}^{g},u_{i,2}^{g},u_{i,3}^{g},\ldots ,u_{i,D}^{g}\right )\), to enhance the potential diversity of the population by exchanging the components of the donor vector \({\overset {\rightharpoonup }{V}}_{i}^{g}\) and target vector \({\overset {\rightharpoonup }{X}}_{i}^{g}\) according to the given probability, defined as CR. Each component of \({\overset {\rightharpoonup }{U}}_{i}^{g}\) is generated by the scheme outlined as:

$$ u_{i,j}^{g} = \left\{ {\begin{array}{cc} {v_{i,j}^{g}} & {if({r_{j}} \le \textit{CR})||j = {r_{i}}}\\ {x_{i,j}^{g}}&{\textit{otherwise}} \end{array}} \right. $$
(3)

where i denotes the i th individual, j denotes the j th dimension, g indicates the g th generation, r j ∈[0,1] is the j th evaluation of a uniform random number generator. CR is the crossover constant in the range [0,1], where zero means no crossover. r i ∈(1,2,3,...,D) is a randomly chosen index that ensures trial vector \({\overset {\rightharpoonup }{U}}_{i}^{g}\) gets at least one element from the donor vector \({\overset {\rightharpoonup }{V}}_{i}^{g}\), which is instantiated once per generation for each vector. Otherwise, no new parent vector would be produced and the population would remain unchanged. If the value of any dimension of the newly generated trial vector exceeds the pre-specified upper and lower bounds, it is set to the closest boundary value.

Selection

To keep the population size constant over subsequent generations, one-to-one greedy selection between a parent and its corresponding offspring is employed to decide whether the trial individual \({\overset {\rightharpoonup }{U}}_{i}^{g}\) should replace the target vector \({\overset {\rightharpoonup }{X}}_{i}^{g}\) as a member of the next generation according to their fitness values. For minimization problems, the one-to-one selection scheme is formulated as:

$$ {\overset{\rightharpoonup}{X}}_{i}^{g+1} = \left\{ {\begin{array}{cccc} {\overset{\rightharpoonup}{U}}_{i}^{g} & \textit{if} \left(f\left({\overset{\rightharpoonup}{U}}_{i}^{g}\right)\right. & \left. \le f \left({\overset{\rightharpoonup}{X}}_{i}^{g}\right)\right)\\ {\overset{\rightharpoonup}{X}}_{i}^{g} & \textit{if} \left(f\left({\overset{\rightharpoonup}{U}}_{i}^{g}\right)\right. & \left. > f \left({\overset{\rightharpoonup}{X}}_{i}^{g}\right)\right) \end{array}} \right. $$
(4)

where \(f ({\overset {\rightharpoonup }{X}})\) is the objective function to be minimized. From the above description, if and only if the trial vector yields a better cost function value compared with its corresponding target vector in the current generation, it is accepted as the new parent vector in the next generation; otherwise, the target is once again retained in the population. As a result, the population either improves or remains the same in terms of fitness status, but never deteriorates.

The iterative procedure is terminated when any one of the following criteria is met: an acceptable solution is obtained, a state with no further improvement in the solution is reached, control variables have converged to a stable state, or a predefined number of iterations have been executed. Our proposed algorithm adopts a similar main framework as the traditional DE algorithm, but employs different strategies in the process of evolution.

3 Adaptive differential evolutionary algorithm with directional strategy and cloud model

Based on the conventional DE algorithm, the Adaptive Differential Evolutionary Algorithm with Directional Strategy and Cloud Model, ADEwDC for short, constructs a trial vector generation pool to effect the dynamic adjustment strategy. The whole population is divided into three subgroups, with each subgroup selecting a different trial vector generation strategy from the pool according to its own convergence degree. In each iteration, the control parameters are dynamically set based on the fitness of the individual compared with the optimal one in the whole group, with the specific value obtained by control parameter generation. To improve the convergence rate, the evolutionary direction is introduced, and ADEwDC chooses the best individuals to evolve in the optimal evolution direction, which is defined according to the evolution potential. To avoid premature convergence, ADEwDC utilizes the cloud model to increase the diversity of the whole population, and proposes the learning operator with cloud model by applying forward and reverse cloud generators, defined as MCG. The specific operations of ADEwDC are discussed in detail in this section.

3.1 Dynamic adjustment strategy

The self-adapting strategy is an important research area for DE algorithms [36], of which there are many prominent variants. However, the performance of DE is sensitive to the choice of mutation strategy and associated control parameters [17]. In other words, different mutation strategies with different parameter settings at different stages of the evolution may be more appropriate than a single mutation strategy with unique parameter settings. Therefore, as opposed to self-adaptation [8], this paper implements dynamic adjustment by adopting a distributed topology [7] and constructing a trial vector generation pool. This means that different mutation and crossover operations, chosen from the mutation and crossover operator pool are used for each subgroup, although a uniform selection operation is adopted. Meanwhile, the control parameters for mutation and crossover obtain their values adaptively based on the fitness space, particularly from information of the marked optimal individual in the population.

Since mainly mutation and crossover in ADEwDC are used to obtain the optimal value, which exceeds the overall performance of the parent generation, rapidly, this paper constructs a mutation and crossover operator pool to generate the donor vector \({\mathrm {P}}_{\overset {\rightharpoonup }{V}^{g}} = \left ({\overset {\rightharpoonup }{V}}_{1}^{g}, {\overset {\rightharpoonup }{V}}_{2}^{g}, {\overset {\rightharpoonup }{V}}_{3}^{g}, \ldots , {\overset {\rightharpoonup }{V}}^{g}_{NP}\right )\) using the following variants:

  1. 1)

    DE/rand/1/bin

    $$ {\mathrm{v}}_{i,j}^{g} = \left\{ {\begin{array}{cc} {x_{{r_{1}},j}^{g} + {F_{1}} \cdot \left(x_{{r_{2}},j}^{g} - x_{{r_{3}},j}^{g}\right)} & {\textit{if}({r_{j}} \le \textit{CR}||{n_{j}} = j)}\\ {x_{i,j}^{g}}&{\textit{otherwise}} \end{array}} \right. $$
    (5)
  2. 2)

    DE/rand/2/bin

    $$ \begin{array}{lll} {\mathrm{v}}_{i,j}^{g} &= x_{{r_{1}},j}^{g} + {F_{1}} \cdot \left(x_{{r_{2}},j}^{g} - x_{{r_{3}},j}^{g}\right)& \\ &{\kern10pt}+ {F_{2}} \cdot \left(x_{{r_{4}},j}^{g} - x_{{r_{5}},j}^{g}\right)& {if({r_{j}} \le CR||{n_{j}} = j)}\\ &= {x_{i,j}^{g}} &{\kern43pt} {\textit{otherwise}} \end{array} $$
    (6)
  3. 3)

    DE/target-to-best/1/bin

    $$ \begin{array}{lll} {\mathrm{v}}_{i,j}^{g} &= x_{i,j}^{g} + {F_{1}} \cdot \left(x_{\textit{gbest},j}^{g} - x_{i,j}^{g}\right)\\ & \quad + {F_{2}} \cdot \left(x_{{r_{1}},j}^{g} - x_{{r_{2}},j}^{g}\right)&{\textit{if}({r_{j}} \le \textit{CR}||{n_{j}} = j)}\\ &={x_{i,j}^{g}}&{\kern43pt} {\textit{otherwise}} \end{array} $$
    (7)
  4. 4)

    DE/target-to-best/2/bin

    $$ \begin{array}{lll} {\mathrm{v}}_{i,j}^{g} &=& x_{i,j}^{g} + {F_{1}} \cdot \left(x_{\textit{gbest},j}^{g} - x_{i,j}^{g}\right) +{F_{2}} \cdot \left(x_{{r_{1}},j}^{g} - x_{{r_{2}},j}^{g}\right) \\ && \quad + {F_{3}} \cdot \left(x_{{r_{3}},j}^{g} - x_{{r_{4}},j}^{g}\right)~~~~~~~~~~{\kern3pt}{if({r_{j}} \le \textit{CR}||{n_{j}} = j)} \\ &=&{x_{i,j}^{g}}\qquad\qquad\qquad\qquad~~~~~~~~~~~~~{\kern32pt} {\textit{otherwise}} \end{array} $$
    (8)
  5. 5)

    DE with a neighborhood-based scheme

    1. 5.1)

      Neighborhood vector

      $$ {\overset{\rightharpoonup}{L}}_{i}^{g} = {\overset{\rightharpoonup}{X}}_{i}^{g} + \alpha_{1}\cdot \left({\overset{\rightharpoonup}{X}}_{\textit{nbest}} - {\overset{\rightharpoonup}{X}}_{i}^{g}\right) + \beta_{1} \cdot \left({\overset{\rightharpoonup}{X}}_{r1}^{g} - {\overset{\rightharpoonup}{X}}_{r2}^{g}\right) $$
      (9)
    2. 5.2)

      Population vector

      $$ {\overset{\rightharpoonup}{G}}_{i}^{g} = {\overset{\rightharpoonup}{X}}_{i}^{g} + \alpha_{2}\cdot \left({\overset{\rightharpoonup}{X}}_{\textit{gbest}} - {\overset{\rightharpoonup}{X}}_{i}^{g}\right) + \beta_{2} \cdot \left({\overset{\rightharpoonup}{X}}_{r1}^{g} - {\overset{\rightharpoonup}{X}}_{r2}^{g}\right) $$
      (10)
    3. 5.3)

      Component of donor vector

      $$ {\mathrm{v}}_{i,j}^{g} = \left\{ {\begin{array}{ll} {\omega \cdot g_{i,j}^{g} + (1 - \omega ) \cdot l_{i,j}^{g}}&{if({r_{j}} \le \textit{CR}||{n_{j}} = j)}\\ {\kern40pt}{x_{i,j}^{g}}&{\kern30pt} {\textit{otherwise}} \end{array}} \right. $$
      (11)

where r 1r 2r 3r 4r 5i∈{1,2,3,...,N P}, j denotes the j th component, F 1,F 2,F 2,α 1,α 2,β 1,β 2 are all scaling factors to control the scale of the differential vectors, ω is a weighting factor applied to the information of the neighborhood and global population, and C R is the crossover rate to determine the source of each dimension of the offspring.

Of the five variants given above, since DE/target-to-best/1/bin and DE/target-to-best/2/bin rely on the best solution found, they usually have a faster convergence speed and perform well when solving unimodal problems. However, these algorithms are more likely to be trapped in a local optimum and to converge prematurely when solving multimodal problems. DE/rand/1/bin and DE/rand/2/bin usually demonstrate slow convergence speed with superior exploration capability. Therefore, they are usually better suited to solving multimodal problems. By including a neighborhood-based scheme [3, 5], different variants are employed to evolve individuals and set parameters, as discussed in the next section. Obviously, these five donor vector strategies constitute a strategy candidate pool with diverse characteristics.

In this paper, the distributed topology is simplified by dividing the population into three subgroups, while the selection criteria are strengthened by defining a variable, called the convergence degree.

Definition 1

Convergence degree is the average value of the accumulative fitness of each individual in the subgroup as the standard for determining the specific mutation and crossover operation. It is defined as:

$$ \textit{fi}{t^{s}} = \textit{Fit}_{\Sigma}^{s}/\textit{np}; $$
(12)

where the accumulative fitness \(\textit {Fit}_{\Sigma }^{s} = \sum \limits _{i = 1}^{np} {f\left ({\overset {\rightharpoonup }{X}}_{i}^{g}\right )}\), s denotes the number of subgroups, and n p denotes the number of individuals in each subgroup. Considering the stability of the algorithm, the subgroup with the maximum convergence degree generates the donor vector by DE/rand/1/bin or DE/rand/2/bin, the one with the minimum degree uses DE/target-to-best/1/bin or DE/target-to-best/1/bin, and the remaining ones choose their variant randomly.

ADEwDC makes full use of the historical optimal information of the population and the fitness of each individual to set the control parameters, referred to as PSO. The ultimate evolution goal of the current individual can be regarded as the current group optimal individual with the optimal fitness, and this forms the important theoretical basis for setting the control parameters.

figure f
figure g

With reference to the relevant papers, the scaling factors F,α,β take values in [0.1,1] [19], the weighting factor ω is restricted to the range [0.05,0.95] [5], and C R takes a value in [0.1,0.9] [17]. Thus, integrating the above discussion, the dynamic adjustment strategy is given as follows.

At this point, ADEwDC generates the same number of individuals as in the parent generation, that is, the donor vector space \({\mathrm {P}}_{\overset {\rightharpoonup }{V}^{g}} = \left ({\overset {\rightharpoonup }{V}}_{1}^{g}, {\overset {\rightharpoonup }{V}}_{2}^{g}, {\overset {\rightharpoonup }{V}}_{3}^{g}, \ldots , {\overset {\rightharpoonup }{V}}^{g}_{\textit {NP}}\right )\).

3.2 Design of evolutionary direction

DE belongs to the class of stochastic optimization algorithms with randomness and unpredictability, leading to the arbitrariness of the offspring and the fact that good characteristics of parental individuals cannot be fully passed to the next generation. However, living things are not passive victims of their environment in nature and human society. Instead, they struggle to fit into the environment by constantly adjusting the evolutionary direction. Therefore, this paper adopts an indicative directional propagation strategy to alter the blindness of conventional DE and improve the convergence rate of the DE algorithm before the selection process.

Definition 2

Evolutionary direction is a description of the direction vector from the current individual to its offspring for any individual \({\overset {\rightharpoonup }{X}}_{i}^{g}\) in the population P g , such that ∀Xi gP g . It is defined as:

$$ {\text{DR}}\left({\overset{\rightharpoonup}{X}}_{i}^{g}\right) = \left\{ {\begin{array}{ll} \overrightarrow{\overset{\rightharpoonup}{X}_{i}^{(g-1)^{\prime}}{\overset{\rightharpoonup}{X}}_{i}^{g}}& \textit{while} (g > 1)\\ {\textit{rand}\_{\textit{direction}}}&{\kern20pt} {\textit{else}} \end{array}} \right. $$
(13)

where \({\overset {\rightharpoonup }{X}}_{i}^{g} = \textit {offspring} \left ({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}}\right )\), \({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}} \in P_{(g -1)}\), and \({\overset {\rightharpoonup }{X}_{i}^{g}} \in P_{g}\).

In the current population, the evolutionary direction of individual \({\overset {\rightharpoonup }{X}_{i}^{g}}\) depends on itself and its corresponding parent in the previous generation of population \({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}}\), shown in the previous equation as \(\overrightarrow {\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}{\overset {\rightharpoonup }{X}}_{i}^{g}}\). The evolutionary directions of individuals in the first generation are selected randomly, and the optimal evolution direction is described below.

Definition 3

Optimal evolution direction D R o p t (P g ) is selected from the evolutionary orientation of the whole population through a series of specific operations, which is the best evolution direction. It is described as:

$$ D{R_{\textit{opt}}}({P_{g}}) = {\varTheta_{s}}\left({\text{DR}}\left({\overset{\rightharpoonup}{X}_{i}^{g}}\right)\right) $$
(14)

where \({\overset {\rightharpoonup }{X}_{i}^{g}} \in P_{g}\) and \({f\left ({\overset {\rightharpoonup }{X}}_{i}^{g}\right )}\geq f{\left ({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}}\right )}\), and Θ s denotes the selection operation on the evolution direction.

ADEwDC defines the directional strategy as follows: Select the top n individuals with the maximum fitness in the donor vector space as generated in the previous subsection. Select the top m optimal evolution directions. Then, evolve the selected individuals along the selected directions and shape the progeny space with nm individuals, where nm>N P.

There are many criteria for selecting the optimal evolution direction, and the overall requirement is to improve the fitness of an individual after evolving along the selected orientation. In fact, the ultimate evolution goal of the current individual can be regarded as the current group optimal individual \({\overset {\rightharpoonup }{X}}_{gbest}\), which has the optimal fitness of the whole population, fming for a minimization problem. Thus, the goal vector is defined as \(\overrightarrow {\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}{\overset {\rightharpoonup }{X}}_{gbest}^{g}}\), and the motion vector as \(\overrightarrow {\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}{\overset {\rightharpoonup }{X}}_{i}^{g}}\). Then, ADEwDC defines evolution potential as the factor used to choose the optimal directions.

Definition 4

Evolution potential D R is defined as the disparity between the goal vector and the motion vector in the fitness space, defined as:

$$ \nabla_{{\text{DR}}({\overset{\rightharpoonup}{X}}_{i}^{g})} = \left[f({\overset{\rightharpoonup}{X}}_{i}^{g}) - f\left({\overset{\rightharpoonup}{X}_{i}^{(g-1)^{\prime}}}\right)\right]\bigg/\left[f\left({\overset{\rightharpoonup}{X}}_{\textit{gbest}}^{g}\right) - f\left({\overset{\rightharpoonup}{X}_{i}^{(g-1)^{\prime}}}\right)\right] $$
(15)

Therefore, the optimal evolution direction based on maximum evolution potential is defined as:

$$ D{R_{\textit{opt}\_ep}}({P_{g}}) = {\text{DR}}\left({\overset{\rightharpoonup}{X}}_{i}^{g}\right) {\kern 4pt} \textit{where} {\kern 4pt} j = \arg {\underset{j}{\mathrm{\max }}} {\nabla_{{\text{DR}}({\overset{\rightharpoonup}{X}}_{i}^{g})}} $$
(16)

The directional evolution strategy is given by the following algorithm.

figure h

From the optimal individuals and optimal evolution directions, the trail vector space \({\mathrm {P}}_{\overset {\rightharpoonup }{U}^{g}} = \left ({\overset {\rightharpoonup }{U}}_{1}^{g}, {\overset {\rightharpoonup }{U}}_{2}^{g}, {\overset {\rightharpoonup }{U}}_{3}^{g}, \ldots , {\overset {\rightharpoonup }{U}}^{g}_{n{\cdot }m}\right )\) is constructed, providing choices for the new generation.

3.3 Specific application of cloud model

By introducing the evolutionary direction, the convergence speed of the DE algorithm is improved further; however, the possibility of the algorithm falling into a local optimum rises markedly because of the rapid decrease in population diversity. To improve the diversity of the population, we introduce the cloud model [13], which describes individuals of the population through expectation Ex, entropy En, and hyper-entropy He.

Definition 5

Membership cloud [11] Let U denote a quantitative domain composed of precise numerical variables, with C the qualitative concept on U. If the quantitative value xU is a random realization of qualitative concept C, the confirmation of x on C can be denoted as μ(x)∈[0,1], which is a random number with stable tendency.

$$ \begin{array}{c} \mu :U \to \left[ {0,1} \right] \\ \forall x \in U,x \to \mu (x) \end{array} $$
(17)

The distribution of x on U is called a cloud, x is called a cloud droplet, and the cloud consists of a series of cloud droplets.

The cloud droplets have a certain randomness, which maintains the diversity of individual stocks thereby avoiding a search for local extreme values, and stability characteristics, which protect the population of a more outstanding individual and thus the overall situation of extreme adaptive positioning. The membership cloud describes a concrete concept through expectation Ex, entropy En, and hyper-entropy He. Expectation Ex expresses the point that is most able to represent the domain of the concept and is the most typical sample of this concept to quantify. Entropy En represents the granularity of a concept that can be measured (the larger the entropy is, and the larger the granularity is, the more macro is the concept). It reflects the range of the domain space that can be accepted by the specific concept. Hyper-entropy He describes the uncertain measurement of entropy. It can be used to express the relationship between randomness and fuzziness.

As a specific kind of cloud, the normal cloud model has been proven to be universal [12], based on the normal distribution and Gauss membership function. Regarding probability, the normal distribution is the most commonly used form, which is described by expectation E and variance D. In fuzzy set theory, the bell-shape membership function \(\mu (x) = e^{\frac {{ - (x - a)^{2} }}{{2b^{2} }}}\) is also the most common membership function used in fuzzy sets. The normal cloud, described below, combines the characteristics of the two with an additional expansion.

Definition 6

Normal Cloud Model [14] Let U be the universe of discourse and à be a qualitative concept in U. If xU is a random instantiation of concept Ã, satisfying xN(E x ,E n 2) and E n 2N(E n ,H e 2), and the certainty degree of x belonging to concept à satisfies

$$ \mu = {e^{- \frac{{{{(x - {E_{x}})}^{2}}}}{{2{{({E_{n}}^{\prime} )}^{2}}}}}} $$
(18)

then the distribution of x in universe U is called a normal cloud.

Knowledge is usually the association between concepts in the real world, between which the cause and effect relationship can be described by the membership cloud generator (MCG), including both a forward and reverse generator (for more information, see [11]). With the associated numerical characteristics, that is, Ex, En and He, the forward cloud generator can generate cloud drops (x,μ), where x is the quantity values and μ is the membership degree of x [10]. The reverse cloud generator is the other conversion model that can convert quantity numbers to a quality concept. It can convert accurate data (x 1,x 2,...,x n ) with membership degrees (μ 1,μ 2,...,μ n ) to a quality cloud concept expressed as numerical characteristics (E x ,E n ,H e ) [10]. The MCG based on cloud theory is summarized below.

figure i

Based on our analysis, ADEwDC incorporates the cloud model into the selection operation to remedy the diversity of population, and presents a novel operator called the learning operator with cloud model.

Definition 7

Learning operator with cloud model View the current population I λ as a cloud defined using C l o u d M (E x ,E n ,H e ), the eigenvalue calculated by the reverse normal cloud generator r C G (P λ ) to describe the overall information owned by it. For each individual in the suboptimal population, ∀d i I λe, generate a new individual using the forward normal cloud generator f C G (E x ,E n ,H e ,(λe)), then the learning operator with cloud model is defined as:

$$\begin{array}{@{}rcl@{}} \forall {d_{i}} \in {I^{\lambda - e}},\exists {d_{c}} \in \varUpsilon_{\textit{Cloud}}^{\lambda - e}:{d_{i}} \to {d_{c}},\\ \;\varUpsilon_{\textit{Cloud}}^{\lambda - e} = fCG({E_{x}},{E_{n}},{H_{e}},(\lambda - e)) \end{array} $$
(19)

where λ denotes the parents’ individuals space and e means the individuals selected from trail vector space. That is, part of the subgeneration is selected directly from the trial vector space, with count e, and the rest is generated randomly by the learning operator with cloud model according to the feature information about the known population.

The specific application of the cloud model is summarized below.

figure j

Thus far, we have introduced each part of ADEwDC: the dynamic adjustment strategy is used in the mutation and crossover process, while the design of the evolutionary direction and specific application of the cloud model are both applied in the selection process. The overall framework of ADEwDC is presented next.

3.4 ADEwDC

The proposed ADEwDC employs the dynamic adjustment strategy (see Section 3.1), the directional strategy (see Section 3.2), and the cloud model (see Section 3.3), introduced previously. In this section, we first present the overall framework of the proposed algorithm, and then analyze its discipline holistically.

3.4.1 Framework of the algorithm

3.4.2 Discipline of the algorithm

The main framework of ADEwDC is taken from the traditional DE algorithm, but with changed mutation and crossover strategies and an improved selection mechanism. As with the classical DE algorithm, the proposed algorithm defines a donor vector and a trial vector, where the former is the result of the mutation and crossover operation, while the latter is the result of directional evolution. The next generation may come from the parent generation, donor vector space (Algorithm 2), or trial vector space (Algorithm 3), but most new individuals must be generated by the cloud model (Algorithm 5).

  1. a)

    Dynamic adjustment strategy ADEwDC constructs a mutation and crossover pool to realize the dynamic adjustment strategy, including DE/rand/1/bin, DE/rand/ 2/bin, DE/target-to-best/1/bin, DE/target-to-best/2/bin, and DE with neighborhood-based scheme variants. It then defines the strategies to select the evolution operators (Definition 1) and set the control parameters (Algorithm 1) based on the constructed topology, which are summarized as the donor vector generation (Algorithm 2).

  2. b)

    Design of evolutionary direction The evolutionary direction is determined based on the theory that the individuals are not passive in joining the environment, but struggle to fit into the environment by constantly adjusting the direction of evolution. This algorithm defines the optimal direction (Definition 3) with evolution potential (Definition 4), and synthesizes the trial vector generation(Algorithm 3).

  3. c)

    Specific application of cloud model The cloud model is a useful mathematical tool with randomness and a stability tendency, which is employed to record the eigenvalues of the parent generation using the reverse cloud generator and generate cloud droplets randomly for most of the filial generation using the forward cloud generator (Definition 4). The specific operation is described by Algorithm 5.

  4. d)

    Selection strategy The next generation consists of two parts: some are directly selected as offspring from the vector space, consisting of the parent population, the donor vector space, and the trial vector space, whereas the greater part is generated by the MCG (Algorithm 4) based on the cloud model to compensate for the loss of population diversity, and which is described as:

    $$ {P_{g + 1}} = {\varTheta_{s}^{e}}\left({P_{g}} \cup {P_{\overset{\rightharpoonup}{{V}}^{g}}} \cup {P_{\overset{\rightharpoonup}{{U}}^{g}}}\right) \cup \varUpsilon_{Cloud}^{\lambda - e} $$
    (20)

    where Θ s e means selecting e individuals from the specific vector space.

figure k

ADEwDC utilizes the dynamic adjustment strategy to select evolutionary variants and set control parameters based on the constructed topology. In fact, the subgroup with the maximum f i t s evolves with a slow convergence speed and superior exploration capability, while the minimum one evolves with a fast convergence speed and superior exploitation capability. Furthermore, each individual in the same subgroup sets its control parameters based on the gap between its fitness and the best fitness of the population. In short, this algorithm has good adaptability.

Furthermore, ADEwDC designs an evolutionary direction to improve the convergence rate, introduces the cloud model to ensure good diversity of the population, and applies the special selection strategy to balance the performance of the whole group. A series of experiments were carried out to confirm the effectiveness of ADEwDC as reported in the next section.

4 Experimental setup and results

To validate the ADEwDC algorithm, we selected several categories of global minimization benchmark functions to evaluate the proposed algorithm against other DE variants. These benchmark functions provide a balance between unimodal and multimodal functions, and were chosen from the set of 13 classic benchmark problems [30] that have frequently been used in the literature [25, 38]. The following problems were used:

  1. (1)

    Rastrigin function:

    $${f_{1}}(X) = 10 \cdot D + \sum\limits_{i = 1}^{D} {\left[{x_{i}^{2}} - 10 \cdot \textit{cos}(2\varPi {x_{i}})\right]} $$

    with global optimum X = 0 and f(X ) = 0 for −5 ≤ x i ≤ 5.

  2. (2)

    Sphere function:

    $${f_{2}}(X) = \sum\limits_{i = 1}^{D} {{x_{i}^{2}}} $$

    with global optimum X =0 and f(X )=0 for −100≤x i ≤100.

  3. (3)

    Rosenbrock function:

    $${f_{3}}(X) = \sum\limits_{i = 1}^{D - 1} {\left(100{{({x_{i}^{2}} - {x_{i + 1}})}^{2}} + {{({x_{i}} - 1)}^{2}}\right)} $$

    with global optimum X =(1,1,...,1) and f(X )=0 for −100≤x i ≤100.

  4. (4)

    Ackley function:

    $${f_{4}}(X) = - 20 \cdot {e^{- 0.2\sqrt {\frac{1}{D}\sum\limits_{i = 1}^{D} {{x_{i}^{2}}} } }} - {e^{\frac{1}{{\mathrm{D}}}\sum\limits_{{\mathrm{i}} = 1}^{D} {\cos (2{\Pi} {x_{i}})} }} + 20 + e $$

    with global optimum X =0 and f(X )=0 for −32≤x i ≤32.

  5. (5)

    Griewank function:

    $${f_{5}}(X) = \sum\limits_{i = 1}^{D} {\frac{{{x_{i}^{2}}}}{{4000}} - \prod\limits_{i = 1}^{D} {\cos \left(\frac{{{x_{i}}}}{{\sqrt i }}\right) + 1} } $$

    with global optimum X =0 and f(X )=0 for −600≤x i ≤600.

Experiment 1 used a two-dimensional (2D) f 1 to explore the impact of the relevant parameters of Algorithm 6, while experiment 2 used the same test function to show the distribution of individuals in the solution space of each generation and verify its convergence. Experiment 3 used a 2D f 2, f 3, f 4, f 5 to explain the effectiveness of ADEwDC, compared with the conventional DE algorithm, DE/rand/1/bin. Next, ADEwDCf was evaluated on the CEC2013 benchmark problem set [15], as well as the CEC2005 benchmarks [23] in Experiment 4. In this paper, we compare ADEwDC with the following state-of-the-art DE algorithms: SHADE [24], CoDE [28], EPSDE [17], JADE [34], and dynNP-jDE [1] (an improved version of jDE [2]). Our comparisons were done with dimensions D=30 to analyze the effectiveness of the ADEwDC algorithm comprehensively. Finally, ADEwDC was used to solve the CEC2013 benchmark problems with dimensions D = 50 to prove its universality and robustness.

All experiments were executed on the following system:

  • OS: Windows 7 Professional

  • CPU: Intel(R) Xeon(R) CPU E5620 @ 2.40GHz 2.40GHz

  • RAM: 12.0GB

  • Language: Matlab

  • Compiler: Microsoft Visual C ++ 2012

4.1 Experiment 1 - related parameter settings

To investigate the relevant parameter settings, we utilized a 2D f 1 as the test function to explore the influence of parameters in Algorithm 6, with the function described as:

$$f(x_{1} ,x_{2} ) = 20 + {x_{1}^{2}} + {x_{2}^{2}} - 10(\cos 2\pi x_{1} + \cos 2\pi x_{2} ). $$

There are many local minima in the region of the value distribution of this function, and thus it is a good test case for measuring the effect of the relevant parameter settings. In this experiment, the number of top individuals (n) and the number of top directions (m) in step 2.3 of Algorithm 6 were observed, as well as the number of remaining individuals (e) from step 2.4 of Algorithm 6.

First, keeping the same e, we set n>m, n=m, n<m<2n, m=2n and m>2n as five cases. This experiment was repeated 10,000 times, and we randomly selected the set of experimental results shown in Fig. 1a. Then we calculate the average best solution of all the experimental results, which are shown in Table 1.

Fig. 1
figure 1

Influence of the number of top individuals (n), the number of top directions (m), and the number of remaining individuals (e) when optimizing the Rastrigin function with ADEwDC

Table 1 Average best solution of all the experimental results varying n and m with a fixed e when optimizing the Rastrigin function with ADEwDC over 10000 trials

Then, keeping the same n and m, set e=20 % N P,e=40 % N P,e=50 % N P,e=60 % N P,e=65 % N P,e=70 % N P,e=75% N P and e=80 % N P as eight cases. This experiment was also repeated 10,000 times, and we randomly selected the set of experimental results shown in Fig. 1b, and then calculate the average best solution of all the experimental results in Table 2.

Table 2 Average best solution of all the experimental results varying e with fixed values of n and m when optimizing the Rastrigin function with ADEwDC over 10000 trials

Based on the above analysis, ADEwDC achieves relatively good results when the number of top directions (m) is more than twice the number of top individuals (n), and the number of remaining individuals (e) is 6 0 % of the population size. In fact, n satisfies:

$$ \left\{ {\begin{array}{ll} {m > 2n}\\ {n \cdot m > NP} \end{array}\;\; \Rightarrow n \cdot m > 2{n^{2}} > NP \Rightarrow n > \sqrt {\frac{1}{2} \cdot NP} } \right. $$
(21)

4.2 Experiment 2 - convergence analysis

As defined above, the 2D Rastrigin function was also used in this experiment, although it was applied to verify the convergence of Algorithm 6.

The initial population consisted of 40 points, 20 times the dimension recommended by Price [20], distribution randomly in the region between [−5,5] and [−5,5]. Meanwhile, n=5,m=12, ensure nm>N P and m>2n, then e=24, which is 60 % of the population size. The distribution of the initial population is shown in Fig. 1a. The fitness function is the function value of the concrete points in the area. From the test results shown in Figs. 2a–f, we can see that after the evolution of 12 generations, the individuals distribution converges well. We listed the individuals of the 2nd, 4th, 8th, 11th, 12th generations on the region as shown in Figs. 2b–f, respectively.

Fig. 2
figure 2

Distribution of individuals in each generation while computing the minimum of the Rastrigin function

To further verify the convergence of ADEwDC, it was compared with the standard DE, sDE for short, with DE/rand/1/bin (with F=0.5,C R = 0.3 in our experiments). The results of evolving 15 generations shown in Fig. 3 were randomly selected from 50 trial runs. In the search procedure for the optimal solution, we used the distribution area of all the points to represent the search space, changes in which are illustrated in Fig. 3a. Obviously, the search space of ADEwDC tends to zero much faster than that of sDE, after evolving only five or six generations. We plot the best individual for each evolutionary generation, as well as its abscissa and ordinates in Figs. 3b–d, respectively. In Fig. 3, it can be seen that ADEwDC converges to the optimal solution much faster than sDE.

Fig. 3
figure 3

Changes in the search space and the optimal value in each generation, comparing ADEwDC with the standard DE

Overall, the results of experiment 2 show that ADEwDC can effectively reduce the search space and rapidly converge to the optimal solution, which is consistent with the directional strategy. In the course of evolution, excellent individuals evolve along the optimal evolution directions, enabling the entire population to obtain the optimal value easily and quickly.

4.3 Experiment 3 - effectiveness analysis

In this experiment, ADEwDC was used to optimize the 2D f 2, f 3, f 4,f 5 to investigate the effectiveness of the algorithm compared with the standard DE algorithm (sDE). The results of evolving 50 generations for each function, shown in Fig. 4, were randomly selected from 100 trial runs.

Fig. 4
figure 4

Comparison of the standard DE and the proposed algorithms while computing the minimum of f 2, f 3, f 4, f 5

For the Ackley and Griewank functions, the results by ADEwDC are better than those by sDE in terms of both convergence speed and accuracy, as clearly shown in Figs. 4c and (d). For the Sphere and Rosenbrock functions, as shown in Figs. 4a and b, ADEwDC converges faster than sDE and obtains the desired optimal solution. To illustrate the effectiveness of ADEwDC further, in Table 3 we list one full trial result, randomly selected from the 100 experimental results for each function.

Table 3 Comparison of the standard DE algorithm and ADEwDC when computing the minimum of f 2, f 3, f 4, f 5

From all the test results, we find that ADEwDC converges to the optimal solution quickly and has good robustness. In contrast, sDE is more likely to fall into a local optimum as shown in Fig. 5. Moreover, 1.0×10−3 is defined as the acceptable level, which means that the run is judged to be successful if a solution obtained by an algorithm falls between the acceptable level and the actual global optimum [33]. And we compare the frequency of premature convergence and the average value of all the test results between ADEwDC and sDE in Table 4.

Fig. 5
figure 5

Optimal individuals of f 2, f 3, f 4, f 5 in each generation computed by ADEwDC and sDE, showing that ADEwDC has better robustness than sDE

Table 4 The frequency of premature convergence and the average best solution when computing the minimum of f 2, f 3, f 4, f 5 with ADEwDC and sDE

Generally, the payoff for increasing diversity is a slower (although more efficient) convergence speed. Nevertheless, Table 4 indicates that ADEwDC is not at the cost of the convergence speed to obtain the diversity of the solutions; conversely, it improves the diversity with MCG (Algorithm 4), maintaining a good convergence rate at the same time which has been confirmed in the previous content [6, 35].

From the above experimental results and analysis, it can been seen that compared with the standard DE algorithm, Algorithm 6 has a faster convergence rate, obtains more accurate optimization results, and has better robustness. In short, ADEwDC is more effective than the standard DE because of the cloud model, which enhances the population diversity and prevents the algorithm from falling into a local optimum.

4.4 Experiment 4 - competitiveness analysis

In this section, we evaluate the performance of ADEwDC on the CEC2013 benchmark problem set [15], compared with SHADE [24], CoDE [28], EPSDE [17], JADE [34] and dynNP-jDE [1]. Then, using the CEC2005 benchmarks [23], ADEwDC is compared with SHADE to verify its validity and accuracy further. For each comparative algorithm, we used the control parameter values suggested in the cited papers. ADEwDC was executed on the same system as that given for the previous experiments, while comparative data for SHADE were taken from its original paper [24]. The source programs for CoDE, EPSDE, and JADE were based on code received from the original authors. The number of dimensions was set to D=30 and the maximum number of objective function calls per run was calculated as D×10,000 (that is, 300,000) when comparing ADEwDC with the other algorithms. Meanwhile, N P=100,n=8,m=20, and e=60. Finally, we executed ADEwDC on the CEC2013 benchmarks with the number of dimensions set to D=50 as a high-dimensional case to prove its universality and robustness.

4.4.1 Experiment on the CEC2013 benchmarks

In this experiment, we performed our evaluation following the guidelines of the CEC2013 benchmark competition [15]. The search space was set as [−100,100]D for all the selected problems, the results of which are shown in Table 5. In the table, the mean and standard deviation of the error (difference) between the best fitness values found in each run and the optimal value are shown. The + , −, ≈ indicate whether a given algorithm performed significantly better (+ ), significantly worse (−), or not significantly different better or worse (≈) compared to ADEwDC according to the Wilcoxon rank-sum test, which is a nonparametric alternative to the two-sample t-test based solely on the order of the observations from the two samples (significance threshold p≤0.05) [29]. Functions f 1f 5 are unimodal, while f 6f 20 are multimodal. f 21f 28 are composite functions combining multiple test problems into a complex landscape [15].

The best result for each benchmark function is highlighted in Table 5, while Table 6 shows the statistical ranking according to the respective performance of the DE algorithms. The algorithms are ranked based on the average best solution in each row of Table 5, and the final rank of the algorithms is determined by the average rank calculated according to problem feature in each column. From Table 6, we see that SHADE, ADEwDC and JADE achieve the best performance on unimodal problems. The good performance of JADE on the unimodal functions is consistent with previous results [28]. For the basic multimodal functions, ADEwDC performs relatively well, although for several of the problems including f 13,f 18, JADE and CoDE perform particularly well. For the complex, composite functions, the best performer is dynNP-jDE (possibly owing to its population size reduction strategy), followed by SHADE, ADEwDC, JADE, CoDE and EPSDE. Finally, based on the statistical data given in the bottom three rows of Table 5, ADEwDC achieves better performance than CoDE, EPSDE and JADE, and similar performance to dynNP-jDE, but does not surpass the performance of SHADE on these 28 problems. Still, ADEwDC is better suited to multimodal problems than SHADE. This is probably because the dynamic adjustment strategy gives ADEwDC extensive adaptability, which has advantages in dealing with multimodal problems.

Table 5 Comparison of ADEwDC and state-of-the-art DE algorithms while computing the minimum of the benchmark functions in CEC2013. For all problems, dimensionality was set to D=30, and the maximum number of objective function evaluations to D×10,000=300,000. All results are averaged over 51 trials
Table 6 Statistical ranking of ADEwDC and state-of-the-art DE algorithms while computing the minimum of the benchmark functions in CEC2013

4.4.2 Experiment on the CEC2005 benchmarks

To verify the validity and accuracy further, we compared ADEwDC with SHADE on the CEC2005 benchmarks [23]. CEC2005 provides many classic benchmark functions, which have been tested by many original authors of algorithms. Functions f 1f 5 are unimodal, while the others are multimodal. f 6f 12 are basic functions, f 13f 14 are expanded functions, and f 15f 25 are hybrid composition functions. For this experiment, comparative data were once again taken from [24].

In Table 7, it is shown that ADEwDC performs very well with multimodal problems, especially on the basic and hybrid composition functions, which coincides with the previous experimental analysis. Overall, ADEwDC is slightly better than SHADE. For the expanded functions, ADEwDC achieves similar performance to SHADE. Nonetheless, ADEwDC does not perform as well on the unimodal problems, especially for f 4,f 5, which is also consistent with the previous analysis. Generally, ADEwDC is ideal for multimodal optimization problems.

Table 7 Comparison of ADEwDC and SHADE while computing the minimum of benchmark functions in CEC2005. For all problems, we set the dimensionality D=30, and the maximum number of objective function evaluations D×10,000=300,000. All results are averaged over 25 trials

4.4.3 High dimensional experiment on the CEC2013 benchmarks

We also executed ADEwDC on the CEC2013 benchmarks, setting the number of dimensions D=50 as a high-dimensional case to prove the universality and robustness of the proposed algorithm. The maximum number of objective function calls per run was set as D×10,000 (that is, 500,000). From Table 8, it can clearly be seen that ADEwDC is also suitable for high-dimensional problems, especially for multimodal optimization functions, which is consistent with our previous experimental analysis and conclusions.

Table 8 ADEwDC applied to the CEC2013 benchmarks, with dimensionality D=50, and maximum number of objective function evaluations D×10,000=500,000. The results of this high-dimensional test confirm the universality and robustness of ADEwDC, with all results averaged over 51 trials

5 Conclusions and future work

A novel adaptive DE algorithm with directional strategy and cloud model was presented in this paper. The DE algorithm is known to be an efficient and powerful optimization algorithm, which is used widely in scientific and engineering fields. Considering that most DE algorithms still suffer from a number of problems such as difficult parameter setting, slow convergence rate, and premature convergence, the proposed algorithm utilizes a dynamic adjustment strategy to select evolutionary variants and set control parameters based on the constructed topology, adopts a directional strategy to evolve outstanding individuals and improve the convergence speed of the algorithm, and maintains the diversity of the population by employing the cloud model. Results of the experiments confirm that ADEwDC achieves good performance in terms of convergence, stability, and precision, which greatly contributes to overcoming the low efficiency in conventional DE algorithms. Meanwhile, ADEwDC effectively avoids falling into a local optimum. Our future work involves strengthening the theoretical analysis of ADEwDC, further improving its accuracy and stability, and generalizing it to solve constraint and multi-objective optimization problems in practical applications.