Abstract
Recently, many studies have focused on differential evolution (DE), which is arguably one of the most powerful stochastic real-parameter optimization algorithms. Prominent variants of this approach largely optimize the DE algorithm; however, almost all the DE algorithms still suffer from problems like difficult parameter setting, slow convergence rate, and premature convergence. This paper presents a novel adaptive DE algorithm by constructing a trial vector generation pool, and dynamically setting control parameters according to current fitness information. The proposed algorithm adopts a distributed topology, which means the whole population is divided into three subgroups with different mutation and crossover operations used for each subgroup. However, a uniform selection strategy is employed. To improve convergence speed, a directional strategy is introduced based on the greedy strategy, which means that an individual with good performance can evolve rapidly in the optimal evolution direction. It is well known that the faster an algorithm converges, the greater the probability of premature convergence. Aimed at solving the local optimum problem, the proposed algorithm introduces a new mathematical tool in the selection process, called the membership cloud model. In essence, the cloud model improves the diversity of the population by randomly generating cloud droplets. Experimental results from executing typical benchmark functions show high quality performance of the proposed algorithm in terms of convergence, stability, and precision. They also indicate that this improved differential evolutionary algorithm can overcome the shortcoming of conventional differential evolutionary algorithms of low efficiency, while effectively avoiding falling into a local optimum.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Since its introduction in 1995 by Storn and Price [22], the differential evolution (DE) algorithm has been used as a simple yet powerful search technique for solving complex nonlinear and non-differentiable continuous functions. It belongs to the class of stochastic optimization algorithms, which are used to find the best-suited solution to a problem by minimizing an objective function, which is a mapping from a parameter vector \({\overset {\rightharpoonup }{X}}\in {\Re ^{D}}\) to R, within the given constraints and flexibilities.
Starting with a population initialized randomly, the DE algorithm uses simple mutation and crossover operators to generate new candidate solutions, and adopts a one-to-one competitive scheme to determine whether the offspring should replace their parents in the next generation [19]. Owing to its ease of implementation and simplicity, DE has attracted much attention from researchers all over the world, resulting in many variations of the basic algorithm with improved performance. The strategies of the variants often represent varying search capabilities in different search phases of the evolution process.
Most of the approaches to improve the standard DE algorithm mainly concentrate on four aspects: the structure and size of the population, associated control parameter setting, trial vector generation, and hybrid strategies. Of these, control parameter setting and trial vector generation directly affect the search accuracy and convergence speed of the DE algorithm, which is why associated control parameter setting is generally considered together with the trial vector generation strategy. Qin et al. [21] developed a self-adaptive DE (SaDE) algorithm for constrained real-parameter optimization, in which both the trial vector generation strategy and the associated control parameter values are gradually self-adapted according to the learning experience. This algorithm performs much better than both the traditional DE algorithm and several state-of-the-art adaptive parameter DE variants, such as the ADE [32], SDE [18], and JDE [2] algorithms. In [19], a self-adaptive DE algorithm, called SspDE, is proposed with each target individual having its own trial vector generation strategy, scaling factor F, and crossover rate CR, which gradually self-adapt from previous experience in generating promising solutions. Mallipeddi et al. [17] employed an ensemble of mutation strategies and control parameters in their DE algorithm (called EPSDE), in which a pool of distinct mutation strategies coexists with a pool of values for each control parameter throughout the evolution process competing to produce offspring. In [7], the authors designed a heterogeneous distributed algorithm (HdDe) by proposing a new mutation strategy, GPBX- α, and parallel execution in two separate islands using the classic DE/rand/1/bin algorithm. Wang and Zhao [27] presented a DE algorithm with a self-adaptive population resizing mechanism based on JADE [34], called SapsDE. This algorithm gradually self-adapts NP according to previous experience in generating promising solutions and enhances the performance of DE by dynamically choosing one of two mutation strategies and tuning control parameters in a self-adaptive manner. Many hybrid strategies exist that improve the efficiency of the DE algorithms. For the unconstrained global optimization problems, a novel optimization model is proposed, called clustering-based differential evolution with 2 multi-parent crossovers (2-MPCs-CDE) [16], hybridizing DE with the one-step k-means clustering and 2 multi-parent crossovers. Yildiz [31] developed a novel hybrid optimization algorithm called hybrid robust differential evolution (HRDE) by adding positive properties of Taguchis method to the DE algorithm to minimize the production cost associated with multi-pass turning problems.
Although these prominent variants of the DE algorithm largely optimize the DE process, almost all of them still suffer from problems such as difficult parameter setting, slow convergence rate, and premature convergence. Thus, this paper presents a novel adaptive DE algorithm that adopts a dynamic adjustment strategy including a directional strategy and cloud model, which we call ADEwDC. In the ADEwDC algorithm, the whole population is divided into three subgroups with each group dynamically selecting its trial vector generation from a constructed mutation and crossover pool according to its own convergence degree. The proposed algorithm realizes parameter control adaptively by referring to current fitness information. By introducing evolutionary direction [37], the convergence speed of the DE algorithm is further improved. The directional strategy is based on a greedy strategy, which means that individuals with good performance can evolve rapidly in the optimal evolution direction. These good individuals are selected according to their fitness, and the optimal direction is chosen according to the disparity between the target and trial vectors in the fitness space. However, the possibility of the algorithm falling into a local optimum is increased because of the rapid decline in population diversity. Thus, after further analysis, the ADEwDC algorithm improved the diversity of the population by employing the cloud model [13] in each generation. The characteristics of the cloud model, including randomness and stability, are included in the entire cloud droplet group, where randomness maintains the diversity of the population, while stability preserves the performance of excellent cloud droplets [14]. In other words, the proposed algorithm improves the convergence speed of the population, while at the same time, increasing the diversity and stability of the group. Computational experiments and comparisons show that ADEwDC overcomes the shortcomings of slow convergence rate and low efficiency in conventional DE algorithms, effectively avoids falling into a local optimum, and overall performs better than many state-of-the-art DE variants [4], such as JDE and JaDE, when applied to the optimization of benchmark global optimization problems.
The rest of the paper is arranged as follows. Section 2 introduces the traditional DE algorithm. In Section 3, the proposed ADEwDC algorithm is described in detail. The experimental design and results are presented and discussed in Section 4. Finally, the paper is concluded in Section 5.
2 The DE algorithm
Scientists and engineers from many disciplines often have to deal with the classic problems of search and optimization [4]. The DE algorithm is a simple population-based, stochastic parallel search evolutionary algorithm for global optimization and is capable of handling non-differentiable, nonlinear, and multimodal objective functions [9, 26]. In the DE algorithm, the population consists of real-valued vectors with dimension R D, which equals the number of design parameters. The size of the population is adjusted by parameter NP. The initial population is uniformly distributed in the search space \(\left [{\overset {\rightharpoonup }{X}}_{\min },{\overset {\rightharpoonup }{X}}_{\max }\right ]\), where \({\overset {\rightharpoonup }{X}}_{\min } = \left ({x_{\textit {min},1}},{x_{\textit {min},2}},{x_{\textit {min},3}},\ldots ,{x_{\textit {min},D}}\right )\) and \({\overset {\rightharpoonup }{X}}_{\max } = \left ({x_{\max ,1}},{x_{\max ,2}},{x_{\max ,3}},...,{x_{\max ,D}}\right )\). Each component is determined as follows:
where x i,j denotes the j th component of the i th individual, 0 denotes the initialized subsequent generation, and rand is a uniformly distributed random number between 0 and 1, which is instantiated independently for each component of the i th vector.
The traditional DE algorithm works through a simple cycle of stages, including mutation, crossover, and selection. Mutation and crossover are applied to each individual to produce the new population, followed by the selection phase, where each individual of the new population is compared with the corresponding individual of the old population, and the better of the two is selected as a member of the population in the next generation. A brief description of each of the evolutionary operators is given below.
Mutation
In the DE literature, a parent vector from the current generation is called the target vector, a mutant vector obtained through the differential mutation operation is known as the donor vector, and finally an offspring formed by combining the donor with the target vector is called a trial vector. There are many mutation strategies to generate donor vector \({\overset {\rightharpoonup }{V}}_{i}^{g}\), of which the most commonly used operator and one with the simplest form is ′ D E/r a n d/1/b i n ′, which is expressed as:
where \({\overset {\rightharpoonup }{X}}_{r_{1}^{i}}^{g}, {\overset {\rightharpoonup }{X}}_{r_{2}^{i}}^{g}, {\overset {\rightharpoonup }{X}}_{r_{3}^{i}}^{g}\) are sampled randomly from the current population in the g th generation. Indices r1i, r2i, and r3i are mutually exclusive integers randomly chosen from the range [1,N P], and which also differ from the index of the i th target vector, meaning r1i≠r2i≠r3i≠i∈{1,2,3,...,N P}. \({\overset {\rightharpoonup }{X}}_{r_{2}^{i}}^{g} - {\overset {\rightharpoonup }{X}}_{r_{3}^{i}}^{g}\) is a differential vector, and F is a real-valued mutation scaling factor that controls the amplification of the differential variation.
Crossover
After mutation, a binary crossover operation is applied to form the trial vector, \({\overset {\rightharpoonup }{U}}_{i}^{g} = \left (u_{i,1}^{g},u_{i,2}^{g},u_{i,3}^{g},\ldots ,u_{i,D}^{g}\right )\), to enhance the potential diversity of the population by exchanging the components of the donor vector \({\overset {\rightharpoonup }{V}}_{i}^{g}\) and target vector \({\overset {\rightharpoonup }{X}}_{i}^{g}\) according to the given probability, defined as CR. Each component of \({\overset {\rightharpoonup }{U}}_{i}^{g}\) is generated by the scheme outlined as:
where i denotes the i th individual, j denotes the j th dimension, g indicates the g th generation, r j ∈[0,1] is the j th evaluation of a uniform random number generator. CR is the crossover constant in the range [0,1], where zero means no crossover. r i ∈(1,2,3,...,D) is a randomly chosen index that ensures trial vector \({\overset {\rightharpoonup }{U}}_{i}^{g}\) gets at least one element from the donor vector \({\overset {\rightharpoonup }{V}}_{i}^{g}\), which is instantiated once per generation for each vector. Otherwise, no new parent vector would be produced and the population would remain unchanged. If the value of any dimension of the newly generated trial vector exceeds the pre-specified upper and lower bounds, it is set to the closest boundary value.
Selection
To keep the population size constant over subsequent generations, one-to-one greedy selection between a parent and its corresponding offspring is employed to decide whether the trial individual \({\overset {\rightharpoonup }{U}}_{i}^{g}\) should replace the target vector \({\overset {\rightharpoonup }{X}}_{i}^{g}\) as a member of the next generation according to their fitness values. For minimization problems, the one-to-one selection scheme is formulated as:
where \(f ({\overset {\rightharpoonup }{X}})\) is the objective function to be minimized. From the above description, if and only if the trial vector yields a better cost function value compared with its corresponding target vector in the current generation, it is accepted as the new parent vector in the next generation; otherwise, the target is once again retained in the population. As a result, the population either improves or remains the same in terms of fitness status, but never deteriorates.
The iterative procedure is terminated when any one of the following criteria is met: an acceptable solution is obtained, a state with no further improvement in the solution is reached, control variables have converged to a stable state, or a predefined number of iterations have been executed. Our proposed algorithm adopts a similar main framework as the traditional DE algorithm, but employs different strategies in the process of evolution.
3 Adaptive differential evolutionary algorithm with directional strategy and cloud model
Based on the conventional DE algorithm, the Adaptive Differential Evolutionary Algorithm with Directional Strategy and Cloud Model, ADEwDC for short, constructs a trial vector generation pool to effect the dynamic adjustment strategy. The whole population is divided into three subgroups, with each subgroup selecting a different trial vector generation strategy from the pool according to its own convergence degree. In each iteration, the control parameters are dynamically set based on the fitness of the individual compared with the optimal one in the whole group, with the specific value obtained by control parameter generation. To improve the convergence rate, the evolutionary direction is introduced, and ADEwDC chooses the best individuals to evolve in the optimal evolution direction, which is defined according to the evolution potential. To avoid premature convergence, ADEwDC utilizes the cloud model to increase the diversity of the whole population, and proposes the learning operator with cloud model by applying forward and reverse cloud generators, defined as MCG. The specific operations of ADEwDC are discussed in detail in this section.
3.1 Dynamic adjustment strategy
The self-adapting strategy is an important research area for DE algorithms [36], of which there are many prominent variants. However, the performance of DE is sensitive to the choice of mutation strategy and associated control parameters [17]. In other words, different mutation strategies with different parameter settings at different stages of the evolution may be more appropriate than a single mutation strategy with unique parameter settings. Therefore, as opposed to self-adaptation [8], this paper implements dynamic adjustment by adopting a distributed topology [7] and constructing a trial vector generation pool. This means that different mutation and crossover operations, chosen from the mutation and crossover operator pool are used for each subgroup, although a uniform selection operation is adopted. Meanwhile, the control parameters for mutation and crossover obtain their values adaptively based on the fitness space, particularly from information of the marked optimal individual in the population.
Since mainly mutation and crossover in ADEwDC are used to obtain the optimal value, which exceeds the overall performance of the parent generation, rapidly, this paper constructs a mutation and crossover operator pool to generate the donor vector \({\mathrm {P}}_{\overset {\rightharpoonup }{V}^{g}} = \left ({\overset {\rightharpoonup }{V}}_{1}^{g}, {\overset {\rightharpoonup }{V}}_{2}^{g}, {\overset {\rightharpoonup }{V}}_{3}^{g}, \ldots , {\overset {\rightharpoonup }{V}}^{g}_{NP}\right )\) using the following variants:
-
1)
DE/rand/1/bin
$$ {\mathrm{v}}_{i,j}^{g} = \left\{ {\begin{array}{cc} {x_{{r_{1}},j}^{g} + {F_{1}} \cdot \left(x_{{r_{2}},j}^{g} - x_{{r_{3}},j}^{g}\right)} & {\textit{if}({r_{j}} \le \textit{CR}||{n_{j}} = j)}\\ {x_{i,j}^{g}}&{\textit{otherwise}} \end{array}} \right. $$(5) -
2)
DE/rand/2/bin
$$ \begin{array}{lll} {\mathrm{v}}_{i,j}^{g} &= x_{{r_{1}},j}^{g} + {F_{1}} \cdot \left(x_{{r_{2}},j}^{g} - x_{{r_{3}},j}^{g}\right)& \\ &{\kern10pt}+ {F_{2}} \cdot \left(x_{{r_{4}},j}^{g} - x_{{r_{5}},j}^{g}\right)& {if({r_{j}} \le CR||{n_{j}} = j)}\\ &= {x_{i,j}^{g}} &{\kern43pt} {\textit{otherwise}} \end{array} $$(6) -
3)
DE/target-to-best/1/bin
$$ \begin{array}{lll} {\mathrm{v}}_{i,j}^{g} &= x_{i,j}^{g} + {F_{1}} \cdot \left(x_{\textit{gbest},j}^{g} - x_{i,j}^{g}\right)\\ & \quad + {F_{2}} \cdot \left(x_{{r_{1}},j}^{g} - x_{{r_{2}},j}^{g}\right)&{\textit{if}({r_{j}} \le \textit{CR}||{n_{j}} = j)}\\ &={x_{i,j}^{g}}&{\kern43pt} {\textit{otherwise}} \end{array} $$(7) -
4)
DE/target-to-best/2/bin
$$ \begin{array}{lll} {\mathrm{v}}_{i,j}^{g} &=& x_{i,j}^{g} + {F_{1}} \cdot \left(x_{\textit{gbest},j}^{g} - x_{i,j}^{g}\right) +{F_{2}} \cdot \left(x_{{r_{1}},j}^{g} - x_{{r_{2}},j}^{g}\right) \\ && \quad + {F_{3}} \cdot \left(x_{{r_{3}},j}^{g} - x_{{r_{4}},j}^{g}\right)~~~~~~~~~~{\kern3pt}{if({r_{j}} \le \textit{CR}||{n_{j}} = j)} \\ &=&{x_{i,j}^{g}}\qquad\qquad\qquad\qquad~~~~~~~~~~~~~{\kern32pt} {\textit{otherwise}} \end{array} $$(8) -
5)
DE with a neighborhood-based scheme
-
5.1)
Neighborhood vector
$$ {\overset{\rightharpoonup}{L}}_{i}^{g} = {\overset{\rightharpoonup}{X}}_{i}^{g} + \alpha_{1}\cdot \left({\overset{\rightharpoonup}{X}}_{\textit{nbest}} - {\overset{\rightharpoonup}{X}}_{i}^{g}\right) + \beta_{1} \cdot \left({\overset{\rightharpoonup}{X}}_{r1}^{g} - {\overset{\rightharpoonup}{X}}_{r2}^{g}\right) $$(9) -
5.2)
Population vector
$$ {\overset{\rightharpoonup}{G}}_{i}^{g} = {\overset{\rightharpoonup}{X}}_{i}^{g} + \alpha_{2}\cdot \left({\overset{\rightharpoonup}{X}}_{\textit{gbest}} - {\overset{\rightharpoonup}{X}}_{i}^{g}\right) + \beta_{2} \cdot \left({\overset{\rightharpoonup}{X}}_{r1}^{g} - {\overset{\rightharpoonup}{X}}_{r2}^{g}\right) $$(10) -
5.3)
Component of donor vector
$$ {\mathrm{v}}_{i,j}^{g} = \left\{ {\begin{array}{ll} {\omega \cdot g_{i,j}^{g} + (1 - \omega ) \cdot l_{i,j}^{g}}&{if({r_{j}} \le \textit{CR}||{n_{j}} = j)}\\ {\kern40pt}{x_{i,j}^{g}}&{\kern30pt} {\textit{otherwise}} \end{array}} \right. $$(11)
-
5.1)
where r 1≠r 2≠r 3≠r 4≠r 5≠i∈{1,2,3,...,N P}, j denotes the j th component, F 1,F 2,F 2,α 1,α 2,β 1,β 2 are all scaling factors to control the scale of the differential vectors, ω is a weighting factor applied to the information of the neighborhood and global population, and C R is the crossover rate to determine the source of each dimension of the offspring.
Of the five variants given above, since DE/target-to-best/1/bin and DE/target-to-best/2/bin rely on the best solution found, they usually have a faster convergence speed and perform well when solving unimodal problems. However, these algorithms are more likely to be trapped in a local optimum and to converge prematurely when solving multimodal problems. DE/rand/1/bin and DE/rand/2/bin usually demonstrate slow convergence speed with superior exploration capability. Therefore, they are usually better suited to solving multimodal problems. By including a neighborhood-based scheme [3, 5], different variants are employed to evolve individuals and set parameters, as discussed in the next section. Obviously, these five donor vector strategies constitute a strategy candidate pool with diverse characteristics.
In this paper, the distributed topology is simplified by dividing the population into three subgroups, while the selection criteria are strengthened by defining a variable, called the convergence degree.
Definition 1
Convergence degree is the average value of the accumulative fitness of each individual in the subgroup as the standard for determining the specific mutation and crossover operation. It is defined as:
where the accumulative fitness \(\textit {Fit}_{\Sigma }^{s} = \sum \limits _{i = 1}^{np} {f\left ({\overset {\rightharpoonup }{X}}_{i}^{g}\right )}\), s denotes the number of subgroups, and n p denotes the number of individuals in each subgroup. Considering the stability of the algorithm, the subgroup with the maximum convergence degree generates the donor vector by DE/rand/1/bin or DE/rand/2/bin, the one with the minimum degree uses DE/target-to-best/1/bin or DE/target-to-best/1/bin, and the remaining ones choose their variant randomly.
ADEwDC makes full use of the historical optimal information of the population and the fitness of each individual to set the control parameters, referred to as PSO. The ultimate evolution goal of the current individual can be regarded as the current group optimal individual with the optimal fitness, and this forms the important theoretical basis for setting the control parameters.
With reference to the relevant papers, the scaling factors F,α,β take values in [0.1,1] [19], the weighting factor ω is restricted to the range [0.05,0.95] [5], and C R takes a value in [0.1,0.9] [17]. Thus, integrating the above discussion, the dynamic adjustment strategy is given as follows.
At this point, ADEwDC generates the same number of individuals as in the parent generation, that is, the donor vector space \({\mathrm {P}}_{\overset {\rightharpoonup }{V}^{g}} = \left ({\overset {\rightharpoonup }{V}}_{1}^{g}, {\overset {\rightharpoonup }{V}}_{2}^{g}, {\overset {\rightharpoonup }{V}}_{3}^{g}, \ldots , {\overset {\rightharpoonup }{V}}^{g}_{\textit {NP}}\right )\).
3.2 Design of evolutionary direction
DE belongs to the class of stochastic optimization algorithms with randomness and unpredictability, leading to the arbitrariness of the offspring and the fact that good characteristics of parental individuals cannot be fully passed to the next generation. However, living things are not passive victims of their environment in nature and human society. Instead, they struggle to fit into the environment by constantly adjusting the evolutionary direction. Therefore, this paper adopts an indicative directional propagation strategy to alter the blindness of conventional DE and improve the convergence rate of the DE algorithm before the selection process.
Definition 2
Evolutionary direction is a description of the direction vector from the current individual to its offspring for any individual \({\overset {\rightharpoonup }{X}}_{i}^{g}\) in the population P g , such that ∀X⇀i g∈P g . It is defined as:
where \({\overset {\rightharpoonup }{X}}_{i}^{g} = \textit {offspring} \left ({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}}\right )\), \({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}} \in P_{(g -1)}\), and \({\overset {\rightharpoonup }{X}_{i}^{g}} \in P_{g}\).
In the current population, the evolutionary direction of individual \({\overset {\rightharpoonup }{X}_{i}^{g}}\) depends on itself and its corresponding parent in the previous generation of population \({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}}\), shown in the previous equation as \(\overrightarrow {\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}{\overset {\rightharpoonup }{X}}_{i}^{g}}\). The evolutionary directions of individuals in the first generation are selected randomly, and the optimal evolution direction is described below.
Definition 3
Optimal evolution direction D R o p t (P g ) is selected from the evolutionary orientation of the whole population through a series of specific operations, which is the best evolution direction. It is described as:
where \({\overset {\rightharpoonup }{X}_{i}^{g}} \in P_{g}\) and \({f\left ({\overset {\rightharpoonup }{X}}_{i}^{g}\right )}\geq f{\left ({\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}}\right )}\), and Θ s denotes the selection operation on the evolution direction.
ADEwDC defines the directional strategy as follows: Select the top n individuals with the maximum fitness in the donor vector space as generated in the previous subsection. Select the top m optimal evolution directions. Then, evolve the selected individuals along the selected directions and shape the progeny space with n⋅m individuals, where n⋅m>N P.
There are many criteria for selecting the optimal evolution direction, and the overall requirement is to improve the fitness of an individual after evolving along the selected orientation. In fact, the ultimate evolution goal of the current individual can be regarded as the current group optimal individual \({\overset {\rightharpoonup }{X}}_{gbest}\), which has the optimal fitness of the whole population, fming for a minimization problem. Thus, the goal vector is defined as \(\overrightarrow {\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}{\overset {\rightharpoonup }{X}}_{gbest}^{g}}\), and the motion vector as \(\overrightarrow {\overset {\rightharpoonup }{X}_{i}^{(g-1)^{\prime }}{\overset {\rightharpoonup }{X}}_{i}^{g}}\). Then, ADEwDC defines evolution potential as the factor used to choose the optimal directions.
Definition 4
Evolution potential ∇ D R is defined as the disparity between the goal vector and the motion vector in the fitness space, defined as:
Therefore, the optimal evolution direction based on maximum evolution potential is defined as:
The directional evolution strategy is given by the following algorithm.
From the optimal individuals and optimal evolution directions, the trail vector space \({\mathrm {P}}_{\overset {\rightharpoonup }{U}^{g}} = \left ({\overset {\rightharpoonup }{U}}_{1}^{g}, {\overset {\rightharpoonup }{U}}_{2}^{g}, {\overset {\rightharpoonup }{U}}_{3}^{g}, \ldots , {\overset {\rightharpoonup }{U}}^{g}_{n{\cdot }m}\right )\) is constructed, providing choices for the new generation.
3.3 Specific application of cloud model
By introducing the evolutionary direction, the convergence speed of the DE algorithm is improved further; however, the possibility of the algorithm falling into a local optimum rises markedly because of the rapid decrease in population diversity. To improve the diversity of the population, we introduce the cloud model [13], which describes individuals of the population through expectation Ex, entropy En, and hyper-entropy He.
Definition 5
Membership cloud [11] Let U denote a quantitative domain composed of precise numerical variables, with C the qualitative concept on U. If the quantitative value x∈U is a random realization of qualitative concept C, the confirmation of x on C can be denoted as μ(x)∈[0,1], which is a random number with stable tendency.
The distribution of x on U is called a cloud, x is called a cloud droplet, and the cloud consists of a series of cloud droplets.
The cloud droplets have a certain randomness, which maintains the diversity of individual stocks thereby avoiding a search for local extreme values, and stability characteristics, which protect the population of a more outstanding individual and thus the overall situation of extreme adaptive positioning. The membership cloud describes a concrete concept through expectation Ex, entropy En, and hyper-entropy He. Expectation Ex expresses the point that is most able to represent the domain of the concept and is the most typical sample of this concept to quantify. Entropy En represents the granularity of a concept that can be measured (the larger the entropy is, and the larger the granularity is, the more macro is the concept). It reflects the range of the domain space that can be accepted by the specific concept. Hyper-entropy He describes the uncertain measurement of entropy. It can be used to express the relationship between randomness and fuzziness.
As a specific kind of cloud, the normal cloud model has been proven to be universal [12], based on the normal distribution and Gauss membership function. Regarding probability, the normal distribution is the most commonly used form, which is described by expectation E and variance D. In fuzzy set theory, the bell-shape membership function \(\mu (x) = e^{\frac {{ - (x - a)^{2} }}{{2b^{2} }}}\) is also the most common membership function used in fuzzy sets. The normal cloud, described below, combines the characteristics of the two with an additional expansion.
Definition 6
Normal Cloud Model [14] Let U be the universe of discourse and à be a qualitative concept in U. If x∈U is a random instantiation of concept Ã, satisfying x∼N(E x ,E n ′ 2) and E n ′ 2∼N(E n ,H e 2), and the certainty degree of x belonging to concept à satisfies
then the distribution of x in universe U is called a normal cloud.
Knowledge is usually the association between concepts in the real world, between which the cause and effect relationship can be described by the membership cloud generator (MCG), including both a forward and reverse generator (for more information, see [11]). With the associated numerical characteristics, that is, Ex, En and He, the forward cloud generator can generate cloud drops (x,μ), where x is the quantity values and μ is the membership degree of x [10]. The reverse cloud generator is the other conversion model that can convert quantity numbers to a quality concept. It can convert accurate data (x 1,x 2,...,x n ) with membership degrees (μ 1,μ 2,...,μ n ) to a quality cloud concept expressed as numerical characteristics (E x ,E n ,H e ) [10]. The MCG based on cloud theory is summarized below.
Based on our analysis, ADEwDC incorporates the cloud model into the selection operation to remedy the diversity of population, and presents a novel operator called the learning operator with cloud model.
Definition 7
Learning operator with cloud model View the current population I λ as a cloud defined using C l o u d M (E x ,E n ,H e ), the eigenvalue calculated by the reverse normal cloud generator r C G (P λ ) to describe the overall information owned by it. For each individual in the suboptimal population, ∀d i ∈I λ−e, generate a new individual using the forward normal cloud generator f C G (E x ,E n ,H e ,(λ−e)), then the learning operator with cloud model is defined as:
where λ denotes the parents’ individuals space and e means the individuals selected from trail vector space. That is, part of the subgeneration is selected directly from the trial vector space, with count e, and the rest is generated randomly by the learning operator with cloud model according to the feature information about the known population.
The specific application of the cloud model is summarized below.
Thus far, we have introduced each part of ADEwDC: the dynamic adjustment strategy is used in the mutation and crossover process, while the design of the evolutionary direction and specific application of the cloud model are both applied in the selection process. The overall framework of ADEwDC is presented next.
3.4 ADEwDC
The proposed ADEwDC employs the dynamic adjustment strategy (see Section 3.1), the directional strategy (see Section 3.2), and the cloud model (see Section 3.3), introduced previously. In this section, we first present the overall framework of the proposed algorithm, and then analyze its discipline holistically.
3.4.1 Framework of the algorithm
3.4.2 Discipline of the algorithm
The main framework of ADEwDC is taken from the traditional DE algorithm, but with changed mutation and crossover strategies and an improved selection mechanism. As with the classical DE algorithm, the proposed algorithm defines a donor vector and a trial vector, where the former is the result of the mutation and crossover operation, while the latter is the result of directional evolution. The next generation may come from the parent generation, donor vector space (Algorithm 2), or trial vector space (Algorithm 3), but most new individuals must be generated by the cloud model (Algorithm 5).
-
a)
Dynamic adjustment strategy ADEwDC constructs a mutation and crossover pool to realize the dynamic adjustment strategy, including DE/rand/1/bin, DE/rand/ 2/bin, DE/target-to-best/1/bin, DE/target-to-best/2/bin, and DE with neighborhood-based scheme variants. It then defines the strategies to select the evolution operators (Definition 1) and set the control parameters (Algorithm 1) based on the constructed topology, which are summarized as the donor vector generation (Algorithm 2).
-
b)
Design of evolutionary direction The evolutionary direction is determined based on the theory that the individuals are not passive in joining the environment, but struggle to fit into the environment by constantly adjusting the direction of evolution. This algorithm defines the optimal direction (Definition 3) with evolution potential (Definition 4), and synthesizes the trial vector generation(Algorithm 3).
-
c)
Specific application of cloud model The cloud model is a useful mathematical tool with randomness and a stability tendency, which is employed to record the eigenvalues of the parent generation using the reverse cloud generator and generate cloud droplets randomly for most of the filial generation using the forward cloud generator (Definition 4). The specific operation is described by Algorithm 5.
-
d)
Selection strategy The next generation consists of two parts: some are directly selected as offspring from the vector space, consisting of the parent population, the donor vector space, and the trial vector space, whereas the greater part is generated by the MCG (Algorithm 4) based on the cloud model to compensate for the loss of population diversity, and which is described as:
$$ {P_{g + 1}} = {\varTheta_{s}^{e}}\left({P_{g}} \cup {P_{\overset{\rightharpoonup}{{V}}^{g}}} \cup {P_{\overset{\rightharpoonup}{{U}}^{g}}}\right) \cup \varUpsilon_{Cloud}^{\lambda - e} $$(20)where Θ s e means selecting e individuals from the specific vector space.
ADEwDC utilizes the dynamic adjustment strategy to select evolutionary variants and set control parameters based on the constructed topology. In fact, the subgroup with the maximum f i t s evolves with a slow convergence speed and superior exploration capability, while the minimum one evolves with a fast convergence speed and superior exploitation capability. Furthermore, each individual in the same subgroup sets its control parameters based on the gap between its fitness and the best fitness of the population. In short, this algorithm has good adaptability.
Furthermore, ADEwDC designs an evolutionary direction to improve the convergence rate, introduces the cloud model to ensure good diversity of the population, and applies the special selection strategy to balance the performance of the whole group. A series of experiments were carried out to confirm the effectiveness of ADEwDC as reported in the next section.
4 Experimental setup and results
To validate the ADEwDC algorithm, we selected several categories of global minimization benchmark functions to evaluate the proposed algorithm against other DE variants. These benchmark functions provide a balance between unimodal and multimodal functions, and were chosen from the set of 13 classic benchmark problems [30] that have frequently been used in the literature [25, 38]. The following problems were used:
-
(1)
Rastrigin function:
$${f_{1}}(X) = 10 \cdot D + \sum\limits_{i = 1}^{D} {\left[{x_{i}^{2}} - 10 \cdot \textit{cos}(2\varPi {x_{i}})\right]} $$with global optimum X ∗ = 0 and f(X ∗) = 0 for −5 ≤ x i ≤ 5.
-
(2)
Sphere function:
$${f_{2}}(X) = \sum\limits_{i = 1}^{D} {{x_{i}^{2}}} $$with global optimum X ∗=0 and f(X ∗)=0 for −100≤x i ≤100.
-
(3)
Rosenbrock function:
$${f_{3}}(X) = \sum\limits_{i = 1}^{D - 1} {\left(100{{({x_{i}^{2}} - {x_{i + 1}})}^{2}} + {{({x_{i}} - 1)}^{2}}\right)} $$with global optimum X ∗=(1,1,...,1) and f(X ∗)=0 for −100≤x i ≤100.
-
(4)
Ackley function:
$${f_{4}}(X) = - 20 \cdot {e^{- 0.2\sqrt {\frac{1}{D}\sum\limits_{i = 1}^{D} {{x_{i}^{2}}} } }} - {e^{\frac{1}{{\mathrm{D}}}\sum\limits_{{\mathrm{i}} = 1}^{D} {\cos (2{\Pi} {x_{i}})} }} + 20 + e $$with global optimum X ∗=0 and f(X ∗)=0 for −32≤x i ≤32.
-
(5)
Griewank function:
$${f_{5}}(X) = \sum\limits_{i = 1}^{D} {\frac{{{x_{i}^{2}}}}{{4000}} - \prod\limits_{i = 1}^{D} {\cos \left(\frac{{{x_{i}}}}{{\sqrt i }}\right) + 1} } $$with global optimum X ∗=0 and f(X ∗)=0 for −600≤x i ≤600.
Experiment 1 used a two-dimensional (2D) f 1 to explore the impact of the relevant parameters of Algorithm 6, while experiment 2 used the same test function to show the distribution of individuals in the solution space of each generation and verify its convergence. Experiment 3 used a 2D f 2, f 3, f 4, f 5 to explain the effectiveness of ADEwDC, compared with the conventional DE algorithm, DE/rand/1/bin. Next, ADEwDCf was evaluated on the CEC2013 benchmark problem set [15], as well as the CEC2005 benchmarks [23] in Experiment 4. In this paper, we compare ADEwDC with the following state-of-the-art DE algorithms: SHADE [24], CoDE [28], EPSDE [17], JADE [34], and dynNP-jDE [1] (an improved version of jDE [2]). Our comparisons were done with dimensions D=30 to analyze the effectiveness of the ADEwDC algorithm comprehensively. Finally, ADEwDC was used to solve the CEC2013 benchmark problems with dimensions D = 50 to prove its universality and robustness.
All experiments were executed on the following system:
-
OS: Windows 7 Professional
-
CPU: Intel(R) Xeon(R) CPU E5620 @ 2.40GHz 2.40GHz
-
RAM: 12.0GB
-
Language: Matlab
-
Compiler: Microsoft Visual C ++ 2012
4.1 Experiment 1 - related parameter settings
To investigate the relevant parameter settings, we utilized a 2D f 1 as the test function to explore the influence of parameters in Algorithm 6, with the function described as:
There are many local minima in the region of the value distribution of this function, and thus it is a good test case for measuring the effect of the relevant parameter settings. In this experiment, the number of top individuals (n) and the number of top directions (m) in step 2.3 of Algorithm 6 were observed, as well as the number of remaining individuals (e) from step 2.4 of Algorithm 6.
First, keeping the same e, we set n>m, n=m, n<m<2n, m=2n and m>2n as five cases. This experiment was repeated 10,000 times, and we randomly selected the set of experimental results shown in Fig. 1a. Then we calculate the average best solution of all the experimental results, which are shown in Table 1.
Then, keeping the same n and m, set e=20 % N P,e=40 % N P,e=50 % N P,e=60 % N P,e=65 % N P,e=70 % N P,e=75% N P and e=80 % N P as eight cases. This experiment was also repeated 10,000 times, and we randomly selected the set of experimental results shown in Fig. 1b, and then calculate the average best solution of all the experimental results in Table 2.
Based on the above analysis, ADEwDC achieves relatively good results when the number of top directions (m) is more than twice the number of top individuals (n), and the number of remaining individuals (e) is 6 0 % of the population size. In fact, n satisfies:
4.2 Experiment 2 - convergence analysis
As defined above, the 2D Rastrigin function was also used in this experiment, although it was applied to verify the convergence of Algorithm 6.
The initial population consisted of 40 points, 20 times the dimension recommended by Price [20], distribution randomly in the region between [−5,5] and [−5,5]. Meanwhile, n=5,m=12, ensure n⋅m>N P and m>2n, then e=24, which is 60 % of the population size. The distribution of the initial population is shown in Fig. 1a. The fitness function is the function value of the concrete points in the area. From the test results shown in Figs. 2a–f, we can see that after the evolution of 12 generations, the individuals distribution converges well. We listed the individuals of the 2nd, 4th, 8th, 11th, 12th generations on the region as shown in Figs. 2b–f, respectively.
To further verify the convergence of ADEwDC, it was compared with the standard DE, sDE for short, with DE/rand/1/bin (with F=0.5,C R = 0.3 in our experiments). The results of evolving 15 generations shown in Fig. 3 were randomly selected from 50 trial runs. In the search procedure for the optimal solution, we used the distribution area of all the points to represent the search space, changes in which are illustrated in Fig. 3a. Obviously, the search space of ADEwDC tends to zero much faster than that of sDE, after evolving only five or six generations. We plot the best individual for each evolutionary generation, as well as its abscissa and ordinates in Figs. 3b–d, respectively. In Fig. 3, it can be seen that ADEwDC converges to the optimal solution much faster than sDE.
Overall, the results of experiment 2 show that ADEwDC can effectively reduce the search space and rapidly converge to the optimal solution, which is consistent with the directional strategy. In the course of evolution, excellent individuals evolve along the optimal evolution directions, enabling the entire population to obtain the optimal value easily and quickly.
4.3 Experiment 3 - effectiveness analysis
In this experiment, ADEwDC was used to optimize the 2D f 2, f 3, f 4,f 5 to investigate the effectiveness of the algorithm compared with the standard DE algorithm (sDE). The results of evolving 50 generations for each function, shown in Fig. 4, were randomly selected from 100 trial runs.
For the Ackley and Griewank functions, the results by ADEwDC are better than those by sDE in terms of both convergence speed and accuracy, as clearly shown in Figs. 4c and (d). For the Sphere and Rosenbrock functions, as shown in Figs. 4a and b, ADEwDC converges faster than sDE and obtains the desired optimal solution. To illustrate the effectiveness of ADEwDC further, in Table 3 we list one full trial result, randomly selected from the 100 experimental results for each function.
From all the test results, we find that ADEwDC converges to the optimal solution quickly and has good robustness. In contrast, sDE is more likely to fall into a local optimum as shown in Fig. 5. Moreover, 1.0×10−3 is defined as the acceptable level, which means that the run is judged to be successful if a solution obtained by an algorithm falls between the acceptable level and the actual global optimum [33]. And we compare the frequency of premature convergence and the average value of all the test results between ADEwDC and sDE in Table 4.
Generally, the payoff for increasing diversity is a slower (although more efficient) convergence speed. Nevertheless, Table 4 indicates that ADEwDC is not at the cost of the convergence speed to obtain the diversity of the solutions; conversely, it improves the diversity with MCG (Algorithm 4), maintaining a good convergence rate at the same time which has been confirmed in the previous content [6, 35].
From the above experimental results and analysis, it can been seen that compared with the standard DE algorithm, Algorithm 6 has a faster convergence rate, obtains more accurate optimization results, and has better robustness. In short, ADEwDC is more effective than the standard DE because of the cloud model, which enhances the population diversity and prevents the algorithm from falling into a local optimum.
4.4 Experiment 4 - competitiveness analysis
In this section, we evaluate the performance of ADEwDC on the CEC2013 benchmark problem set [15], compared with SHADE [24], CoDE [28], EPSDE [17], JADE [34] and dynNP-jDE [1]. Then, using the CEC2005 benchmarks [23], ADEwDC is compared with SHADE to verify its validity and accuracy further. For each comparative algorithm, we used the control parameter values suggested in the cited papers. ADEwDC was executed on the same system as that given for the previous experiments, while comparative data for SHADE were taken from its original paper [24]. The source programs for CoDE, EPSDE, and JADE were based on code received from the original authors. The number of dimensions was set to D=30 and the maximum number of objective function calls per run was calculated as D×10,000 (that is, 300,000) when comparing ADEwDC with the other algorithms. Meanwhile, N P=100,n=8,m=20, and e=60. Finally, we executed ADEwDC on the CEC2013 benchmarks with the number of dimensions set to D=50 as a high-dimensional case to prove its universality and robustness.
4.4.1 Experiment on the CEC2013 benchmarks
In this experiment, we performed our evaluation following the guidelines of the CEC2013 benchmark competition [15]. The search space was set as [−100,100]D for all the selected problems, the results of which are shown in Table 5. In the table, the mean and standard deviation of the error (difference) between the best fitness values found in each run and the optimal value are shown. The + , −, ≈ indicate whether a given algorithm performed significantly better (+ ), significantly worse (−), or not significantly different better or worse (≈) compared to ADEwDC according to the Wilcoxon rank-sum test, which is a nonparametric alternative to the two-sample t-test based solely on the order of the observations from the two samples (significance threshold p≤0.05) [29]. Functions f 1∼f 5 are unimodal, while f 6∼f 20 are multimodal. f 21∼f 28 are composite functions combining multiple test problems into a complex landscape [15].
The best result for each benchmark function is highlighted in Table 5, while Table 6 shows the statistical ranking according to the respective performance of the DE algorithms. The algorithms are ranked based on the average best solution in each row of Table 5, and the final rank of the algorithms is determined by the average rank calculated according to problem feature in each column. From Table 6, we see that SHADE, ADEwDC and JADE achieve the best performance on unimodal problems. The good performance of JADE on the unimodal functions is consistent with previous results [28]. For the basic multimodal functions, ADEwDC performs relatively well, although for several of the problems including f 13,f 18, JADE and CoDE perform particularly well. For the complex, composite functions, the best performer is dynNP-jDE (possibly owing to its population size reduction strategy), followed by SHADE, ADEwDC, JADE, CoDE and EPSDE. Finally, based on the statistical data given in the bottom three rows of Table 5, ADEwDC achieves better performance than CoDE, EPSDE and JADE, and similar performance to dynNP-jDE, but does not surpass the performance of SHADE on these 28 problems. Still, ADEwDC is better suited to multimodal problems than SHADE. This is probably because the dynamic adjustment strategy gives ADEwDC extensive adaptability, which has advantages in dealing with multimodal problems.
4.4.2 Experiment on the CEC2005 benchmarks
To verify the validity and accuracy further, we compared ADEwDC with SHADE on the CEC2005 benchmarks [23]. CEC2005 provides many classic benchmark functions, which have been tested by many original authors of algorithms. Functions f 1∼f 5 are unimodal, while the others are multimodal. f 6∼f 12 are basic functions, f 13∼f 14 are expanded functions, and f 15∼f 25 are hybrid composition functions. For this experiment, comparative data were once again taken from [24].
In Table 7, it is shown that ADEwDC performs very well with multimodal problems, especially on the basic and hybrid composition functions, which coincides with the previous experimental analysis. Overall, ADEwDC is slightly better than SHADE. For the expanded functions, ADEwDC achieves similar performance to SHADE. Nonetheless, ADEwDC does not perform as well on the unimodal problems, especially for f 4,f 5, which is also consistent with the previous analysis. Generally, ADEwDC is ideal for multimodal optimization problems.
4.4.3 High dimensional experiment on the CEC2013 benchmarks
We also executed ADEwDC on the CEC2013 benchmarks, setting the number of dimensions D=50 as a high-dimensional case to prove the universality and robustness of the proposed algorithm. The maximum number of objective function calls per run was set as D×10,000 (that is, 500,000). From Table 8, it can clearly be seen that ADEwDC is also suitable for high-dimensional problems, especially for multimodal optimization functions, which is consistent with our previous experimental analysis and conclusions.
5 Conclusions and future work
A novel adaptive DE algorithm with directional strategy and cloud model was presented in this paper. The DE algorithm is known to be an efficient and powerful optimization algorithm, which is used widely in scientific and engineering fields. Considering that most DE algorithms still suffer from a number of problems such as difficult parameter setting, slow convergence rate, and premature convergence, the proposed algorithm utilizes a dynamic adjustment strategy to select evolutionary variants and set control parameters based on the constructed topology, adopts a directional strategy to evolve outstanding individuals and improve the convergence speed of the algorithm, and maintains the diversity of the population by employing the cloud model. Results of the experiments confirm that ADEwDC achieves good performance in terms of convergence, stability, and precision, which greatly contributes to overcoming the low efficiency in conventional DE algorithms. Meanwhile, ADEwDC effectively avoids falling into a local optimum. Our future work involves strengthening the theoretical analysis of ADEwDC, further improving its accuracy and stability, and generalizing it to solve constraint and multi-objective optimization problems in practical applications.
References
Brest J, Mauec MS (2008) Population size reduction for the differential evolution algorithm. Appl Intell 29(3):228–247
Brest J, Greiner S, Boskovic B, Mernik M, Zumer V (2006) Self-adapting control parameters in differential evolution: a comparative study on numerical benchmark problems. IEEE Trans Evol Comput 10(6):646–657
Cai YQ, Wang JH (2013) Differential evolutionwith neighborhood and direction information for numerical optimization. IEEE Trans on Cybernetics 43(6):2202–2215
Das S, Suganthan PN (2011) Differential evolution: A survey of the state-of-the-art. Evol Comput 15(1):4–31
Das S, Abraham A, Chakraborty UK, Konar A (2009) Differential evolution using a neighborhood based mutation operator. IEEE Trans Evol Comput 13(3):526–553
Ding JL, Liu J, Chowdhury KR, Zhang WS, Hu QP, Lei J (2014) A particle swarm optimization using local stochastic search and enhancing diversity for continuous optimization. Neurocomputing 137(0):261–267
Dorronsoro B, Bouvry P (2011) Improving classical and decentralized differential evolution with new mutation operator and population topologies. IEEE Trans Evol Comput 15(1):67– 98
Eiben AE, Hinterding R, Michalewicz Z (1999) Parameter control in evolutionary algorithms. IEEE Trans Evol Comput 3(2):124–141
Gao XZ, Wang XL, Ovaska SJ (2009) Fusion of clonal selection algorithm and differential evolution method in training cascade-correlation neural network. Neurocomputing 72(10–12):2483–2490
Gao Y (2009) An optimization algorithm based on cloud model. In: 2009 International Conference on Computational Intelligence and Security, pp 84–87
Li DY, Du Y (2005) Artificial Intelligence with Uncertainty (in Chinese). National Defense Industry Press, Beijing
Li D Y, Liu C Y (2005) Study on the universality of the normal cloud model. Eng Sci 3(2):18–24
Li DY, Meng HJ, Shi XM (1995) Membership clouds and membership cloud generators (in chinese). Journal of Computer Research and Development 32(6):15–20
Li DY, Liu CY, Gan WY (2009) A new cognitive model: Cloud model. Int J Intell Syst 24(3):357–375
Liang JJ, Qu BY, Suganthan PN, Hernández-Díaz AG (2013) Problem definitions and evaluation criteria for the cec 2013 special session on real-parameter optimization. Tech. rep., Computational Intelligence Laboratory, Zhengzhou University, Zhengzhou China
Liu G, Li YX, Nie X, Zheng H (2012) A novel clustering-based differential evolution with 2 multi-parent crossovers for global optimization. Appl Soft Comput 12(2):663–681
Mallipeddi R, Suganthan PN, Pan QK, Tasgetiren MF (2011) Differential evolution algorithm with ensemble of parameters and mutation strategies. Appl Soft Comput 11(2):1679–1696
Omran MGH, Salman A, Engelbrecht AP (2005) Self-adaptive differential evolution. Springer, Berlin Heidelberg, pp 192–199. chap International Conference, CIS 2005, Xian, China, December 15–19, 2005, Proceedings Part I
Pan QK, Suganthan PN, Ling W, Liang G, Mallipeddi R (2011) A differential evolution algorithm with self-adapting strategy and control parameters. Comput Oper Res 38(1):394–408
Price KV (1999) An introduction to differential evolution. In: Corne D, Dorigo M, Glover F (eds) New ideas in optimization. McGraw-Hill, Ltd., UK, pp 79–108
Qin AK, Huang VL, Suganthan PN (2009) Differential evolution algorithm with strategy adaptation for global numerical optimization. IEEE Trans Evol Comput 13(2):398– 417
Storn R, Price K (1997) Differential evolution - a simple and efficient adaptive scheme for global optimization over continuous spaces. J Glob Optim 11(4):341–359
Suganthan PN, Hansen N, Liang JJ, Deb K, Chen YP, Auger A, Tiwari S (2005) Problem definitions and evaluation criteria for the cec 2005 special session on real-parameter optimization. Nanyang Technological University, Tech. rep.
Tanabe R, Fukunaga A (2013) Success-history based parameter adaptation for differential evolution. In: Evolutionary Computation (CEC), 2013 IEEE Congress on, pp 71–78
Vafashoar R, Meybodi MR, Azandaryani AHM (2012) Cla-de: a hybrid model based on cellular learning automata for numerical optimization. Appl Intell 36(3):735–748
Varadarajan M, Swarup KS (2008) Differential evolutionary algorithm for optimal reactive power dispatch. Electrical Power and Energy Systems 30(8):435–441
Wang X, Zhao SG (2013) Differential evolution algorithm with self-adaptive population resizing mechanism. Math Probl Eng 2013(2013):1–14
Wang Y, Cai Z, Zhang Q (2011) Differential evolution with composite trial vector generation strategies and control parameters. IEEE Trans Evol Comput 15(1):55–66
Wilcoxon F (1945) Individual comparisons by ranking methods. Biometrics Bulletin 1(6):80–83
Yao X, Liu Y, Lin G (1999) Evolutionary programming made faster. IEEE Trans Evol Comput 3(2):82–102
Yildiz AR (2013) Hybrid taguchi-differential evolution algorithm for optimization of multi-pass turning operations. Appl Soft Comput 13(3):1433–1439
Zaharie D (2003) Control of population diversity and adaptation in differential evolution algorithms. In: Matousek R, Osmera P (eds) In: 2003 9th international conference on soft computing, pp 41–46
Zhan ZH, Zhang J, Li Y, Shi YH (2011) Orthogonal learning particle swarm optimization. IEEE Trans Evol Comput 15(6):832–847
Zhang JQ, Sanderson AC (2009) Jade: adaptive differential evolution with optional external archive. IEEE Trans Evol Comput 13(5):945–958
Zhang JZ, Ding XM (2011) A multi-swarm self-adaptive and cooperative particle swarm optimization. Eng Appl Artif Intell 24(6):958–967
Zhao SZ, Suganthan PN, Das S (2011) Self-adaptive differential evolution with multi-trajectory search for large-scale optimization. Soft Comput 15(11):2175–2185
Zhao ZQ, Gou J, Wang J (2010) Directional evolutionary algorithm based on fitness gradient of individuals (in chinese). Pattern Recognit Artif Intell 23(1):29–37
Zheng YJ, Ling HF, Xue JY (2014) Ecogeography-based optimization: Enhancing biogeography-based optimization with ecogeographic barriers and differentiations. Comput Oper Res 50(0):115–127
Zhu CM, Ni J (2012). Cloud model-based differential evolution algorithm for optimization problems. In: 2012 Sixth International Conference on Internet Computing for Science and Engineering, pp 55–59
Acknowledgements
This work was supported by the National Natural Science Foundation of China (No. 61103170, 51305142, 61305085), the Program for Prominent Young Talent in Fujian Province University (No. JA12005), the Program for Prominent Young Talent in Fujian Province University (No. JA12005), and the Promotion Program for Young andMiddle-aged Teachers in Science and Technology Research at Huaqiao University (No. ZQN-PY211).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Gou, J., Guo, WP., Hou, F. et al. Adaptive differential evolution with directional strategy and cloud model. Appl Intell 42, 369–388 (2015). https://doi.org/10.1007/s10489-014-0592-3
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-014-0592-3