1 Introduction

Many engineering and scientific researches involve optimization (Yi et al. 2016; Deb 2000). Some of them contain several constraints which the optimal solution must meet. Without a loss of generality, the mathematical formula of constrained optimization problems (COPs) can be expressed as follows (Sun et al. 2016; Yu et al. 2016; Garg 2016; Zhang et al. 2015; Chuang et al. 2016; Asafuddoula et al. 2015):

$$\begin{aligned}&\min \,\,f(\vec {x})\nonumber \\&\quad \hbox {s.t.}\left\{ \begin{array}{lll} g_j (\vec {x})\le 0,&{}j=1,2,\ldots ,q \\ h_j (\vec {x})=0,&{}j=q+1,\ldots ,m \\ \end{array}\right. , \end{aligned}$$
(1)

where \(\vec {x}=(x_1 ,x_2 ,\ldots ,x_n )\in \Omega \) is the decision vector generated in the decision space S. \(\Omega \) denotes the feasible region. S is the n-dimensional search space by the boundary constraints, \(L_i \le x_i \le U_i ,i=1,\ldots ,n,\) where \(U_i \) and \(L_i \) are the upper boundary and the lower boundary of \(x_i \), respectively. \(f(\vec {x})\) is the decision function. The constraints \(g_j (\vec {x})\) and \(h_j (\vec {x})\) denote the inequality and the equality, respectively.

To solve COPs, the equality constraints are generally converted into inequality constraints:

$$\begin{aligned} \left| {h_j (\vec {x})} \right| -\delta \le 0, \end{aligned}$$
(2)

where \(j=q+1,\ldots ,m\) and \(\delta \) is the tolerance value for the equality constraints. According to Liang et al. (2006b), \(\delta =0.0001\). The absolute value operator can be removed by transforming the Eq. (2) into inequality constraints:

$$\begin{aligned} -\delta \le h_j (\vec {x})\le \delta \rightarrow \left\{ {\begin{array}{ll} h_j (\vec {x})\le \delta &{}\, j=q\!+\!1,\ldots ,m \\ -h_j (\vec {x})\le \delta &{}\, j=m\!+\!1,\ldots ,2m-q \\ \end{array}}\right. \end{aligned}$$
(3)

So, the constraints of COPs can be presented as follows:

$$\begin{aligned} \left\{ {\begin{array}{ll} g_j (\vec {x})\le 0&{}\quad j=1,\ldots ,q \\ h_j (\vec {x})\le \delta &{}\quad j=q+1,\ldots ,m \\ -h_j (\vec {x})\le \delta &{}\quad j=m+1,\ldots ,2m-q \\ \end{array}} \right. \end{aligned}$$
(4)

Then, the total constraint violation (CV) can be expressed briefly as follows:

$$\begin{aligned} CV(\vec {x})= & {} \sum _{j=1}^{{2}m-q} {cv_j (\vec {x})} ,cv_j (\vec {x})\nonumber \\= & {} \left\{ {\begin{array}{ll} \max (g_j (\vec {x}),0)&{}\, j=1,\ldots ,q \\ \max (h_j (\vec {x})-\delta ,0)&{}\, j=q+1,\ldots ,m \\ \max (-h_j (\vec {x})-\delta ,0) &{}\, j=m+1,\ldots ,2m-q \\ \end{array}} \right. \end{aligned}$$
(5)

There are many different kinds of COPs. The main differences among them are variables and constraint properties. Some of them are integer, real; some of them are discrete (Elsayed et al. 2012). Many optimization algorithms cannot perform well as they are computationally expensive or easily get stuck at local optima (Wang and Cai 2011). On the other hand, many researches have indicated that evolutionary algorithms (EAs) have advantage when solving COPs, especially the genetic algorithm (GA) (Long 2014; Barbosa and Lemonge 2003; Garcia-Martinez et al. 2008), differential evolution (DE) algorithm (Yong et al. 2008; Mezura-Montes and Coello 2004; Price et al. 2005), evolutionary strategies (ES) (Wei et al. 2011; Hansen and Ostermeier 2001), PSO (Kennedy and Eberhart 1995; Liang et al. 2006a; Sun et al. 2011), ant colony-based method (Mahdavi and Shiri 2015; Long et al. 2016), grey wolf (Arora Mehak Kohliand 2017y), and artificial bee colony(ABC) (Liang et al. 2015; Li and Yin 2014; Karaboga and Akay 2011). In the past decades, many techniques have been proposed to solve COPs, such as feasibility rules, stochastic ranking, \(\varepsilon \)-constrained method, penalty function, multi-objective concept, and ensemble of constraint-handling techniques (Mezura-Montes and Coello 2011; Wang et al. 2016; Takahama and Sakai 2006; Runarsson and Yao 2000; Wang et al. 2007; Tessema and Yen 2006).

Differential evolution is one of the EAs created by Storn and Price (1997). The DE is very simple. The algorithm is efficient and effective.DE has attracted increasing attention. It is originally designed to solve continuous unconstrained problems (Fan and Yan 2016; Qin et al. 2009; Mallipeddi et al. 2011; Brest et al. 2006; Zhang and Sanderson 2009; Yu et al. 2015a, b). Recently, an improved differential evolution by differential vector archive and hybrid repair method is proposed to solve single optimization problem. The proposed algorithm has achieved good performance on IEEE CEC 2013 (Zhang and Zhang 2016). A prior knowledge-guided DE (called PKDE) is implemented, in which the macro-levels and micro-levels of prior knowledge are extracted to guide the search direction. To validate the performance of PKDE, two sets of benchmark test functions from IEEE CEC2005 and IEEE CEC2014 are used. The experimental results have indicated that PKDE is competitive (Fan et al. 2016). Besides, enhanced adaptive differential evolution (EADE) algorithm is presented to solve large-scale global optimization problems. A new mutation rule is introduced to use the information of good and bad vectors in the DE population of EADE (Mohamed 2017). There is no doubt that these newly proposed algorithms have made great contributions to the development of DE algorithm and EAs community.

Now DE algorithm is extended to solve COPs. The adaptive ranking mutation operator (ARMOR) for DE is proposed to solve COPs. The ARMOR is to make DE converge faster and achieves feasible solutions faster (Gong et al. 2015). Inspired by the fact that in modern society, a dual-population differential evolution (DPDE) with coevolution is designed for COPs (Gao et al. 2015). For a better coverage of the problem characteristics, a self-adaptive differential evolution algorithm is introduced (Elsayed et al. 2014). A ranking-based mutation operator and an improved dynamic diversity mechanism based on DE are proposed (Gong et al. 2014a). With the improved augmented Lagrangian approach, a cooperative coevolutionary DE algorithm is proposed (Ghasemishabankareh et al. 2016). Multi-objective optimization is combined with differential evolution (CMODE) to deal with COPs (Wang and Cai 2012a). A dynamic hybrid framework (DyHF) is designed for solving COPs (Wang and Cai 2012b). Objective function information is incorporated into the feasibility rule to form a new algorithm: FROFI (Wang et al. 2016). To cope with COPs, an improved DE and a novel archiving-based adaptive trade-off model are employed (Jia et al. 2013).

The above algorithms based on DE make great contributions to solve COPs. When using DE to solve COPs, it generates feasible individuals and infeasible individuals during the evolution. It also has to tackle the constraints. Two issues are of great importance for solving COPs when using DE. One is how to cope with the feasible and infeasible solutions. Feasibility rule emphasizes the fact that feasible solution is better than infeasible solution(Deb 2000). However, a lot of researches have indicated that infeasible solution is also very critical to find the global optima (Singh et al. 2008; Mezura-Montes and Coello 2005; Lin et al. 2014). So, how to extend three criteria to reconsider the infeasible solution should be observed. The second one is how to generate the offspring. The conventional DE uses just single evolution strategy during evolution. However, different problems need different strategies with different parameter values relying on the properties of problems. A single mutation strategy often cannot meet the requirement.

Based on discussed above, an effective DE algorithm is proposed by two efforts. First, the novel algorithm extends three criteria of feasibility rule to seven criteria. The effort is to compare feasible solution over infeasible solution. Then, two kinds of offspring generation methods are designed. The first one is DE/rand/1 as it is said to be the most widely employed and successful mutation strategy according to the current research. The second one is the integration of current-to-rand/1 and current-to-best/1 (Wang and Cai 2011). The idea is significant different from the above research. In order to validate the effectiveness of the proposed algorithm, benchmark functions from CEC2006 are adopted. The experimental results have demonstrated that the proposed algorithm is effective compared with other optimization algorithms. Then, the proposed algorithm is employed to solve the weight in AHP.

The paper is organized as follows. Literature review is given in Sect. 2. In Sect. 3, DE is briefly introduced and the proposed algorithm is presented. In Sect. 4, experiments are done based on standard benchmarks. The proposed method is employed to solve the real application in Sect. 5. The conclusions are made in Sect. 6.

2 Literature review

A comprehensive survey on COPs is introduced, in which following techniques are mainly discussed to solve COPs (Mezura-Montes and Coello 2011).

(1) Penalty function method

The approach introduces a penalty function or a coefficient into the original objective function (Michalewicz 1995; Coit et al. 1996; Tessema and Yen 2009). The aim is to penalize solutions which violate constraints. To effectively cope with constraint problems, Lin proposed a hybrid algorithm the rough penalty genetic algorithm (RPGA), which contained genetic algorithm and rough set theory. The aim of RPGA was to effectively resolve COPs and achieve robust solutions (Lin 2013). As it is difficult for static penalty approach to adjust penalty factors, the dynamic penalty approach is put forward. A novel optimization algorithm was designed by utilizing a penalty function in the objective function to treat violation (Homaifar et al. 1994). However, when using penalty function approach to solve COPs, it is still the most difficult to find appropriate penalty coefficients, which guide the search direction toward the optimum (Joines and Houck 1994).

(2) Multi-objective optimization techniques

The technique usually transforms the COPs into two objective function problems, the original objective function and the violation objective function. Then, the problems are handled by multi-objective optimization techniques. A generic optimization framework based on genetic algorithms was proposed to solve COPs (Venkatraman and Yen 2005). Wang and Cai proposed an effective evolutionary algorithm to optimize COPs according to multi-objective optimization principle (Cai and Wang 2006). Multi-objective optimization technique based on differential evolution algorithm was employed to tackle these problems (Qu and Suganthan 2011). Multi-objective optimization was combined with differential evolution to deal with COPs (Wang and Cai 2012c). However, solving multi-objective optimization problems themselves is still very complicated.

Fig. 1
figure 1

The flowchart of conventional DE

(3) Gradient method

An improved \(\varepsilon \)-constrained differential evolution algorithm was presented, which was integrated with pre-estimated comparison gradient (Yi et al. 2016). A novel distributed gradient algorithm was proposed for COPs (Yi et al. 2015). A new line search method as well as combining it with the spectral projected gradient approach was used to solve COPs (Yu 2008). A non-monotonic interior backtracking line search method combined with an affine scaling reduced preconditional conjugate gradient path approach was proposed (Zhu 2007). However, gradient method is applicable to two kinds of problems: a few variables or having convex feasible regions. The burden of computational cost will become increasingly heavier when the amount of variables increases (Deb 1995; Reklaitis et al. 1983).

(4) Ensemble of constraint-handling methods

Motivated by the no free lunch theorems, Mallipeddi and Suganthan proposed an ensemble of four constraint techniques (Mallipeddi and Suganthan 2010). Elsayed et al. employed two constraint-handling techniques to solve CMOPs (Ali and Kajee-Bagdadi 2009). Feasibility rule, \(\upvarepsilon \)-constrained, and an adaptive penalty function were combined to solve CMOPs (Tasgetiren et al. 2010).

(5) Feasibility rule

Let \(x_{1}\) and \(x_{2}\) be feasible solutions, and \(x_{3}\) and \(x_{4}\) are infeasible solutions.

  1. 1.

    Feasible solution is preferred to infeasible solutions without considering their objective value. So, \(x_{1}\) and \(x_{2}\) are preferred to \(x_{3}\), \(x_{4}\), respectively.

  2. 2.

    If \(f(x_{1})\) is smaller than \(f(x_{2})\), then \(x_{1}\) is preferred to \(x_{2}\).

  3. 3.

    If \(CV(x_{3})<CV(x_{4})\), then \(x_{3}\) is preferred to \(x_{4}\).

There are also a number of other criteria, which are similar to the above. They are imposed to deal with COPs (Deb 2000; Richardson et al. 1989). These implementations employed different measures of constraint violations. By designing new comparison rules, the infeasible individuals with better objective function were made full use of during the evolution (Zheng et al. 2012). For instance, a novel diversity mechanism was designed (Mezura-Montes and Coello 2005). However, these criteria are too simple. They directly consider that infeasible solution is worse than feasible solution. In fact, infeasible solution is also of importance to maintain the diversity of the population.

Table 1 The pseudo codes of offspring generation methods

3 DE

3.1 Conventional DE

In conventional DE, mutation, crossover, and selection are very critical. The main framework of DE is similar to EAs, which is exhibited in Fig. 1.

According to the main steps of DE, the trial vector \(V_{i,G} =\{v_{i,G}^1 ,v_{i,G}^2 ,\ldots ,v_{i,G}^D \}\) and \(U_{i,G}^ =\{u_{i,G}^1 ,u_{i,G}^2 ,\ldots ,u_{i,G}^D \}\) are the trial vectors. The former is generated by mutation strategy in the mutation step. The latter is generated in crossover step.

Mutation strategy plays an important role during evolution. Price and Storn employed DE/rand/1/bin. It is widely used. The main strategies implemented in the DE are listed as follows (Sharma et al. 2012):

  • DE/best/1:

    $$\begin{aligned} V_{i,G} =X_{best,G} +F.(X_{r_1^i ,G} -X_{r_2^i ,G} ) \end{aligned}$$
    (6)
  • DE/best/2:

    $$\begin{aligned} V_{i,G} =X_{best,G} +F.(X_{r_1^i ,G} -X_{r_2^i ,G} )+F.(X_{r_3^i ,G} -X_{r_4^i ,G} ) \end{aligned}$$
    (7)
  • DE/current-to-best/1:

    $$\begin{aligned} V_{i,G} =X_{i,G} +F.(X_{best,G} -X_{i,G} )+F.(X_{r_1^i ,G} -X_{r_2^i ,G} ) \end{aligned}$$
    (8)
  • DE/rand/1:

    $$\begin{aligned} V_{i,G} =X_{r_1^i ,G} +F.(X_{r_2^i ,G} -X_{r_3^i ,G} ) \end{aligned}$$
    (9)
  • DE/rand/2:

    $$\begin{aligned} V_{i,G} =X_{r_1^i ,G} +F.(X_{r_2^i ,G} -X_{r_3^i ,G} )+F.(X_{r_4^i ,G} -X_{r_5^i ,G} ) \end{aligned}$$
    (10)
  • DE/current-to-rand/1 :

    $$\begin{aligned} U_{i,G} =X_{i,G} +K.(X_{r_1^i ,G} -X_{r_i ,G} )+F.(X_{r_2^i ,G} -X_{r_3^i ,G} ). \end{aligned}$$
    (11)

The indices \(r_1^i ,r_2^i ,r_3^i \) are mutually exclusive integers randomly generated within the range [0, 1], which are also different from the index i. F and K are the mutation scale factors, which are used in controlling the amplification of the differential variation.

3.2 The proposed algorithm

3.2.1 Mutation strategies

The first mutation strategy DE/rand/1 was developed for DE (Price et al. 2005) and is one of the most widely and successful employed strategies (Babu and Jehan 2003). But DE/best/2 may have some superiority to DE/rand/1 (Gamperle et al. 2002; Pahner and Hameyer 2000). Besides, it is beneficial to incorporate best solution information during evolution (Mezura-Montes et al. 2006). In contrast to DE/rand, algorithm can get benefits by holding best solution information. However, the behavior may also cause problems such as premature due as population diversity is reduced (Zhang and Sanderson 2009). Besides, it has been proved that current-to-best/1 strategy performs well when solving un-modal problems as it has the best individual information (Iorio and Li 2004). However, when solving multimodal problems, it easily traps in a local optimum. From the above discussion, it can be found that only one strategy cannot perform well during evolution. Here, two generation methods are used. The first one is DE/rand/1 as it is said to be the most wide and successful strategy (Babu and Jehan 2003). The second one is the integration of current-to-best/1 and current-to-rand/1 (Mezura-Montes et al. 2006).

Based on above discussion, the implementation of the mutation process is given in Table 1, where parameter \(rand\in [0,1]\) is a uniformly random number and \(\phi \) is the feasible solution proportion of the last population. At the early stage, there are little feasible solutions and \(\phi \) is very small. Thus, condition of current-to-rand can be met more frequently. At the middle and later stage, more and more feasible solutions are generated. Mutation strategy current-to-best will be employed more frequently, in which best solution is incorporated. By balancing the two strategies with the help of parameter \(\varphi \), the algorithm can realize global and local search.

The crossover operation is the same as conventional DE. So, the two trial vectors \(U_{i,G}^1 ,U_{i,G}^2 \) are generated. Now, the key is how to compare the three vectors \(X_{i,G} ,U_{i,G}^1 ,U_{i,G}^2 \).

3.2.2 Selection mechanism

The comparison among the three vectors\(X_{i,G} ,U_{i,G}^1 ,U_{i,G}^2 \) is to select the one which is more suitable for the next evolution. There are two kinds: feasible solution and infeasible solution. If all the solutions are feasible, the one having lower objective function value is chosen. The main problems are to compare the feasible solution over infeasible solution and infeasible solution over infeasible solution. As the infeasible solution is also very important to keep the diversity of population, the algorithm should make good use of information from infeasible solution. The \(\upvarepsilon \)-constrained method introduces a relaxation of the limit \(\upvarepsilon \) to consider a solution as feasible (Takahama et al. 2005).

Let \(f(x_{1} ),f(x_2 )\) and \(CV(x_1 ),CV(x_2 )\) be the fitness values and the constraint violation. Then, \(\varepsilon \) level comparison between \(f(x_{1} )\) and \(f(x_2 )\) is defined as follows:

$$\begin{aligned}&(f(x_1 ),CV(x_1 ))<_\varepsilon (f(x_2 ),CV(x_2 ))\nonumber \\&\quad \Leftrightarrow \left\{ {\begin{array}{ll} f(x_1 )<f(x_2 )&{} if\,{CV(x_1 ),CV(x_2 )\le \varepsilon } \\ f(x_1 )<f(x_2 )&{} if\, {CV(x_1 )=CV(x_2 )}\\ {CV(x_1 )<CV(x_2 )}&{} {otherwise} \\ \end{array}} \right. \end{aligned}$$
(12)

However, when the mechanism deals with complicated COPs, the global search ability is also limited. Combined with \(\upvarepsilon \)-constrained method and feasibility rule, the novel criteria are proposed in Table 2. In the procedure, there are two parameters e and P. The parameter e is the tolerance parameter, which is used to compare with total constraints CV in Eq. (5). The function of parameter e is similar to \(\varepsilon \) in Eq. (12). The selection probability parameter Pis the probability parameter, which is between 0.9 and 1.

For the first criteria, the lower objective function value is preferred, which is the same to the feasibility rule (Deb 2000). In the second criterion, if the total constraints of two vectors are both less than e, the lower objective function value is also chosen. This rule is similar to \(\upvarepsilon \)-constrained (Takahama et al. 2005). The total constraints of two vectors are both more than e in the third criterion, and the lower total constraints vector is selected with probability P. The rule is similar to feasibility rule (Deb 2000). If \({\textit{rand}}>P\), the vector with lower fitness value is selected to keep the diversity of population in the fourth criterion. If \(U_{i,G}^1 \) is infeasible solution and \(U_{i,G}^2 \) is feasible solution, \(U_{i,G}^2 \) can be accepted in the fifth and sixth criteria. For the last criterion, the total constraints of \(U_{i,G}^2 \) are less than e and it is selected to maintain the diversity of population.

3.2.3 The procedure of the proposed algorithm e-DE

According to the discussion above, the main procedure of the novel algorithm e-DE is presented as follows:

figure a

4 Experiments

4.1 Test functions

Table 2 The pseudo codes of selection mechanism for feasible solution and infeasible solution
Table 3 Benchmark test functions

For the purpose of validating the performance of the algorithm e-DE, twenty-two functions are selected from CEC2006 (Liang et al. 2006b) in Table 3.These well-known benchmark functions include linear, nonlinear, polynomial, quadratic, and so on, where \(f(x^{*})\) is the objective function value for optimal solution \(x^{*}\).

For these test functions, 25 independent runs were performed. Parameters setting: NP=50,CR=0.9, F=0.8, e=1, selection probability parameter \(P\in [0.9,1]\). These parameters were maintained in all runs.

4.2 Experimental results

Table 4 Function errors acquired when \(5\times 10^{3}\), \(5\times 10^{4}\), and \(5\times 10^{5}\) for test functions g01–g12

The results of e-DE are presented in Table 4 and Table 5.On the basis of the reference (Liang et al. 2006b), the best, the worst, median, mean, and standard deviation (Std) of the function error value (\(f(x)\hbox {-}f(x^{*}))\) of the acquired best results with \(5\times 10^{3}\), \(5\times 10^{4}\), and \(5\times 10^{5}\) FES for each benchmark function are presented, where \(x^{*}\) is the best known solution and \(f(x^{*})\) is the best function value.

Table 5 Function errors achieved when \(5\times 10^{3}\), \(5\times 10^{4}\), and \(5\times 10^{5}\) for test functions g13–g19, g23, g24

The experimental results have demonstrated that the e-DE has the ability to succeed in finding feasible solutions, which are close to the best known solutions for g08 in \(5\times 10^{3} \) function evaluations (FES) and for g03, g04, g06, g09, g11, g12, g16, and g24 in \(5 \times 10^{4}\) FES. When FES is set to \(5\times 10^{5}\), the best solutions of experiments are close or equal to the optimal known solutions.

When FES is equal to \(5\times 10^{5}\), all the experimental results for functions g01, g11, and g12 are the same as the optimal values. The e-DE finds solutions close to the best solutions for the fourteen test functions (g04-g10, g14-g16, g18, g19, g23, g24). The acquired best solutions approximate the known optimal values with differences \((10^{-10})\). It has indicated that the e-DE can obtain the results, which are approximate to optimal solutions for these five test functions. The best results for the g03, g05, g09 problems are even beyond to its optimal solutions as equality constraints are converted into inequality constraints with a tolerance of 0.0001, revealing that the algorithm e-DE can consistently find the best solutions in 25 experiments.

Notably, the standard deviations of the e-DE are very small in fourteen test problems (g01, g03-g09, g11, g12, g14-g16, g24). This finding implies that the novel algorithm is robust when solving COPs.

4.3 The performance of success performance

Table 6 Number of FES to reach the success condition, success rate, feasible rate, and success performance

The number of min, median, max, mean FES required in each run to find a feasible solution is shown in Table 6. Besides, the success rate, the feasible rate, and the success performance are also given in Table 6.

Once one feasible solution is found during the run, it is a feasible run. If the feasible solution \(\vec {x}\) satisfies \(f(\vec {x})-f(\vec {x}^{*})\le 0.0001\), it is a successful run.

  • Feasible rate = (feasible runs)/(total runs)

  • Success rate = (successful runs)/(total runs)

  • Success performance = mean (FES for successful runs)\(\times \)(total runs)/(successful runs)

Table 6 exhibits the overall performance of e-DE. It can be noticed that the feasible rate is 100% for these functions. With the exception of test functions g02, g13, g17, and g21, the success rate of 100% has been achieved for the rest functions.

The e-DE needs fewer than 5000 FES for one function to reach the success condition and requires fewer than 50,000 FES to achieve success for ten functions. No more than 200,000 FES for 19 test functions and no more than 300,000 FES for 21 functions are demanded, respectively.

In practical terms, the proposed e-DE presents superior SP that has a lot of inequality constraints and disconnected feasible regions, for instance g16, g12, and g24. Moreover, it is observed that the proposed e-DE can successfully cope with great majority of test problems. Functions include g01, g07, g10, g16, and g18 with a larger amount of inequality constraints compared to the other COPs. The results have indicated that the e-DE with the dynamic control parameters is effective to deal with such constraints.

In addition, it is mentioned that some functions are considered as the comparatively more difficult functions among these test functions (Liang and Suganthan 2006; Huang et al. 2006). For instance, g02, g13, and g17 are the multimodal COPs. Similar to other EAs, the algorithm also has difficulty to tackle these complicated functions. However, the algorithm achieves success rates as high as 92, 44, and 44% for g02, g13, and g17, respectively. It is obviously better than majority of EAs.

Table 7 Mean number of FES to find the first feasible solution

In order to approve the velocity that e-DE finds the feasible region, Table 7 presents the mean number of FES to find feasible solutions over 25 runs. The proposed algorithm e-DE requires no more than 1000 FES for ten test problems, between 1000 and 10,000 FES for five test problems. For the rest test problems, the numbers of FES finding the first feasible solutions are between 10,000 and 50,000. The finding reveals that the algorithm has good convergence speed.

4.4 Comparison against the latest DE variants

Table 8 The best, mean, and standard variation of six algorithms for g01–g13
Table 9 The best, mean, and standard variation of six algorithms for g14–g19, g21, and g24

In the previous subsection, the efficacy of e-DE is verified through the benchmark functions. In this section, e-DE is compared with state-of-the-art EAs for the COPs. These algorithms are CMODE (Wang and Cai 2012a), DyHF (Wang and Cai 2012b), ICDE (Jia et al. 2013), rank-iMDDE (Gong et al. 2014a), and CCiALF (Ghasemishabankareh et al. 2016). The CMODE algorithm is proposed, which combines multi-objective optimization with differential evolution to deal with COPs. A dynamic hybrid framework, (DyHF) is designed for solving COPs. This framework consists of two major steps: global search model and local search model. To solve COPs, an improved DE and a novel archiving-based adaptive trade-off model are employed to form a new algorithm ICDE. An improved constrained differential evolution (rank-iMDDE) method is proposed, where a ranking-based mutation and an improved dynamic diversity mechanism are presented. A cooperative coevolutionary differential evolution algorithm based on the improved augmented Lagrangian approach (CCiALF) is proposed for solving COPs. We choose these five EAs for comparisons due to their good performance obtained and the same Max_NFEs (240,000) used. As it is fair to keep the same Max_NFEs for each algorithm, the Max_NFEs of e-DE is also set to 240,000. Otherwise, the comparisons are unfair and unfaithful.

In Tables 8 and 9, the best, mean, and standard deviation of the objective function values of the final solutions for each algorithms are shown. The overall best results among the six EAs are highlighted in boldface. Note that the results of rank-iMDDE and CCiALF are directly obtained from their corresponding literature.

From the Tables 8 and 9, it is very clear that the six algorithms have achieved the same performance in ten functions (g03, g07, g08, g09, g10, g11, g12, g15, g16, and g24), in which three equality functions and seven inequality functions are included. These results have indicated that the proposed algorithm has the same search ability to the five algorithms. CMODE has achieved the best performance in g14 and g18. DyHF and e-DE have the best performance in g01, g06, g13, g14, g18, g19, and g23. ICDE has the superior performance in g01, g06, g14, g17, g18, g19, g21, and g23. Rank-iMDDE has the superior performance in g01,g06,g13,g18, and g21.

According to the results shown in Tables 8 and 9, it can be clearly seen that e-DE consistently obtains highly competitive results in all functions compared with other five EAs. In terms of the best results, e-DE gets the best or similar values among the six EAs in all 22 functions. With respect to the mean results, only in two functions g17 and g21, e-DE is slightly worse. In the rest 20 functions, e-DE is able to obtain better or similar results compared with other five EAs.

Friedman test has been used to rank these algorithms. It is a nonparametric statistical test .It is used to detect differences in treatments across multiple test attempts. Based on the mean values in Tables 8 and 9, the final rankings obtained by the Friedman test are shown in Table 10. With respect to the average rankings of all algorithms by the Friedman test, Table 10 shows that our proposed e-DE and rank-iMDDE get the first ranking among six algorithms, followed by DyHF, CMODE, ICDE, and CCiALF. It denotes that the e-DE can perform as well as or even superior to other algorithms. In other words, the advantages of the proposed e-DE method are stated and proven.

Table 10 Average rankings of CMODE, DyHF, ICDE, rank-iMDDE, CCiALF, and e-DE by the Friedman test for the 22 functions in terms of the mean values

It is reported that the test function g17 has the new best known function value 8853.533875 (Brajevic 2015). The results also have been achieved for the four algorithms (CMODE, DyHF, ICDE, and e-DE). The finding also indicates that the four algorithms have good abilities to deal with COPs.

4.5 The parameter sensitivity analysis

Fig. 2
figure 2

e-DE with different parameters on g01–g03

Fig. 3
figure 3

e-DE with different parameters on g04–g19, g21

Fig. 4
figure 4

e-DE with different parameters on g23 and g24

The parameter e needs to be tested as it is one of the most important parameters during the evolution. In this section, the impact of this parameter on the algorithm is focused on. The e-DE runs 25 times on each function with ten different parameters of 1, 2, 3, 4, 5,6,7,8,9,10, random between 1 and 10, respectively.

Figures 2, 3, and 4 reveal the box plots of \(\log (f(\vec {x})-f(\vec {x}^{*})+\exp (-10))\) that e-DE algorithm obtained in the 25 runs. \(f(\vec {x}^{*})\) is the objective function value for known optimal solution \(\vec {x}^{*}\), and \(f(\vec {x})\) is the optimal solution obtained by the algorithm in each run.

From Figs.  2, 3, and 4, it is found that the proposed algorithm is not sensitive to parameter value on g01, g03, g04, g06, g08, g11, g12, g15, g16, and g24,in which three equality constrains and seven inequality constrains are included. The parameter e has little influence on g02, g05, g07, g09, g10, g13, g18, and g19 and has some impact on g14, g17, g21, and g23.

In particular, test function g02 has a nonlinear objective function and 20 decision variables; test function g16 has thirty-four nonlinear inequality constraints and four linear inequality constraints. The feasible region is limited. With ten different parameters, the algorithm can obtain the known optimal values in 25 runs. The performance of e-DE algorithm is also stable to find the known optimal solutions. Therefore, the e-DE algorithm is less sensitive to the parameter e between 1 and 10 for most of test functions. Therefore, the algorithm is robust.

4.6 Effectiveness of the selection mechanism

Experiments were performed to validate the effectiveness of the algorithm for rand/1/bin and combined mutation strategies. The former one is the e-DE with the rand/1/bin (e-DE-1), and the latter one is the current-to-best/1 and current-to-rand/1 (e-DE-2). The results obtained from the above experiments are compared with the e-DE. The comparisons of four representative test functions g14, g17, g19, and g21 are listed in Table 11.

Test function g14 has ten decision variables, three equality constraints, and a nonlinear objective function. Note that e-DE-2 can obtain similar results with e-DE, which reveals that adding rand/1/bin cannot give negative effects on the diversity of population. The mean and STD results of e-DE and e-DE-2 are better than those of e-DE-1. With respect to test function g17, e-DE-1 and e-DE-2 have not achieved any success during the evolution. The success rate of proposed algorithm is 44%, which is the highest among the three algorithms. The above phenomenon signifies that the algorithm including two strategies has higher convergence speed. Function g19 has a nonlinear objective function with fifteen decision variables. The results of e-DE and e-DE-2 are much better than those of e-DE-1. It manifests that rand/1/bin strategy deteriorates the convergence speed. However, the success rate of e-DE-1 is higher than e-DE-2 for function g21. Function g21 is a complex constrained problem. As the e-DE-2 has the information of the optimal individual, it degrades the diversity of the population.

According to the above discussion, e-DE-1 with rand/1/bin strategy can enhance the diversity of the population and deteriorate the convergence speed, e-DE-2 with current-to-best/1 and current-to-rand/1 can accelerate the convergence speed and degrade the diversity of the population, and e-DE with the above two strategies can trade off the diversity of the population and the convergence speed very well. The comparison above confirms that strategy selection is reasonable.

5 Solve weights in analytic hierarchy process (AHP)

AHP, developed by Saaty (1980), is used to tack multiple criteria decision making (MCDM) in real applications (Wu et al. 2016; Xiaobing et al. 2011; Gong et al. 2014b). MCDM is to screen, prioritize, rank, or select a set of alternatives under usually independent, incommensurate, or conflicting attributes. In AHP, multiple pair-wise comparisons are from a standardized comparison scale of nine levels shown in Table 12.

Suppose that \(C = \{C_{j}{\vert } j =1,2{\ldots }n \}\) be the set of criteria. Evaluation matrix Acan be gotten, in which every element \(a_{ij}\, (i,j=1,2{\ldots }n)\) represents the relative weights of the criteria illustrated:

$$\begin{aligned} A=\left[ {{\begin{array}{llll} {a_{11} }&{} {a_{12} }&{} \ldots &{} {a_{1n} } \\ {a_{21} }&{} {a_{22} }&{} \ldots &{} \ldots \\ \ldots &{} \ldots &{} {a_{{ii}} }&{} \ldots \\ {a_{n1} }&{} {a_{n2} }&{} \ldots &{} {a_{nn} } \\ \end{array} }} \right] \end{aligned}$$
(13)

If matrix A is completely consistent, it has to comply with following condition:

$$\begin{aligned} a_{ii}= & {} \frac{w_i }{w_i }=1 \nonumber \\ a_{ji}= & {} \frac{w_j }{w_i }=\frac{1}{a_{ij} } \nonumber \\ a_{ij} a_{jk}= & {} \frac{w_i }{w_j }\times \frac{w_j }{w_k }=\frac{w_i }{w_k }=a_{ik}. \end{aligned}$$
(14)

According to above properties, the following equations can be obtained:

$$\begin{aligned}&\sum _{k=1}^n (a_{ik} w_k )= \sum _{k=1}^n {({w_i }/{w_k })w_k =nw_i ,i=1,2,\ldots ,n} \end{aligned}$$
(15)
$$\begin{aligned}&\sum _{i=1}^n {\left| {\sum _{k=1}^n {(a_{ik} w_k )-nw_i } } \right| } =0. \end{aligned}$$
(16)

In other words, if the judgment matrix Ameets the Eq. (16), it is completely consistent. However, it is very difficult to achieve the condition in the real application. In fact, the matrix A has just to meet the satisfactory consistency. The Eq. (16) can be converted into the following format:

$$\begin{aligned}&\min \, CIF(w)=\frac{\mathop {\sum }\nolimits _{i=1}^n {\left| {\mathop {\sum }\nolimits _{k=1}^n {(a_{ik} w_k )-nw_i } } \right| } }{n} \end{aligned}$$
(17)
$$\begin{aligned}&0<w_k <1,\mathop {\sum }\nolimits _{k=1}^n {w_k =1} \end{aligned}$$
(18)

The smaller the CIF is, the more consistent the matrix A is. So, the weight acquisition is converted into the single objective optimization with constraint. The objective is to minimize CIF, and the constraint is:\(0<w_k <1,\sum _{k=1}^n {w_k =1} \). Thus, it is a typical COP and can be solved by the proposed algorithm. For example,

$$\begin{aligned} A=\left[ {{\begin{array}{ccc} 1&{} 3&{} 7 \\ {1/3}&{} 1&{} 5 \\ {1/7}&{} {1/5}&{} 1 \\ \end{array} }} \right] \end{aligned}$$

The min CIF is as follows:

$$\begin{aligned} \min CIF(w)=\frac{\mathop {\sum }\nolimits _{i=1}^3 {\left| {\mathop {\sum }\nolimits _{k=1}^3 {(A(k,i)\times w_k )-nw_i } } \right| } }{n}. \end{aligned}$$
(19)

The CIF can be solved by the proposed algorithm, and convergence graph is presented in Fig. 5. To verify the rational results, both weights from the proposed algorithm and AHP theory are presented in Table 13, respectively. It can be observed that the weights are consistent with the AHP theory. Both weights have little difference. The results have indicated that the proposed algorithm has the ability to solve real-application COPs.

Table 11 Comparison of e-DE with e-DE-1 and e-DE-2
Table 12 Standardized comparison scale of nine levels
Fig. 5
figure 5

Convergence graph for min CIF

Table 13 The weights comparison between algorithm and AHP theory

6 Conclusions

An extended effective differential evolution algorithm is designed, named e-DE. The framework of e-DE is the same as the conventional DE. It is mainly composed of seven criteria and two offspring generation approaches. The seven criteria are used to compare the feasible solutions over infeasible solutions, which is quite different from traditional research. The two kinds of offspring generation approaches combine rand/1/bin, current-to-best/1, and current-to-rand/1. The experimental results based on CEC2006 have revealed the effectiveness of the proposed algorithm. The proposed algorithm can reach 100% feasible rate. In addition, compare to some state-of-the-art optimization algorithms, the algorithm is competitive. Besides, the proposed algorithm is used to solve the weights in AHP. The results are consistent with the weights from AHP theory.

In the following studies, we will use the algorithm to solve emergency management optimization problems, in which a lot of COPs are involved.