1 Introduction

Most real-world engineering optimization problems are multiobjective in nature, since they normally have several (possible conflicting) objectives that must be satisfied at the same time. Instead of aiming to find a single solution, the multiobjective optimization methods try to produce a set of good “trade-offs” from which the decision maker may select one.

Over the past two decades, a great amount of multiobjective evolutionary algorithms have been proposed [1]. Evolutionary algorithms seem particularly suitable to solve multiobjective optimization problems, because they deal simultaneously with a set of possible solutions. This allows us to find several Pareto optimal solutions in a single run of the algorithm, instead of performing a series of separate runs as in the case of the traditional mathematical programming techniques [1]. In addition, evolutionary algorithms can easily deal with discontinuous or concave Pareto fronts, whereas these two issues are a real concern for mathematical programming techniques.

Recently, a general-purpose local-search heuristic algorithm named Extremal Optimization (EO) was presented by Boettcher and Percus [2]. EO is based on the Bak–Sneppen model [3], which shows the emergence of self-organized criticality (SOC) [4] in ecosystems. The evolution in this model is driven by a process where the weakest species in the population, together with its nearest neighbors, is always forced to mutate. The dynamics of this extremal process exhibits the characteristics of SOC, such as punctuated equilibrium [3]. EO opens the door to applying non-equilibrium process, while the simulated annealing (SA) applies equilibrium statistical mechanics. In contrast to genetic algorithm (GA) which operates on an entire “gene-pool” of huge number of possible solutions, EO successively eliminates those worst components in the sub-optimal solutions. Its large fluctuations provide significant hill-climbing ability, which enables EO to perform well particularly at the phase transitions. EO has been successfully applied to some NP-hard combinatorial optimization problems such as graph bi-partitioning [2], TSP [2], graph coloring [5], spin glasses [6], MAXSAT [7].

So far there have been some papers which studied the multiobjective optimization using extremal dynamics. Ahmed and Elettreby [8] introduced random version of Bak–Sneppen model. They also generalized the single objective Bak–Sneppen model to a multiobjective one by weighted sum of objectives method. The method is easy to implement but its most serious drawback is that it cannot generate proper members of the Pareto-optimal set when the Pareto front is concave regardless of the weights used [9]. Galski et al. [10] presented a multiobjective version of the Generalized Extremal Optimization (GEO) algorithm, called M-GEO. Since the fitness assignment in the M-GEO is not based on the Pareto dominance strategy, M-GEO belongs to the non-Pareto approach [10].

In this paper, we develop a novel Pareto-based algorithm named Multiobjective Population-based Extremal Optimization (MOPEO), which can be considered as a variation of EO to solve multiobjective optimization problems. The fitness assignment of MOPEO is based on the Pareto domination, which is popularly used by many existing multiobjective evolutionary algorithms. Our approach has been validated by five standard test functions reported in the specialized literature and compared against three highly competitive multiobjective evolutionary algorithms: the Nondominated Sorting Genetic Algorithm-II (NSGA-II) [11], the Pareto Archived Evolution Strategy (PAES) [12] and the Strength Pareto Evolutionary Algorithm (SPEA) [13]. The simulation results indicate that MOPEO may be a good alternative to solve the multiobjective optimization problems.

2 Extremal optimization

2.1 Bak–Sneppen model

To make a good understanding of the underlying mechanism of EO, we will first describe the Bak–Sneppen model [3] in detail. Species in the Bak–Sneppen model are located on the sites of a lattice. Each species is assigned a fitness value randomly with uniform distribution. At each update step, the worst adapted species is always forced to mutate. The change in the fitness of the worst adapted species will cause the alteration of the fitness landscape of its neighbors. After a number of iterations, the system evolves to a highly correlated state known as self-organized criticality (SOC). In the SOC state, a little change of one species will result in co-evolutionary chain reactions called “avalanches”.

2.2 Extremal optimization

Inspired by the Bak–Sneppen model, Boettcher and Percus proposed the EO algorithm for a minimization problem as follows [2]:

  1. 1.

    Randomly generate a solution S. Set optimal solution S best = S and the minimum cost function C(S best) = C(S).

  2. 2.

    For the current solution S,

    1. (a)

      evaluate the fitness λ i for each variable x i , i∈{1, 2, ..., n},

    2. (b)

      rank all the fitnesses and find the variable x j with the lowest fitness, i.e., λ j  ≤ λ i for all i,

    3. (c)

      choose one solution S′, in the neighborhood of S, such that the jth variable must change its state,

    4. (d)

      accept S = S′, unconditionally,

    5. (e)

      if C(S) < C(S best) then set S best = S.

  3. 3.

    Repeat at Step 2 as long as desired.

  4. 4.

    Return S best and C(S best).

3 Multiobjective population-based extremal optimization

It has been proved that evolutionary algorithms are suitable for the multiobjective optimization problems due to its ability to create multiple Pareto-optimal solutions in a single simulation run [1]. Furthermore, a simple evolutionary algorithm can be extended to maintain a diverse set of solutions also because of its population mechanism. However, the traditional EO performs a search through sequential changes on a single solution, namely, the point-to-point search rather than the population based search applied in GA. In order to extend EO to solve the multiobjective problems, we developed a novel real-coded EO search algorithm, so-called MOPEO, through introducing the population search strategies being popularly used in evolutionary algorithms into EO. Similar to the evolutionary algorithms, MOPEO operates on the evolution of solutions generation after generation. Let each solution consists of n decision variables. Similar to EO, the MOPEO performs only one operation, i.e., mutation, on each variable of every solution.

3.1 Main algorithm

For a multiobjective optimization problem, the proposed MOPEO algorithm works as follows.

  1. 1.

    Generate initial population with N solutions, S i  = (x i1,x i2,...,x in ), i∈{1, ..., N}, randomly and uniformly. Set the external archive empty. Set iteration = 0.

  2. 2.

    For each solution S i , i∈{1, ..., N},

    1. (a)

      generate n offspring of the current solution S i by performing mutation on each variable one by one;

    2. (b)

      perform dominance ranking on the n offspring and then obtain their rank numbers, i.e., r ij ∈[0,n−1], j∈{1, ..., n};

    3. (c)

      assign the fitness λ ij  = r ij for each variable x ij , j∈{1, ..., n};

    4. (d)

      if there is only one variable with fitness value of zero, the variable will be considered as the worst adapted species; otherwise, the diversity preservation mechanism is invoked. Assuming that the weakest species is x iw with fitness λ iw  = 0, w∈{1, ..., n},

    5. (e)

      perform mutation only on x iw while keeping other variables unchanged, then get a new solution S iw ;

    6. (f)

      accept S i  = S iw unconditionally.

  3. 3.

    Identify the nondominated solutions in the new population.

  4. 4.

    Update the external archive by comparing the nondominated solutions in the new population with those in the archive.

  5. 5.

    If the external population has exceeded the maximum allowable capacity, reduce the external archive by crowding-distance computation; else if the iterations reach the predefined maximum number of the generations, go to Step 6; otherwise,set iteration = iteration + 1, and go to Step 2.

  6. 6.

    Return the external archive as the Pareto-optimal set.

3.2 Fitness assignment

It is important to note that, in MOPEO, each decision variable in one solution is considered as a species. Being different from the multiobjective evolutionary algorithm, the fitness assignment of MOPEO is carried out in each solution rather than in the population. That is, we will find out all the worst adapted species for all solutions in the current population via fitness assignment and perform mutation on them. In our work, we adopt the Pareto-based fitness assignment strategy. We use the dominance ranking [14], i.e., the number of solutions by which a solution is dominated, to determine the fitness value for each solution. In MOPEO, the dominance ranking is carried out in the offspring, which are generated by mutating the variables of the current solution in turn. Therefore, the nondominated offspring are ranked as zero, whilst the worst possible ranking is the number of decision variables minus one. So the species corresponding to the offspring with the fitness value of zero is considered as the weakest one and will be chosen to mutate.

To be clearer, we illustrate the process of dominance ranking in MOPEO using Fig. 1a. The locations of the solutions in the objective space (marked with black solid circles) will change to new ones (shown with grey dashed circles) in the next generation via mutating their weakest species. For example, given a solution S i  = (x 1, x 2, x 3, x 4), of which the location in the objective space is denoted by the circle i, we can identify the weakest species by mutating the four variables one by one and then performing dominance ranking on the newly generated offspring. First, an offspring S iA = (x 1 , x 2, x 3, x 4) can be obtained by mutating x 1 to x 1 ,, keeping other variables unchanged. Similarly, the other three offspring, i.e., S iB = (x 1, x 2 , x 3, x 4), S iC = (x 1, x 2, x 3 , x 4), S iD = (x 1, x 2, x 3, x 4 ,), are generated. The four white dashed circles (i.e., A, B, C, D) in Fig. 1a indicate the locations of four newly generated offspring (i.e., S iA, S iB, S iC, S iD) in the objective space, respectively. The next location of the solution S i in the objective space depends on the dominance ranking number of the four newly generated offspring. It can be seen from Fig. 1a that, for the circle i, circle A is nondominated by the other three dashed circles (i.e., B, C, D). Thus, the rank number of A is zero. Hence, the species x 1 corresponding to A is considered as the weakest species and the solution S i will change to S iA in the next step.

Fig. 1
figure 1

a Shows the process of dominance ranking in MOPEO; b shows the diversity preservation in MOPEO

If there exists more than one worst adapted species, i.e., at least two variables own the same fitness value of zero, then the following diversity preservation mechanism will be invoked.

3.3 Diversity preservation

The goal of introducing the diversity preservation mechanism is to maintain a good spread of solutions in the obtained set of solutions. In this study, we propose a new approach to keep good diversity of nondominated solutions. It is worth pointing out that our approach does not require any user-defined parameter for maintaining diversity among population members. Assuming that one solution has at least two species with the same fitness value of zero, we consider the species, by which the new solution generated approaches the less crowded region in the external archive, as the weakest one. As shown in Fig. 1b, the solution i has two offspring (i.e., A and B) who do not dominate each other. Suppose two nondominated solutions, i.e., C and D, who exist in the external archive, have the minimum Euclidean distances from A and B, respectively. Then we calculate the density value of C and D in the external population by using the k-nearest neighbor method. It can be seen from Fig. 1b that C is less crowded than D. So we choose the one corresponding to C, i.e., A, as the new location of solution i in the next step.

The k-nearest neighbor method requires sorting the external population according to each objective function value in ascending order of magnitude. The density values of the boundary solutions equal to the Euclidean distances of them from their nearest neighbors, while the density value of all other intermediate solutions can be calculated by the following:

$$\widehat{Y}_k({\mathbf X})=\frac{1}{k} \sum_{x_i\in N_k({\mathbf X})}{y_i},$$
(1)

where y i is Euclidean distance of x i from X, N k (X) neighborhood of X that contains exactly k neighbors. To the best of our knowledge, so far there has been no rational method to determine the value of k, which may be one drawback of k-nearest neighborhood method. Note that the computational cost will increase with the value of k. So, in this study, we set k to 2.

3.4 External archive

The main objective of the external archive is to keep a historical record of the nondominated individuals found along the search process. The external repository consists of two main components: the archive controller and the crowding-distance metric.

  • The archive controller: Inspired by [15], we introduce the archive controller into our approach. The function of the archive controller is to decide whether the nondominated solutions found in the new population should be added to the archive or not. The decision-making process is similar to that in [15]. After comparing all the nondominated solutions in the new population with respect to the external population, if the external population has exceeded its maximum allowable capacity, then the crowding-distance computation procedure is invoked.

  • Crowding-distance metric: In our approach, we adopt the crowding-distance metric proposed by Deb et al. [11] to truncate the external archive when the external population has exceeded its maximum allowable capacity. For more details, the readers can refer to [11].

3.5 Mutation operator

There is merely mutation operator in MOPEO. Therefore, the mutation plays a key role in MOPEO search that generates new solutions through adding or removing genes at the current solutions, i.e., chromosomes. Till now, there have been many mutation operations proposed, such as Gaussian mutation, Cauchy mutation and nonuniform mutation. The mechanisms of Gaussian and Cauchy mutation operations have been studied by Yao et al. [16]. They pointed out that Cauchy mutation is better at coarse-grained search while Gaussian mutation is better at fine-grained search. The non-uniform mutation introduced by Janikow and Michalewicz [17] is designed for fine-tuning the solutions. At early generations, the mutation range is relatively large, while at the latter generations, it is tightened for local refinement. Thus, the non-uniform mutation operation combines the advantages of coarse-grained search and fine-grained search.

In this study, we use the non-uniform mutation as mutation operator. Suppose the kth variable is chosen to mutate, the value of kth variable after mutation is calculated as follows [17]:

$$x_k^{t+1}= \left\{ {\begin{array}{*{20}l} {x_k^t+\Delta{(t,u_k-x_k^t)}, \quad a > 0.5}\\ {x_k^t-\Delta{(t,x_k^t-l_k)}, \quad a\leq{0.5}}\\ \end{array} } \right. $$
(2)

where k = 1, ..., n, n is number of decision variables, a random number in [0,1], l k and u k lower and upper bounds of the kth variable, respectively. The function Δ(t,y) is as follows:

$$ \Delta{(t,y)}=y {r^{(1-t/T) {b}}}$$
(3)

where r is random number with range of [0,1], T maximal generation number, and b system parameter determining the degree of non-uniformity. Thus, there is only one adjustable parameter b in the non-uniform mutation. The function Δ(t,y) returns a value in the range [0,y] such that the probability of Δ(t,y) being close to 0 increases as t increases (t is generation number). This property causes this operator to search the space uniformly initially (when t is small), and very locally at the later stages. The value of b affects the performance of non-uniform mutation. As we can see from Eq. 3, the searching step size, i.e., Δ(t,y), increases with the decreasing value of b. So the value of b can adjust the searching step size of non-uniform mutation. As can be seen from the aforementioned, the non-uniform mutation is helpful for our approach to perform exploration and exploitation.

4 Experiments and test results

4.1 Test problems

We choose five problems out of six test problems proposed by Zitzler et al. [18] and call them ZDT1, ZDT2, ZDT3, ZDT4, and ZDT6. All problems have two objective functions. None of these problems have any constraint. We describe these problems in Table 1. The table also shows the number of variables, their bounds, the Pareto-optimal solutions, and the nature of Pareto-optimal front for each problem.

Table 1 Test problems in this study

Deb et al. [11] have compared the performance of NSGA-II with PAES and SPEA and obtained their experimental results in [11]. The experimental results in [11] demonstrated that, in most problems, NSGA-II is able to find much better spread of solutions and better convergence near the true Pareto-optimal front compared to PAES and SPEA. In this paper, we apply the MOPEO to solve five difficult test problems in Table 1 and compare our simulation results with those obtained in [11] under the same conditions. To be fair, we adopt the identical parameter settings as suggested in [11]. For four test problems (ZDT1, ZDT2, ZDT3 and ZDT6), the adjustable parameter b in the non-uniform mutation is set to 0.9, while b is set to 0.1 for ZDT4. In this study, all the algorithms developed were encoded in the floating point representation. The fitness function evaluations for all the algorithms is 25,000. The maximum population size of the external archive is 100. Additionally, ten independent runs were carried out. The source codes of all experiments were coded in JAVA.

4.2 Performance measures

In this article, we used the two performance metrics proposed by Deb et al. [11] to assess the performance of our approach. For more detail, the readers can refer to [11]. The first metric \(\Upsilon\) measures the extent of convergence to a known set of Pareto-optimal solutions. In all simulations, we address the average \(\bar{\Upsilon}\) and variance \(\sigma_\Upsilon\) of this metric obtained in ten independent runs. Deb et al. [11] has pointed out that even when all solutions converge to the Pareto-optimal front, the convergence metric does not have a value of zero. The metric will yield zero only when each obtained solution lies exactly on each of the chosen solutions.

The second metric Δ measures the extent of spread of the obtained nondominated solutions. It is desirable to get a set of solutions that spans the entire Pareto-optimal region. The second metric Δ can be calculated as follows [11]:

$$\Delta=\frac{d_f+d_l+\sum_{i=1}^{N-1}{|d_i-{\bar{d}}|}} {d_f+d_l+(N-1){\bar{d}}}$$
(4)

where, d f and d l are Euclidean distances between the extreme solutions and the boundary solutions of the obtained nondominated set, d i Euclidean distance between consecutive solutions in the obtained nondominated set of solutions, and \(\bar{d}\) is the average of all distances d i (i = 1, 2, ..., N−1), assuming that there are N solutions on the best nondominated front. Note that a good distribution would make all distances d i equal to \(\bar{d}\) and would make d f  = d l  = 0 (with existence of extreme solutions in the nondominated set). Consequently, for the most widely and uniformly spreadout set of nondominated solutions, Δ would be zero.

4.3 Discussion of the results

Table 2 shows the mean and variance of the convergence metric \(\Upsilon\) obtained using four algorithms, i.e., MOPEO, NSGA-II (real-coded), SPEA and PAES. Here, all the experimental results of NSGA-II (real-coded), SPEA and PAES come from [11].

Table 2 Mean (first rows) and variance (second rows) of the convergence metric \(\Upsilon\)

As can be observed from Table 2, MOPEO is able to converge better than the other algorithms in three problems, i.e., ZDT1, ZDT3 and ZDT6 and it is the next-best algorithm for problem ZDT2. For problem ZDT4, MOPEO performs worse than NSGA-II and PAES in terms of the convergence to the Pareto-optimal front. In all cases with MOPEO, the variance of the convergence metric in ten runs is very small except in ZDT4.

Table 3 shows the mean and variance of the diversity metric Δ obtained using all the algorithms. On problems ZDT3 and ZDT6, MOPEO is able to find a better spread of solutions than any other algorithm. MOPEO is the next-best algorithm for problems ZDT1 and ZDT2 while NSGA-II is the best algorithm. For problem ZDT4, MOPEO performs worse than any other algorithm with respect to diversity. In all cases with MOPEO, the variance of the diversity metric in ten runs is also small.

Table 3 Mean (first rows) and variance (second rows) of the diversity metric Δ

In order to compare the running times of MOPEO with those of other three algorithms, i.e., NSGA-II, SPEA and PAES, we also show the mean and variance of running times of each algorithm in ten runs in Table 4. To avoid any bias or misinterpretation when implementing each of the three other approaches, we adopted the public-domain versions of NSGA-II, PAES and SPEA. Note that all the algorithms were run on the same hardware (i.e., Intel Pentium M with 900 MHz CPU and 256 M memory) and software (i.e., JAVA) platform. The fitness function evaluations for all the algorithms is 25,000. In order to compare the running times of each algorithm fairly, we use the crowding-distance metric as the archive truncation method for MOPEO, SPEA and PAES, and as the diversity preservation mechanism for NSGA-II. From Table 4, we can see that, in terms of running times, MOPEO is competitive with other three algorithms on all the problems.

Table 4 Mean (first rows) and variance (second rows) of the running times (in milliseconds)

For illustration, we show one of ten runs of MOPEO on the five test problems in Fig. 2. From Fig. 2, we can see that, with respect to convergence and diversity of the obtained set of nondominated solutions, MOPEO performs well in all test problems except in ZDT4. Note that the problem ZDT4 has 219 different local Pareto-optimal fronts in the search space, of which only one corresponds to the global Pareto-optimal front. As can be seen from Fig. 2, MOPEO got stuck into the local Pareto-optimal fronts of problem ZDT4.

Fig. 2
figure 2

Nondominated solutions with MOPEO on ZDT1, ZDT2, ZDT3, ZDT4 and ZDT6, respectively

4.4 Advantages of proposed approach

From the above analysis, it can be seen that our approach has the following advantages:

  • Our approach is capable of finding multiple Pareto-optimal solutions in one single simulation run due to its EO and population mechanism.

  • There exists only one adjustable parameter, i.e., the system parameter b in the non-uniform mutation. Note that b can be tuned easily. This makes our approach easier in real applications than other state-of-the-art methods.

  • Only one operator, i.e., mutation operator, exists in our approach, which makes our approach simple and convenient.

  • Similar to evolutionary algorithms, MOPEO is less susceptible to the shape or continuity of the Pareto front.

  • The proposed approach has a good performance in both aspects of convergence and distribution of solutions.

5 Conclusion and future work

In this paper, we present a novel Pareto-based algorithm, called MOPEO, to extend EO to handle multiobjective problems. Our approach has many advantages such as less adjustable parameters, only mutation operation, easily implementation. Our approach is validated to be competitive with respect to three algorithms representative of the state-of-the-art in the area. In the future work, it is desirable to improve our approach to solving those problems with multiple local Pareto-optimal fronts. In addition, we will extend MOPEO to solve constrained or discrete multiobjective optimization problems, e.g., multiobjective knapsack problems. The future work also includes the applications of MOPEO to solving those complex engineering optimization problems in the real world, e.g., multiobjective portfolio optimization problems.