Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Motivation

By far most studies in the field of evolutionary multiobjective optimization (EMO) are concerned with the following set problem: find a set of solutions that as a whole represents a good approximation of the Pareto-optimal set. To this end, the original multiobjective problem consisting of:

  • The decision space X

  • The objective space \(Z = {\mathbb{R}}^{n}\)

  • A vector function f = (f 1, f 2, , f n ) comprising n objective functions \({f}_{i} : X \rightarrow \mathbb{R}\), which are without loss of generality to be minimized

  • A relation ≤ on Z, which induces a preference relation \(\preccurlyeq \) on X with \(a \preccurlyeq b :\Leftrightarrow f(a) \leq f(b)\) for a, bX

is usually transformed into a single-objective set problem (Zitzler et al. 2008).

The search space Ψ of the resulting set problem includes all possible Pareto set approximations,Footnote 1 i.e., Ψ contains all multisets over X. The preference relation \(\preccurlyeq \) can be used to define a corresponding set preference relation ≼ on Ψ where

$$A \preccurlyeq B :\Leftrightarrow \forall b \in B\,\exists a \in A :\ a \preccurlyeq b$$
(1)

for all Pareto set approximations A, BΨ. In the following, we will assume that weak Pareto dominance is the underlying preference relation, i.e., \(a \preccurlyeq b :\Leftrightarrow f(a) \leq f(b)\) (cf. Zitzler et al. 2008).Footnote 2

A key question when tackling such a set problem is how to define the optimization criterion. Many multiobjective evolutionary algorithms (MOEAs) implement a combination of Pareto dominance on sets and a diversity measure based on Euclidean distance in the objective space, e.g., NSGA-II (Deb et al. 2000) and SPEA2 (Zitzler et al. 2002). While these methods have been successfully employed in various biobjective optimization scenarios, they appear to have difficulties when the number of objectives increases (Wagner et al. 2007). As a consequence, researchers have tried to develop alternative concepts, and a recent trend is to use set quality measures, also denoted as quality indicators, for search – so far, they have mainly been used for performance assessment. Of particular interest in this context is the hypervolume indicator (Zitzler and Thiele 1998a, 1999) as it is the only quality indicator known to be fully sensitive to Pareto dominance, i.e., whenever a set of solutions dominates another set, it has a higher hypervolume indicator value than the second set. This property is especially desirable when many objective functions are involved.

Several hypervolume-based MOEAs have been proposed meanwhile (e.g., Emmerich et al. 2005; Igel et al. 2007; Brockhoff and Zitzler 2007), but their main drawback is their extreme computational overhead. Although there have been recent studies presenting improved algorithms for hypervolume calculation, currently high-dimensional problems with six or more objectives are infeasible for these MOEAs. Therefore, the question is whether and how fast hypervolume-based search algorithms can be designed that exploit the advantages of the hypervolume indicator and at the same time are scalable with respect to the number of objectives.

2 Related Work

The hypervolume indicator was originally proposed and employed in Zitzler and Thiele (1998b, 1999) to compare quantitatively the outcomes of different MOEAs. In these two first publications, the indicator was denoted as “size of the space covered”, and later also other terms such as “hyperarea metric” (Van Veldhuizen 1999), “S-metric” (Zitzler 1999), “hypervolume indicator” (Zitzler et al. 2003), and “hypervolume measure” (Beume et al. 2007b) were used. Besides the names, there are also different definitions available, based on polytopes (Zitzler and Thiele 1999), the attainment function (Zitzler et al. 2007), or the Lebesgue measure (Laumanns et al. 1999; Knowles 2002; Fleischer 2003).

Knowles (2002) and Knowles and Corne (2003) were the first to propose the integration of the hypervolume indicator into the optimization process. In particular, they described a strategy to maintain a separate, bounded archive of nondominated solutions based on the hypervolume indicator. Huband et al. (2003) presented an MOEA which includes a modified SPEA2 environmental selection procedure where a hypervolume-related measure replaces the original density estimation technique. In Zitzler and Künzli (2004), the binary hypervolume indicator was used to compare individuals and to assign corresponding fitness values within a general indicator-based evolutionary algorithm (IBEA). The first MOEA tailored specifically to the hypervolume indicator was described in Emmerich et al. (2005); it combines nondominated sorting with the hypervolume indicator and considers one offspring per generation (steady state). Similar fitness assignment strategies were later adopted in Zitzler et al. (2007) and Igel et al. (2007), and also other search algorithms were proposed where the hypervolume indicator is partially used for search guidance (Nicolini 2005; Mostaghim et al. 2007). Moreover, specific aspects like hypervolume-based environmental selection (Bradstreet et al. 2006), cf. Sect. 3.3, and explicit gradient determination for hypervolume landscapes (Emmerich et al. 2007) have been investigated recently.

The major drawback of the hypervolume indicator is its high computation effort; all known algorithms have a worst-case runtime complexity that is exponential in the number of objectives, more specifically \(\mathcal{O}({N}^{n-1})\) where N is the number of solutions considered (Knowles 2002; While et al. 2006). A different approach was presented by Fleischer (2003) who mistakenly claimed a polynomial worst-case runtime complexity – While (2005) showed that it is exponential in n as well. Recently, advanced algorithms for hypervolume calculation have been proposed, a dimension-sweep method (Fonseca et al. 2006) with a worst-case runtime complexity of \(\mathcal{O}({N}^{n-2}\log N)\), and a specialized algorithm related to the Klee measure problem (Beume and Rudolph, 2006) the runtime of which is in the worst case of order \(\mathcal{O}(N\log N + {N}^{n/2})\). Furthermore, Yang and Ding (2007) described an algorithm for which they claim a worst-case runtime complexity of \(\mathcal{O}({(n/2)}^{N})\). The fact that there is no exact polynomial algorithm available gave rise to the hypothesis that this problem in general is hard to solve, although the tightest known lower bound is of order Ω(NlogN) (Beume et al. 2007a). New results substantiate this hypothesis: Bringmann and Friedrich (2008) have proven that the problem of computing the hypervolume is #P-complete, i.e., it is expected that no polynomial algorithm exists since this would imply NP = P.

The issue of speeding up the hypervolume indicator has been addressed in different ways: by automatically reducing the number of objectives (Brockhoff and Zitzler 2007) and by approximating the indicator values using Monte Carlo simulation (Everson et al. 2002; Bader et al. 2008; Bringmann and Friedrich 2008). Everson et al. (2002) used a basic Monte Carlo technique for performance assessment in order to estimate the values of the binary hypervolume indicator (Wagner et al. 2007); with their approach the error ratio is not polynomially bounded. In contrast, the scheme presented in Bringmann and Friedrich (2008) is a fully polynomial randomized approximation scheme where the error ratio is polynomial in the input size. Another study (Bader et al. 2008) – a precursor study for the present paper – employed Monte Carlo sampling for fast hypervolume-based search. The main idea is to estimate – by means of Monte Carlo simulation – the ranking of the individuals that is induced by the hypervolume indicator and not to determine the exact indicator values. This paper proposes an advanced method called HypE (Hypervolume Estimation Algorithm for Multiobjective Optimization) that is based on the same idea, but uses a novel fitness assignment scheme for both mating and environmental selection, that can be effectively approximated.

As we will show in the following, the proposed search algorithm can be easily tuned regarding the available computing resources and the number of objectives involved. Thereby, it opens a new perspective on how to treat many-objective problems, and the presented concepts may also be helpful for other types of quality indicators to be integrated in the optimization process.

3 HypE: Hypervolume Estimation Algorithm for Multiobjective Optimization

When considering the hypervolume indicator as the objective function of the underlying set problem, the main question is how to make use of this measure within a multiobjective optimizer to guide the search. In the context of a MOEA, this refers to selection and one can distinguish two situations:

  1. 1.

    The selection of solutions to be varied (mating selection)

  2. 2.

    The selection of solutions to be kept in the population (environmental selection)

In the following, we outline a new algorithm based on the hypervolume indicator called HypE. Thereafter, the two selection steps mentioned above as realized in HypE are presented in detail.

3.1 Algorithm

HypE belongs to the class of simple indicator-based evolutionary algorithm, as for instance discussed in Zitzler et al. (2007). As outlined in Algorithm 1, HypE reflects a standard evolutionary algorithm that consists of the successive application of mating selection, variation, and environmental selection. As to mating selection, binary tournament selection is proposed here, although any other selection scheme could be used as well, where the tournament selection is based on the fitness proposed in Sect. 3.2. The procedure { variation} encapsulates the application of mutation and recombination operators to generate N offspring. Finally, environmental selection aims at selecting the most promising N solutions from the multiset-union of parent population and offspring; more precisely, it creates a new population by carrying out the following two steps:

  1. 1.

    First, the union of parents and offspring is divided into disjoint partitions using the principle of nondominated sorting (Goldberg 1989; Deb et al. 2000), also known as dominance depth. Starting with the lowest dominance depth level, the partitions are moved one by one to the new population as long as the first partition is reached that cannot be transferred completely. This corresponds to the scheme used in most hypervolume-based multiobjective optimizers (Emmerich et al. 2005; Igel et al. 2007; Brockhoff and Zitzler 2007).

  2. 2.

    The partition that only fits partially into the new population is then processed using the novel fitness scheme presented in Sect. 3.3. In each step, the fitness values for the partition under consideration are computed and the individual with the worst fitness is removed – if multiple individuals share the same minimal fitness, then one of them is selected uniformly at random. This procedure is repeated until the partition has been reduced to the desired size, i.e., until it fits into the remaining slots left in the new population.

The scheme of first applying non-dominated sorting is similar to other algorithms (e.g., Igel et al. 2007; Emmerich et al. 2005). The differences are: (1) the fitness assignment scheme for mating, (2) the one for environmental selection, and (3) the method how the fitness values are determined. The estimation of the fitness values by means of Monte Carlo sampling is discussed in Sect. 3.4.

3.2 Basic Scheme for Mating Selection

To begin with, we formally define the hypervolume indicator as a basis for the following discussions. Different definitions can be found in the literature, and we here use the one from Zitzler et al. (2008) which draws upon the Lebesgue measure as proposed in Laumanns et al. (1999) and considers a reference set of objective vectors.

Definition 1.

Let AΨ be a Pareto set approximation and RZ be a reference set of mutually nondominating objective vectors. Then the hypervolume indicator I H can be defined as

$${I}_{H}(A,R) := \lambda (H(A,R)),$$
(2)

where

$$H(A,R) :=\{ z \in Z\,;\,\exists a \in A\,\exists r \in R : f(a) \leq z \leq r\}$$
(3)

and λ is the Lebesgue measure with \(\lambda (H(A,R)) ={ \int }_{{\mathbb{R}}^{n}}{\mathbf{1}}_{H(A,R)}(z)dz\) and 1 H(A, R) being the characteristic function of H(A, R).

The set H(A, R) denotes the set of objective vectors that are enclosed by the front f(A) given by A and the reference set R. It can be further split into partitions H(S, A, R), each associated with a specific subset SA:

$$H(S,A,R) := \Big{[}{\bigcap }_{s\in S}H(\{s\},R)\Big{]} \setminus \Big{[}{\bigcup }_{a\in A\setminus S}H(\{a\},R)\Big{]}.$$
(4)

The set H(S, A, R) ⊆ Z represents the portion of the objective space that is jointly weakly dominated by the solutions in S and not weakly dominated by any other solution in A. The partitions H(S, A, R) are disjoint and the union of all partitions is H(A, R) which is illustrated in Fig. 1.

Fig. 1
figure 1_3

Illustration of the notions of H(A, R) and H(S, A, R) in the objective space for a Pareto set approximation A = { a, b, c, d} and reference set R = { r}

In practice, it is infeasible to determine all distinct H(S, A, R) due to combinatorial explosion. Instead, we will consider a more compact splitting of the dominated objective space that refers to single solutions:

$${ H}_{i}(a,A,R) :=\mathop{ \bigcup}\limits_{\begin{array}{l} S \subseteq A \\ a \in S \\ \vert S\vert = i \end{array} }H(S,A,R). $$
(5)

According to this definition, H i (a, A, R) stands for the portion of the objective space that is jointly and solely weakly dominated by a and any i − 1 further solutions from A, see Fig. 2. Note that the sets H 1(a, A, R), H 2(a, A, R), , H | A |(a, A, R) are disjoint for a given aA while the sets H i (a, A, R) and H i (b, A, R) may be overlapping for fixed i and different solutions a, bA. This slightly different notion has reduced the number of subspaces to be considered from 2| A | for H(S, A, R) to | A | 2 for H i (a, A, R).

Fig. 2
figure 2_3

Illustration of the notions of H(A, R) and H i (a, A, R) in the objective space for a Pareto set approximation A = { a, b, c, d} and reference set R = { r}

Now, given an arbitrary population PΨ one obtains for each solution a contained in P a vector (λ(H 1(a, P, R)), λ(H 2(a, P, R)), , λ(H | P |(a, P, R))) of hypervolume contributions. These vectors can be used to assign fitness values to solutions; while most hypervolume-based search algorithms only take the first components, i.e., λ(H 1(a, P, R)), into account, we here propose the following scheme to aggregate the hypervolume contributions into a single scalar value.

Definition 2.

Let AΨ and RZ. Then the function I h with

$${I}_{h}(a,A,R) :=\mathop{ \sum}\limits_{i=1}^{\vert A\vert }\frac{1} {i} \lambda ({H}_{i}(a,A,R))$$
(6)

gives for each solution aA the hypervolume that can be attributed to a with regard to the overall hypervolume I H (A, R).

The motivation behind this definition is simple: the hypervolume contribution of each partition H(S, A, R) is shared equally among the dominating solutions sS. That means the portion of Z solely weakly dominated by a specific solution a is fully attributed to a, the portion of Z that a weakly dominates together with another solution b is attributed half to a and so forth – the principle is illustrated in Fig. 3. Thereby, the overall hypervolume is distributed among the distinct solutions according to their hypervolume contributions as the following theorem shows (the proof can be found in the appendix). Note that this scheme does not require that the solutions of the considered Pareto set approximation A are mutually non-dominating; it applies to nondominated and dominated solutions alike.

Fig. 3
figure 3_3

Illustration of the basic fitness assignment scheme. The fitness of two individuals a and c of a Pareto set approximation A = { a, b, c, d} is set to F a = I h (a, A, R) and F b = I h (b, A, R) respectively (diagonally hatched areas)

Next, we will extend and generalize the fitness assignment scheme with regard to the environmental selection phase.

3.3 Extended Scheme for Environmental Selection

In EMO, environmental selection is mostly carried out by first merging parents and offspring and then truncating the resulting union by choosing the subset that represents the best Pareto set approximation. The number k of solutions to be achieved usually equals the population size, and therefore the exact computation of the best subset is computationally infeasible. Instead, the optimal subset is approximated in terms of a greedy heuristic (Zitzler and Künzli 2004; Brockhoff and Zitzler 2007): all solutions are evaluated with respect to their usefulness and the least important solution is removed; this process is repeated until k solutions have been removed.

The key issue with respect to the above greedy strategy is how to evaluate the usefulness of a solution. The scheme presented in Definition 2 has the drawback that portions of the objective space are taken into account that for sure will not change. Suppose, for instance, a population with four solutions as shown in Fig. 4; when two solutions need to be removed (k = 2), then the subspaces H({a, b, c}, P, R), H({b, c, d}, P, R), and H({a, b, c, d}, P, R) remain weakly dominated independently of which solutions are deleted. This observation led to the idea of considering the expected loss in hypervolume that can be attributed to a particular solution when exactly k solutions are removed. In detail, we consider for each aP the average hypervolume loss over all subsets SP that contain a and k − 1 further solutions; this value can be easily computed by slightly extending the scheme from Definition 2 as follows.

Definition 3.

Let AΨ, RZ, and k ∈ { 0, 1, , | A | }. Then the function \({I}_{h}^{k}\) with

$${I}_{h}^{k}(a,A,R) :=\mathop{ \sum }\limits_{i=1}^{k}\frac{{\alpha }_{i}} {i} \lambda ({H}_{i}(a,A,R)),$$
(7)

where

$${\alpha }_{i} :=\mathop{ \prod}\limits_{j=1}^{i-1} \frac{k - j} {\vert A\vert - j}$$
(8)

gives for each solution aA the expected hypervolume loss that can be attributed to a when a and k − 1 uniformly randomly chosen solutions from A are removed from A.

The correctness of (7) can be proved by mathematical induction (Bader and Zitzler 2008). In the following example we illustrates the calculation of α i to put the idea of expected loss across:

Example 1.

Out of five individuals A = { a, b, c, d, e}, three have to be removed and we want to know \({I}_{h}^{3}(a,A,R)\) (see Fig. 5). To this end, the expected shares of H i (a, A, R) that are lost when removing a – represented by α i – have to be determined. The first coefficient is α1 = 1, because H 1(a, A, R) is lost for sure, since the space is only dominated by a. Calculating the probability of losing the partitions dominated by a and a second individual H 2(a, A, R) is a little bit more involved. Without loss of generality we consider the space dominated only by a and c. In addition to a, there are two individuals to be removed out of the remaining four individuals. Since we assume, by approximation every removal is equally probable, the chance of removing c and thereby losing the partition H({a, c}, A, R) is α2 = 1 ∕ 2. Finally, for α3 three points including a have to be taken off. The probability of removing a and any other individual was just derived to be 1 ∕ 2. Out of three remaining individuals, the chance to remove one particular, say e, is 1 ∕ 3. This gives the coefficient α3 = (1 ∕ 2)(1 ∕ 3). For reasons of symmetry, these calculations hold for any other partition dominated by a and i − 1 points. Therefore, the expected loss of H i (a, A, R) is α i λ(H i (a, A, R)). Since i points share the partition H i (a, A, R), the expected loss is additionally multiplied by 1 ∕ i.

Fig. 4
figure 4_3

The figure shows for A = { a, b, c, d} and R = { r} (1) which portion of the objective space remains dominated if any two solutions are removed from A (shaded area), and (2) the probabilities α that a particular area that can be attributed to aA is lost if a is removed from A together with any other solution in A

Fig. 5
figure 5_3

Illustrates the calculation of α i when removing a and two other individuals from the set A = { a, b, c, d, e}. Each subset of three individuals that contains a is considered equally likely; therefore, the chance of losing a and c is 3 ∕ 6 and the probability of losing a, c, and e is (1 ∕ 2)(1 ∕ 3)

Notice that \({I}_{h}^{1}(a,A,R) = \lambda ({H}_{1}(a,A,R))\) and \({I}_{h}^{\vert A\vert }(a,A,R) = {I}_{h}(a,A,R)\), i.e., this modified scheme can be regarded as a generalization of the scheme presented in Definition 2 and the commonly used fitness assignment strategy for hypervolume-based search (Knowles and Corne 2003; Emmerich et al. 2005; Igel et al. 2007; Bader et al. 2008).

The fitness values \({I}_{h}^{k}(a,A,R)\) can be calculated on the fly for all individuals aA by a slightly extended version of the “hypervolume by slicing objectives” (Zitzler 2001; Knowles 2002; While et al. 2006) algorithm, which traverse recursively one objective after another. It differs from existing methods in that it allows (1) to consider a set R of reference points and (5) to compute all fitness values, e.g., the \({I}_{h}^{1}(a,P,R)\) values for k = 1, in parallel for any number of objectives instead of subsequently as in Beume et al. (2007b). Basically, the modification concerns also considering solutions, which are dominated according to a particular scanline. A detailed description of the extended algorithm can be found in Bader and Zitzler (2008).

Unfortunately, the worst case complexity of the algorithm is \(\mathcal{O}(\vert A{\vert }^{n})\) which renders the exact calculation inapplicable for problems with more than about five objectives. However, in the context of randomized search heuristics one may argue that the exact fitness values are not crucial and approximated values may be sufficient. These considerations lead to the idea of estimating the fitness values by Monte Carlo sampling, whose basic principle is described in the following section.

3.4 Estimating the Fitness Values Using Monte Carlo Sampling

To approximate the fitness values according to Definitions 2 and 3, we need to estimate the Lebesgue measures of the domains H i (a, P, R) where PΨ is the population. Since these domains are all integrable, their Lebesgue measure can be approximated by means of Monte Carlo simulation.

For this purpose, a sampling space SZ has to be defined with the following properties: (1) the hypervolume of S can easily be computed, (2) samples from the space S can be generated fast, and (3) S is a superset of the domains H i (a, P, R) the hypervolumes of which one would like to approximate. The latter condition is met by setting S = H(P, R), but since it is hard both to calculate the Lebesgue measure of this sampling space and to draw samples from it, we propose using the axis-aligned minimum bounding box containing the H i (a, P, R) subspaces instead, i.e.:

$$S :=\{ ({z}_{1},\ldots,{z}_{n}) \in Z\,\vert \,\forall 1 \leq i \leq n : {l}_{i} \leq {z}_{i} \leq {u}_{i}\},$$
(9)

where

$$\begin{array}{lll} {l}_{i} & :=&{\min }_{a\in P}{f}_{i}(a)\\ {u}_{ i}& :=&{\max }_{({r}_{1},\ldots,{r}_{n})\in R}{r}_{i} \end{array}$$
(10)

for 1 ≤ in. Hence, the volume V of the sampling space S is given by \(V ={ \prod }_{i=1}^{n}\max \{0,{u}_{i} - {l}_{i}\}\).

Now given S, sampling is carried out by selecting M objective vectors s 1, , s M from S uniformly at random. For each s j it is checked whether it lies in any partition H i (a, P, R) for 1 ≤ ik and aP. This can be determined in two steps: first, it is verified that s j is “below” the reference set R, i.e., it exists rR that is dominated by s j ; second, it is verified that the multiset A of those population members dominating s j is not empty. If both conditions are fulfilled, then we know that – given A – the sampling point s j lies in all partitions H i (a, P, R) where i = | A | and aA. This situation will be denoted as a hit regarding the ith partition of a. If any of the above two conditions is not fulfilled, then we call s j a miss. Let \({X}_{j}^{(i,a)}\) denote the corresponding random variable that is equal to 1 in case of a hit of s j regarding the ith partition of a and 0 otherwise.

Based on the M sampling points, we obtain an estimate for λ(H i (a, P, R)) by simply counting the number of hits and multiplying the hit ratio with the volume of the sampling box:

$$\hat{\lambda }\big{(}{H}_{i}(a,P,R)\big{)} = \frac{{\sum }_{j=1}^{M}{X}_{j}^{(i,a))}} {M} V.$$
(11)

This value approaches the exact value λ(H i (a, P, R)) with increasing M by the law of large numbers. Due to the linearity of the expectation operator, the fitness scheme according to (7) can be approximated by replacing the Lebesgue measure with the respective estimates given by (11):

$$\hat{{I}}_{h}^{k}(a,P,R) =\mathop{ \sum }\limits_{i=1}^{k}\frac{{\alpha }_{i}} {i} \left (\frac{{\sum }_{j=1}^{M}{X}_{j}^{(i,a))}} {M} V \right ).$$
(12)

Note that the partitions H i (a, P, R) with i > k do not need to be considered for the fitness calculation as they do not contribute to the \({I}_{h}^{k}\) values that we would like to estimate, cf. Definition 3.

4 Experiments

4.1 Experimental Setup

HypE is implemented within the PISA framework (Bleuler et al. 2003) and tested in two versions: the first, denoted by HypE, uses fitness-based mating selection as described in Sect. 3.2, while the second, HypE*, employs a uniform mating selection scheme where all individuals have the same probability of being chosen for reproduction. Unless stated otherwise, for sampling the number of sampling points is fixed to M = 10, 000 and kept constant during a run.

On the one hand, HypE and HypE* are compared to three popular MOEAs, namely NSGA-II (Deb et al. 2000), SPEA2 (Zitzler et al. 2002), and IBEA (in combination with the ε-indicator) (Zitzler and Künzli 2004). Since these algorithms are not designed to optimize the hypervolume, it cannot be expected that they perform particularly well when measuring the quality of the approximation in terms of the hypervolume indicator. Nevertheless, they serve as an important reference as they are considerably faster than hypervolume-based search algorithms and therefore can execute a substantially larger number of generations when keeping the available computation time fixed. On the other hand, we include the sampling-based optimizer proposed in Bader et al. (2008), here denoted as SHV (sampling-based hypervolume-oriented algorithm); finally, to study the influence of the nondominated sorting we also include a simple HypE variant named RS (random selection) where all individuals of the same dominance depth level are assigned the same constant fitness value. Thereby, the selection pressure is only maintained by the nondominated sorting carried out during the environmental selection phase.

In this paper, we focus on many-objective problems for which exact hypervolume-based methods (e.g., Emmerich et al. 2005; Igel et al. 2007) are not applicable. Therefore, we did not include these algorithms in the experiments. However, the interested reader is referred to Bader and Zitzler (2008) where HypE is compared to an exact hypervolume algorithm.

As basis for the comparisons, the DTLZ (Deb et al. 2005), and the WFG (Huband et al. 2006) test problem suites are considered since they allow the number of objectives to be scaled arbitrarily – here, ranging from 2 to 50 objectives. For the DTLZ problem, the number of decision variables is set to 300, while for the WFG problems individual values are used.Footnote 3

The individuals are represented by real vectors, where a polynomial distribution is used for mutation and the SBX-20 operator is used for recombination (Deb 2001). The recombination and mutation probabilities are set according to Deb et al. (2005).

For each benchmark function, 30 runs are carried out per algorithm using a population size of N = 50. Either the maximum number of generations was set to g max = 200 (results shown in Table 1) or the runtime was fixed to 30 min for each run (results shown in Fig. 6). For each run, the hypervolume of the last population is determined, where for less than six objectives these are calculated exactly and otherwise approximated by Monte Carlo sampling. For each algorithm A i , the hypervolume values are then subsumed under the performance score P(A i ), which represents the number of other algorithms that achieved significantly higher hypervolume values on the particular test case. The test for significance is done using Kruskal–Wallis and the Conover Inman post hoc tests (Conover 1999). For a full description of the performance score, please see Zamora and Burguete (2008).

Table 1 Comparison of HypE to different EAs for the hypervolume indicator, where n denotes the number of objectives. The first number represents the performance score P(A i ), which stands for the number of contenders significantly dominating algorithm A i . The number in brackets denote the hypervolume value (to be maximized), normalized to the minimum and maximum value observed on the testproblem
Fig. 6
figure 6_3figure 6_3

Hypervolume process over 30 min of HypE and SHV for 100; 1,000; 10,000 and 100,000 samples each, denoted by the suffices 100, 1k, 10k and 100k respectively. The test problem is WFG9 for three objectives and NSGA-II, SPEA2, and IBEA are shown as reference. The numbers at the right border of the figures indicate the total number of generations reached after 30 min. The results are split in two figures with identical axis for the sake of clarity

4.2 Results

Table 1 shows the performance score and mean hypervolume of the different algorithms on six test problems. In 18 instances, HypE is better than HypE*, while vice versa HypE* is better than HypE only in four cases. HypE reaches the best performance score overall. Summing up all performance scores, HypE yields the best total (33), followed by HypE* (55), IBEA (55) and SHV, the method proposed in Bader et al. (2008) (97). SPEA2 and NSGA-II reach almost the same score (136 and 140 respectively), clearly outperforming random selection (217). For five out of six testproblems HypE obtains better hypervolume values than SHV. On DTLZ7 however, SHV as well as IBEA outperform HypE. This might be due to the discontinuous shape of the DTLZ7 testfunction, for which the advanced fitness scheme does not give an advantage.

The better Pareto set approximations of HypE come at the expense of longer execution time, e.g., in comparison to SPEA2 or NSGA-II. We therefore investigate, whether the fast NSGA-II and SPEA2 will not overtake HypE given a constant amount of time. Figure 6 shows the hypervolume of the Pareto set approximations over time for HypE using the exact fitness values as well as the estimated values for different samples sizes M.

Even though SPEA2, NSGA-II and even IBEA are able to process twice as many generations as the exact HypE, they do not reach its hypervolume. In the three dimensional example used, HypE can be run sufficiently fast without approximating the fitness values. Nevertheless, the sampled version is used as well to show the dependency of the execution time and quality on the number of samples M. Via M, the execution time of HypE can be traded off against the quality of the Pareto set approximation. The fewer samples are used, the more the behavior of HypE resembles random selection. On the other hand by increasing M, the quality of exact calculation can be achieved, increasing the execution time, though. For example, with M = 1, 000, HypE is able to carry out nearly the same number of generations as SPEA2 or NSGA-II, but the Pareto set is just as good as when 100, 000 samples are used, needing only a fifteenth the number of generations. In the example given, M = 10, 000 represents the best compromise, but the number of samples should be increased in two cases: (1) when the fitness evaluation take more time, and (2) when more generations are used. The former will affect the faster algorithm much more and increasing the number of samples will influence the execution time much less. (ii) More generations are used. In the latter case, HypE using more samples might overtake the faster versions with fewer samples, since those are more vulnerable to stagnation.

In this three objective scenario, SHV can compete with HypE. This is mainly because the sampling boxes that SHV relies on are tight for small number of objectives; as this number increases, however, experiments not shown here indicate that the quality of SHV decreases in relation to HypE.

5 Conclusion

This paper proposes HypE, a novel hypervolume-based evolutionary algorithm. Both its environmental and mating selection step rely on a new fitness assignment scheme based on the Lebesgue measure, where the values can be both exactly calculated and estimated by means of sampling. In contrast to other hypervolume based algorithms, HypE is thus applicable to higher dimensional problems.

The performance of HypE was compared to other algorithms against the hypervolume indicator of the Pareto set approximation. In particular, the algorithms were tested on test problems of the WFG and DTLZ test suite. Simulations results show that HypE is a competitive search algorithm; this especially applies to higher dimensional problems, which indicates using the Lebesgue measure on many objectives is a convincing approach.

A promising direction of future research is the development of advanced adaptive sampling strategies that exploit the available computing resources most effectively.

HypE is available for download at http://www.tik.ee.ethz.ch/sop/download/ supplementary/hype/.