1 Introduction

Multi-objective expected value programming (MEVP) is a kind of stochastic programming with comprehensive engineering application background, involving multiple conflicting expected value objective functions. Many real-world problems can be formulated by MEVP models, such as transportation logistics, project management, facility location, healthcare management, portfolio investment. Although some achievements on single-objective expected value programming and multi-objective programming were reported (Jin and Branke 2005), less work on MEVP has appeared in the literature, due to computational complexity and challenging difficulty. In particular, it is almost impossible to find a Pareto optimal solution when noisy environments become severe. The main challenge includes three points: (i) designing adaptive sampling schemes with low computational complexity; (ii) discriminating between individuals through individual dominance; and (iii) studying efficient nondominated sorting strategies in noisy environments. When the conventional static sampling method, i.e., the same fixed sample size for each individual, is adopted to handle noises, there are two usual ways to cope with MEVP models (Gutjahr and Pichler 2016; Marler and Arora 2010; Hughes 2001; El-Wahed and Lee 2006; Liu 2009): (i) translating them into single-objective expected value ones solved by some well-known methods such as the weighted sum approach (Marler and Arora 2010; Hughes 2001) and goal programming approaches (El-Wahed and Lee 2006; Liu 2009) and (ii) replacing them by the related sample average approximation models handled by multi-objective evolutionary algorithms (MEAs) (Gutjahr and Pichler 2016). The first type of approach is simple and easy to implement, but usually can only find a single approximate Pareto optimal solution in a run; the second type of approach can produce multiple approximate solutions, but it easily causes high computational complexity.

When solving static multi-objective programming problems, traditional mathematical methods can only give poor performances usually, whereas the MEAs are an alternative and competitive tool and become increasingly popular (Deb 2002; Zitzler and Thiele 1999; Corne et al. 2001; Zhang and Li 2007). Some representative MEAs based on Pareto dominance have been reported in the literature, such as nondominated sorting genetic approach (NSGA2) (Deb 2002), strength Pareto evolutionary approach (SPEAII) (Zitzler and Thiele 1999), Pareto envelope-based selection approach (PESAII) (Corne et al. 2001), multi-objective evolutionary algorithm based on decomposition (MOEA/D) (Zhang and Li 2007). Despite the advantages of structural simplicity and strong availability, they also face some difficulties when directly applied to MEVP problems, due to stochastic factors or noises. Even if so, some reported multi-objective approaches in static environments might solve such kind of problem by attaching static sampling strategies, but it is inevitable that they cause high computational complexity in the process of solution search. Consequently, it is still desired to develop novel multi-objective optimization approaches with efficient sampling strategies.

We in this work suggest a multi-objective expected value immune optimization approach (MEIOA) to solve general nonlinear MEVP problems with unknown noise distributions. This approach, inspired by the principles of clonal selection and immune regulation, is an adaptive optimization tool with the merits of simplicity and practicability. It includes mainly three modules of adaptive racing ranking, immune evolution and archive set update. By comparative experiments, we can draw a conclusion that MEIOA presents significant superiority to the compared methods and also prominent potential to complex nonlinear MEVP problems. It is also pointed out that MEIOA is strongly different from any existing immune optimization techniques. On the one hand, a new racing ranking approach in noisy environments is developed to decide the importance of individual; on the other hand, a useful sample bound estimate is derived to control the dynamic sample size of individual, while a novel polymerization degree model is designed to pick up diverse and high-quality individuals which participate in evolution. In particular, our previous work (Zhang and Tu 2007b) also investigated one multi-objective immune optimization approach (PDMIOA) with static sampling. It is emphasized that such approach is completely different from MEIOA, due to different design inspirations. PDMIOA suppresses noisy influence on the optimized qualities of solutions through individual aging with a maximal lifetime and a reproduction scheme. It includes a modified probabilistic dominance model used for discriminating between low- and high-quality individuals. However, since each individual is attached the same sample size, PDMIOA is difficult in avoiding its high computational complexity.

2 Related work survey

2.1 Noise handling approaches

When random variables are with known distributions, analytic formulas are usually adopted to replace MEVP’s subobjective functions. Unfortunately, their noise distributions are generally unknown in engineering applications, and hence studies on model approximation are an alternative way to find Pareto optimal solutions. Monte Carlo simulation as a simple and useful sampling method is usually taken into account, but it is difficult in deciding the sample sizes of such random variables. In order to cope with noises, three kinds of sampling approaches are adopted in general, i.e., static sampling (Zhang and Tu 2007b; Robert and Casella 2013), sample bound estimation (Shapiro et al. 2009; Hoeffding 1963) and adaptive sampling (Cantú-Paz 2004; Higle and Zhao 2004; Chen 2003). Among these approaches, the first is simple and available but usually causes high computational complexity, as each individual is attached to the same large sample size. The second can be used to control the sample sizes of the random variables, but needs many mathematical theory foundations. For example, after exhaustively discussing the relation between single-objective expected value programming and its related sample average approximation model, Shapiro et al. 2009 acquired a valuable lower bound estimate on sample size. They claimed that one such estimate could make any \(\delta \)-optimal solution of the approximation model approach a corresponding \(\varepsilon \)-optimal solution of the true problem with high probability. The third can greatly reduce computational cost and is helpful to rapidly finding the optimum, and therefore, it becomes increasingly popular in the context of stochastic optimization. About the research on sampling termination conditions of individuals, Cantú-Paz (2004) studied an adaptive sampling technique to decide when to terminate the process of sampling in terms of an one-sided t-test. He suggested that such sampling method could dynamically determine the sample size of individual and thus might reduce computational cost. Thereafter, Higle and Zhao (2004) examined experimentally the difference between adaptive and nonadaptive sampling schemes by using two approaches of stochastic decomposition (SD) and sample average approximation (SAA). Their results show that there exists little difference between SD and SAA when only taking the quality of solution into account. However, such two methods have different efficiencies, namely SAA causes high computational complexity but SD does not. Additionally, in order to identify whether an individual is superior to another one, some researchers (Chen 2003; Lee et al. 2012) try to investigate how to dynamically allocate a total sample size to different individuals by means of special sampling techniques. This can help excellent individuals obtain large sample sizes, and conversely poor individuals only get few samples. For example, Chen (2003) developed an efficient optimal computing budget allocation scheme to determine the most important individuals in a given population. The allocation scheme can allocate a given sample size to different individuals among which better individuals can gain larger sample sizes, but usually it can be only used in finding one or multiple best solutions. In our previous work (Zhang et al. 2013a; Zhang and Tu 2007a), two adaptive sampling schemes were designed to provide different individuals with different sample sizes, based on the hypothesis test and a sample bound estimate. Such schemes can find those valuable individuals in the process of evolution.

2.2 Multi-objective immune optimization approaches

Since the end of the twentieth century, immune optimization as an important branch of artificial immune systems has become popular, as three famous artificial immune models of negative selection, clonal selection and artificial immune network (Aickelin et al. 2014) were originally proposed based on immunology principles. Relying upon such models, researchers developed a large number of immune approaches to handle various kinds of optimization problems. In particular, some researchers have paid great attention to multi-objective immune optimization, for which the main work concentrates on two aspects: (i) probing into original multi-objective immune approaches (MOIAs) (Coello and Cortés 2005; Gong et al. 2008; Qi et al. 2012; Lin and Chen 2013) and (ii) developing hybrid multi-objective immune techniques (Tan et al. 2008; Aydin et al. 2011; Gong 2013; Qi et al. 2015). Several valuable and original MOIAs are drawing researchers’ interests in exploring new immune techniques. For example, Coello and Cortés (2005) suggested a multi-objective clonal selection approach by virtue of the concept of dominance and immune evolution inspirations. They claimed that such approach could ensure those elitist individuals to move toward the true Pareto front. Gong et al. (2008) developed a new nondominated neighbor immune approach (NNIA). In the approach, a novel nondominated neighbor-based selection technique is constructed to pick up diverse individuals, and meanwhile, several immune operators and heuristic search methods are chosen to ensure that those survival individuals transform toward the desired region and elitism. Lin and Chen (2013) proposed a valuable multi-objective micro-population immune optimization approach, relying upon a novel adaptive mutation operator and an efficient immune selection strategy; their adaptive mutation operator can strengthen the exploratory capability of boundary and less-crowded individuals. On the other hand, some researchers devoted themselves to hybrid MOIAs by improving some existing techniques. For instance, Tan et al. (2008) presented a multi-objective evolutionary artificial immune approach, which combined the evolutionary approaches’ ability of global search with immune learning. Depending on their new selection strategy and a density preservation mechanism, the presented approach can maintain the balance between exploration and exploitation.

2.3 Intelligent optimization on MEVP

MEVP is a kind of extremely difficult uncertain programming in the field of optimization. Some researchers made great efforts to explore the possible potential of intelligent approaches for this kind of problem. Their work mainly focuses on investigating how to design new dominance methods, noise handling techniques and efficient evolutionary methods. In this research, individual dominance as a metric way of individual importance plays a role in identifying whether an individual is superior to another one under uncertain environments (Hughes 2001; Trautmann et al. 2009; Eskandari and Geiger 2009). Hughes (2001) suggested a classical model of probabilistic dominance to formulate the relation between individuals. Such model has been comprehensively adopted by researchers who investigate multi-objective expected value problems. Particularly, Eskandari and Geiger (2009) introduced the concepts of stochastic dominance and significant dominance to identify competitive solutions, and subsequently, a stochastic fast Pareto genetic approach (FPGA) was designed to deal with stochastic multi-objective optimization problems. Their computational results showed that such approach was superior to NSGA2 (Deb 2002).

Since noise is a crucial factor to interfere the quality of individual evaluation during solution search (Bui 2005; Park and Ryu 2011; Phan and Suzuki 2012; Lee 2010), many researchers have paid great attention to how to explore noise handling techniques. Bui (2005) experimentally studied noisy influence on solution quality by comparing the performances of probabilistic and resampling methods related to NSGA2. They asserted that probabilistic approaches could effectively find Pareto optimal solutions. Park and Ryu (2011) proposed a new optimization technique which mainly included three new operators of accumulative sampling, ranking and selection. Their experiments hinted that such approach performed well over several multi-objective approaches, relying upon multiple kinds of performance criteria. Another focus on solving single or multi-objective expected value programming is to find an appropriate way to accelerate the process of solution search. Although many existing static multi-objective approaches can also handle MEVP problems under the fixed sample sizes of individuals, they are difficult in addressing various kinds of complex MEVP problems with high dimensions or unknown noise distributions, since their solution qualities depend greatly on sufficiently large and fixed sample sizes. For instance, Drugan and Nowe (2013) developed a multi-objective approach to solve the problem of multi-objective multi-armed bandits in the field of reinforcement learning, based on a standard upper confidence bound approach and a Pareto dominance order relationship. Unfortunately, despite being capable of solving the multi-objective problem directly, such approach causes low performance efficiency. Additionally, we have also investigated two immune-inspired multi-objective optimization approaches to solve MEVP (Zhang and Tu 2007b) and multi-objective chance-constrained programming (Zhang et al. 2013b). They involve in several design inspirations such as immune suppression, immune selection, aging and probabilistic dominance, in which a sample allocation technique is designed to assign large sample sizes to high-quality individuals.

3 Problem statement and preliminaries

Consider the following multi-objective expected value programming problem of the form:

$$\begin{aligned} \hbox {MEVP} \qquad {\mathop {\min }\limits _{{{\varvec{x}}}\in {\varvec{D}}}} \;f({{\varvec{x}}})=E\;[f_1 ({{\varvec{x}}},{\varvec{\xi }} ),f_2 ({{\varvec{x}}},{\varvec{\xi }} ),\ldots ,f_q ({{\varvec{x}}},{\varvec{\xi }} )], \end{aligned}$$

with bounded and closed domain D in \(\hbox {R}^{p}\), decision vector x in D, where \(\varvec{\xi }\) is a r-dimensional random vector with unknown distribution; \(\hbox {E}\;[f_1 ({{\varvec{x}}},{\varvec{\xi }} ),f_2 ({{\varvec{x}}},{\varvec{\xi }} )\ldots ,f_q ({{\varvec{x}}},{\varvec{\xi }} )]\) denotes the expected value vector function \((\hbox {E}\;[f_1 ({{\varvec{x}}},{\varvec{\xi }} )],\hbox {E}\;[f_2 ({{\varvec{x}}},{\varvec{\xi }} )],\ldots ,\hbox {E}\;[f_q ({{\varvec{x}}},{\varvec{\xi }} )]);\) E [.] is the operator of expectation; \(f _{j }({{\varvec{x}}}, \varvec{\xi })\) is the jth nonlinear stochastic subobjective function. In order to seek MEVP’s solutions, the concept of \(\varepsilon \)-dominance (Batista et al. 2011) is usually picked up to execute solution comparison. In other words, for two given candidates \({{\varvec{x}}}, {{\varvec{y}}}{\varvec{\in }}D\) we say that \({{\varvec{x}}}\,\varepsilon \)-dominates \({{\varvec{y}}} ({{\varvec{x}}}\prec _\varepsilon {{\varvec{y}}})\), if

$$\begin{aligned} E[f_j ({{\varvec{x}}},{\varvec{\xi } })]+\varepsilon \le E[f_j ({{\varvec{y}}},{\varvec{\xi } })], \end{aligned}$$
(1)

with \(1\le j\le q\), and there exists k satisfying

$$\begin{aligned} E[f_k ({{\varvec{x}}},{\varvec{\xi } })] +\varepsilon <E[f_k ({{\varvec{y}}},{\varvec{\xi } })]. \end{aligned}$$
(2)

This way, \({{\varvec{x}}}^{*}{\varvec{\in }} D\) is called an \(\varepsilon \)-Pareto optimal solution, if there is no candidate \({{\varvec{z}}} {\varvec{\in }} D\) such that \({{\varvec{z}}}\prec _\varepsilon {\varvec{x}}^{*}\). In particular, for a given finite population A, \({{\varvec{x}}}\in A\) is said to be an \(\varepsilon \)-nondominated individual, if there is no individual y in A such that \({{\varvec{y}}}\,{\varepsilon }\)-dominates \({{\varvec{x}}}\). Similarly, the concept of \(\varepsilon \)-dominance above may be naturally extended into the version of \(\beta \)-dominance (Trautmann et al. 2009; Eskandari and Geiger 2009) which will be used in identifying competitive individuals. In other words, \({{\varvec{x}}}\) \(\beta \)-dominates y with given \(\beta =(\beta _1 ,\beta _2 ,\ldots ,\beta _q )\) and \(\beta _i >0\), if inequalities (1) and (2) are true after replacing \(\varepsilon \) by \(\beta _j\) and \(\beta _k\), respectively. Correspondingly, \(x\in A\) is said to be a \(\beta \)-nondominated individual, if there is no individual y in A such that y \(\beta \)-dominates \({{\varvec{x}}}\).

Generally, when \({\varvec{\xi }}\) is with known distribution \({F_{\xi }}({{\varvec{z}}})\), each of the above subobjective functions can be replaced by

$$\begin{aligned} E\;[f_j ({{\varvec{x}}},{\varvec{\xi }} )]=\int \limits _{{{\varvec{z}}}\in R^{r}} {f_j ({{\varvec{x,z}}})} dF_{\xi } ({{\varvec{z}}}), \end{aligned}$$
(3)

and hence the above MEVP can be changed into an analytically deterministic multi-objective programming problem. However, in many practical problems, the noisy information of \({\varvec{\xi }}\) is unknown and accordingly, the model approximation handling method is an alternative way to solve such kind of problem. Sample average approximation is a simple and popular method used in coping with expected value programming problems with unknown noise distributions. Therefore, we use it to transform the above problem into the following multi-objective sample average approximation model:

$$\begin{aligned} \hbox {SAA} \qquad \mathop {\min }\limits _{{{\varvec{x}}}\in D} \;{{\hat{{f}}}}({{\varvec{x}}})= & {} \left( {{{\hat{{f}}}}_1 ({{\varvec{x}}}),{\hat{{f}}}_2 ({{\varvec{x}}}),\ldots ,{\hat{{f}}}_q ({{\varvec{x}}})} \right) ,\\ \text {s.t.,} {\hat{{f}}}_j ({{\varvec{x}}})= & {} \frac{1}{m}{\mathop {\sum }\limits _{i=1}^{m}} {f_j ({{\varvec{x}}},{\varvec{\xi } }_i )} , \end{aligned}$$

with \(1\le j\le q; m\) denotes the fixed sampling size; \({\varvec{\xi } }_i \), \(1\le i\le m\), are the i.i.d samples of \({\varvec{\xi } }\). \({{\varvec{x}}}\) is said to empirically \(\varepsilon \)-dominate \({{\varvec{y}}}\) (simply say \({{\varvec{x}}}\prec _{\hat{{\varepsilon }}} {{\varvec{y}}})\) if satisfying \({\hat{{f}}}_j ({{\varvec{x}}})+\varepsilon \le {\hat{{f}}}_j ({{\varvec{y}}})\) with \(1\le j\le q\) and \({\hat{{f}}}_k ({{\varvec{x}}})+\varepsilon <{\hat{{f}}}_k ({{\varvec{y}}})\) for some k. Similar to the version of \(\varepsilon \)- or \(\beta \)-dominance above, \({{\varvec{x}}}\) is called an \(\varepsilon \)- or \(\beta \)-empirical nondominated individual if there does not exist any individual y in A such that y \(\varepsilon \)- or \(\beta \)-dominates \({{\varvec{x}}}\) empirically.

We easily know that the set of solutions for the above problem SAA can approach that of the MEVP above when m is sufficiently large according to the law of large number. However, in such case any optimization method will cause expensive computational cost. Consequently, we require that different candidates be attached different sample sizes so as to reduce the cost of computation. Hence, the above SAA is transformed into the following multi-objective sample-dependent approximation (SDA) model:

$$\begin{aligned} \hbox {SDA} \qquad \mathop {\min }\limits _{{{\varvec{x}}}\in D} \;{\hat{{f}}}({{\varvec{x}}})= & {} \left( {{\hat{{f}}}_1 ({{\varvec{x}}}),{\hat{{f}}}_2 ({{\varvec{x}}}),\ldots ,{\hat{{f}}}_q ({{\varvec{x}}})} \right) ,\\ \text {s.t.,} {\hat{{f}}}_j ({{\varvec{x}}})= & {} \frac{1}{m({{\varvec{x}}})}{\mathop {\sum }\limits _{k=1}^{m({{\varvec{x}}})}} {f_j ({{\varvec{x}}},{\varvec{\xi } }_k )} ,1\le j\le q, \end{aligned}$$

where \(m({{\varvec{x}}})\) is the sample size of \({\varvec{\xi }}\) at the point \({{\varvec{x}}}\). We next cite the following conclusions to help us design a novel racing ranking approach to be used in deciding those competitive individuals in a given population.

Theorem 1

(Hoeffding’s inequality) (Hoeffding 1963). Let X be a set, F(.) be a probability distribution function on X; \(f _{1},{\ldots }, f _{n}\) denote the real-valued functions defined on X with \(f _{j}\) : \(X \rightarrow [a _{j}, b _{j}]\) for \(j=1,\ldots , n\), where \(a _{j}\) and \(b _{j}\) are real numbers satisfying \(a _{j} < b _{j}\). Let \(x _{1},{\ldots }, x_{n }\) be the samples of i.i.d random variables \(X _{1}\), \(X _{2},{\ldots }, X _{n }\) on X, respectively. Then, the following inequality is true:

$$\begin{aligned}&\Pr \left( {\left| {\frac{1}{n}\sum _{j=1}^n {f_j \left( {{{\varvec{x}}}_j } \right) } -\frac{1}{n}\sum _{j=1}^n {\int _{{{\varvec{x}}}\in X} {f_j \left( {{\varvec{x}}} \right) } dF\left( {{\varvec{x}}} \right) } } \right| \ge \varepsilon } \right) \nonumber \\&\quad \le e^{-\frac{2\varepsilon ^{2}n^{2}}{\sum _{j=1}^n {\left( {b_j -a_j } \right) ^{2}} }}. \end{aligned}$$
(4)

Corollary 1

(Hoeffding 1963) If \(X _{1}, X _{2},{\ldots }, X _{n}\) are i.i.d random variables with \(a\le X_j \le b,1\le j\le n\), and mean \(\mu \), then

$$\begin{aligned} \left| {{\bar{{X}}}_n -\mu } \right| \le (b-a)\sqrt{\frac{1}{2n}\ln \left( {\frac{2}{\delta }} \right) }, \end{aligned}$$
(5)

with probability at least \(1-\delta \), where \({\bar{{X}}}_n =\frac{1}{n}\sum _{j=1}^n {{\hat{{X}}}_j } \);\({\hat{{X}}}_j \) and \(\delta \) denote the observation of \(X _{j}\) and the significance level, respectively.

4 Lower bound estimate theory

We first investigate two lower bound estimates for the sample sizes of random vector \({\varvec{\xi } }\), which can reflect the relationship between two individual sets in a finite population A with size N. Then, an adaptive racing ranking approach is designed to find \(\beta \)-empirical nondominated individuals in such population. In order to ensure each individual in A a rational sampling size, we investigate lower bound estimates to control the value of \(m({{\varvec{x}}})\) with \({{\varvec{x}}}\in A\), based on the sample average approximation model SAA. More precisely, let \(A_\varepsilon \) and \({\hat{{A}}}_\varepsilon \) be two subsets in A composed of \(\varepsilon \)-nondominated and \(\varepsilon \)-empirical nondominated individuals, respectively. We next give the following conclusions for which the proofs can be found in “Appendix 1”.

Lemma 1

Let \(f_j ({{\varvec{x}}},{\varvec{\xi } })\) be uniformly bounded with \(1 \le \quad j \quad \le q\), namely there exists real number a and b such that \(\Pr \{a\le f_j ({{\varvec{x}}},{\varvec{\xi } })\le b\}=1\) with \(\forall {{\varvec{x}}}\in D\). Then, for any \({{\varvec{x}}}\in A_\varepsilon \) and \({{\varvec{y}}}\in {\hat{{A}}}_\varepsilon \), the probability, which the event \(\{{{\varvec{y}}}\prec _{\hat{{\varepsilon }}} {{\varvec{x}}}\}\) occurs, satisfies

$$\begin{aligned} \Pr \{{{\varvec{y}}}\prec _{\hat{{\varepsilon }}} {{\varvec{x}}}\}\le 2e^{-mc\varepsilon ^{2}}, \end{aligned}$$
(6)

where \(c=\frac{1}{2(b-a)^{2}}\).

Theorem 2

If \(f_j ({{\varvec{x}}},{\varvec{\xi } })\) be uniformly bounded with \(1 \le \quad j \quad \le q\), the following formula holds,

$$\begin{aligned} \Pr \{A_\varepsilon \subseteq {\hat{{A}}}_\varepsilon \}\ge 1-2N^{2}e^{-mc\varepsilon ^{2}}. \end{aligned}$$
(7)

Corollary 2

Under the assumption of Theorem 2, \(A_\varepsilon \) is a subset of \({\hat{{A}}}_\varepsilon \) with probability \(1-\delta \), if

$$\begin{aligned} m\ge M_1 \equiv \frac{1}{c\varepsilon ^{2}}\ln \frac{2N^{2}}{\delta }. \end{aligned}$$
(8)

The above corollary shows that the lower bound estimate \(M _{1 }\) depends on \(c,\delta ,\varepsilon \), and in particular \(N^{2}\). Once N is large, \(M _{1}\) is very large, and hence a large sample bound is needed to control the sampling sizes of individuals in population A. If so, such population will cause high computational complexity. Therefore, we improve the above lower bound estimate below.

Theorem 3

Under the assumption of Lemma 1, the following formula holds,

$$\begin{aligned} \Pr \{{\hat{{A}}}_\varepsilon \cap A_\varepsilon \ne \emptyset \}\ge 1-2Ne^{-mc\varepsilon ^{2}}. \end{aligned}$$
(9)

Corollary 3

Under Theorem 3, at least an \(\varepsilon \)-empirical nondominated individual in A is an \(\varepsilon \)-nondominated individual with probability \(1-\delta \), if

$$\begin{aligned} m\ge M_2 \equiv \frac{1}{c\varepsilon ^{2}}\ln \frac{2N}{\delta }. \end{aligned}$$
(10)

We can easily see that \(M_2 \) is smaller than \(M_1 \). This helps for reducing the computational cost of deciding empirical objective values for individuals in A. Thereby, \(M_2 \) is adopted in the subsequent section to control the sample sizes of individuals.

5 Individual evaluation and similarity measure model

5.1 Individual evaluation

As we mentioned above, a sufficiently large sample size can make the empirical objective value approach the expected objective value at a given individual, but the high computational complexity is inevitable. Fortunately, since those inferior individuals will be gradually eliminated in the process of solution search, it is not necessary to assign large sample sizes to them. Conversely, since those high-quality individuals are required to approach \(\varepsilon \)-nondominated solutions gradually, their sample sizes should be large. Therefore, in order to spend less computational cost to seek those empirically nondominated individuals in a given population \(A=\{x_1 ,x_2 ,\ldots ,x_N \}\), we develop a new adaptive racing ranking approach (ARRA) to decide the \(\beta \)-empirical nondominated set of one such population, based on the reported Hoeffding’s race approach (Even-Dar et al. 2006). The main task of ARRA includes two points: (i) Those superior individuals in A gain large sample sizes and (ii) those \(\varepsilon \)-empirical nondominated individuals in A will be found gradually by means of \(\beta \)-empirical dominance. More precisely, let s be a sample increment; r(x) denotes the importance level for \({{\varvec{x}}}{\varvec{\in }} A\). ARRA is formulated below:

  1. Step 1:

    Input the parameters of \(\varepsilon \) and \(\delta \) as mentioned above, population A and empty set Q;

  2. Step 2:

    Define initial sample size \(m _{0}\) and maximal importance level \(R _{\max }\);

  3. Step 3:

    Set \(m\leftarrow m_0 ,\tau \leftarrow 0,s\leftarrow m_0 ,\lambda \leftarrow \ln (1+\tau )\), and \(Q\leftarrow A\);

  4. Step 4:

    Evaluate all the individuals in Q with the same sample size m;

  5. Step 5:

    Calculate the Hoeffding’s bound through Corollary 1 given by

    $$\begin{aligned} \qquad \,\quad \beta _j =(b_j -a_j )\sqrt{\frac{1}{2m}\ln \left( {\frac{2}{\delta }} \right) }, 1\le j\le q, \end{aligned}$$
    (11)

    where \(b _{j }\) and \(a _{j}\) denote the maximum and minimum of \({\hat{{f}}}_j (x_i )\) with \(x_i \in A\) and \(1\le i\le N\), respectively;

  6. Step 6:

    Sort Q into \(P _\mathrm{nond}\) and \(P _{d}\), and set \(r(x)=\tau \) with \(x\in P_d \), where \(P _\mathrm{nond}\) consists of \(\beta \)-empirical nondominated individuals in Q and \(P _{d}\) is composed of other elements in Q;

  7. Step 7:

    Update \(\tau \leftarrow \tau +1\), \(\lambda \leftarrow \ln (1+\tau )\), \(s\leftarrow \lambda s\), and set \(Q\leftarrow \hbox {P}_{nond;}\)

  8. Step 8:

    Update \(m\leftarrow m+s\);

  9. Step 9:

    Re-evaluate all the individuals in Q with the same sample size m;

  10. Step 10:

    If \(m < \quad M _{2}\) and \(\tau <R_{\max }\), go to Step 5; otherwise, set \(r(x)=\tau +1\) with \(x\in Q\);

  11. Step 11:

    Output Q, \(\hbox {A}\backslash Q\) and r(x) with \(x\in A\).

In the above approach, Step 6 employs the nondominated sorting (Zheng 2004) to determine the \(\beta \)-nondominated set under sample size m, where the empirical objective values of individuals in Q, instead of their expected objective values, are used to execute comparison between individuals. In particular, each individual is attached an importance level. \(R _{\max }\) and \(M _{2}\) control the sample sizes of individuals. With the increasing sample size of m, \(\beta _j \) will become small, and thus those \({\varvec{\beta }}\)-nondominated individuals will be found gradually. In other words, those valuable individuals in A will be produced. Consequently, ARRA can determine those high-quality individuals in A. Additionally, each individual in such approach is evaluated at most \(M_{2 }\) times. When executing \(\beta \)-nondominated sorting, all individuals in Q are divided into two subpopulations, in which in the worst case the computational complexity is \(O(N\log N)\). Therefore, ARRA’s complexity is given by \(O(N(M_2 +\log N))\).

5.2 Similarity measure model

Fig. 1
figure 1

Schematic illustration on individual’s similarity and importance

The well-known crowding distance method (Deb 2002) is usually used to measure the similarity degree of a single individual to other individuals in the objective space. Once an individual has a large crowding distance, its neighborhood includes few individuals. Whereas such method can reflect some similarity features of all the individuals in population A, it cannot present any information on individual quality. We here design a metric to emphasize the similarity and importance for an individual in A. Precisely, let the objective vectors of individuals in A appear in Fig. 1; \(P_{1}\) and \(P_{2 }\) are the two closest points of P; so are \(Q_{1}\) and \(Q_{2 }\) for Q. We easily see that P and Q have the same crowding distance but different qualities. Since \(P_{1}\), \(P_{2 }\) and P are close to the origin, we obtain that \(|S(P,P_1 +P_2 )|<|S(Q,Q_1 +Q_2 )|\), and hence P is better than Q (take the problem of minimization for example), where S(., .) and \({\vert }.{\vert }\) represent the inner product of two vectors and the absolute value of real number, respectively. Thus, in order to measure whether an individual has similar individuals in its neighborhood and whether it is of high quality, we give the following polymerization degree model (PDM),

$$\begin{aligned} \hbox {PDM}({{\varvec{x}}})=\frac{d({{\varvec{x}}})}{S^{2}({{\varvec{x}}},{{\varvec{x}}}_1 )+S^{2}({{\varvec{x}}},{{\varvec{x}}}_2 )}, \end{aligned}$$
(12)

where \({{\varvec{x}}}_{1}\) and \({{\varvec{x}}}_{2}\) are the two closest points of x in A based on the objective space and \(d({{\varvec{x}}})\) presents the crowding distance of x. Notice that \(S({{\varvec{x}}},{{\varvec{y}}})\) is the inner product of empirical objective vectors of x and y. If x approaches the origin and \(d({{\varvec{x}}})\) is large, we think preferably that it is a competitive individual. Thus, Eq. (12) can be used to pick up those both high-quality and diverse individuals in A, while prohibiting more extreme individuals from being selected, e.g., those individuals presented on the top left corner or on the lower right corner.

6 Artificial immune model and algorithm formulation

6.1 Artificial immune model

Adaptive immune response is usually triggered by invading pathogens viewed as antigens (Ag), including the humoral immunity and cell-mediated immunity mediated by B cells and T cells (Owen et al. 2013), respectively. T cells can be divided into several types, e.g., Th1, Th2, Ts and Tc, according to their immune functions. Th1 cells can induce the immune system to kill the infected target cells, whereas Th2 cells can assist B cells to produce antibodies in order to eliminate antigens in body fluid. Additionally, Ts and Th2 can regulate humoral immunity and cell-mediated immunity so as to keep the balance of cell populations. When the antigens invade the living body, B cells and T cells will conduct an immune response process of activation, proliferation and differentiation. In such process, antigen’s counter-specific B cells and T cells will be selected to multiply for effectors. These cells include cytotoxic T cells (Tc) and antibody (Ab). Thereafter, some antigen-specific cells as memory cells stay at the immune system for a long time, while others take place a change through somatic maturation. Figure 2 presents the relationship between different cells, while emphasizing a simple learning and evolutionary process of immune cells, which gives us sufficient bio-inspirations to study how to design immune optimization approaches solving the MEVP problem in Sect. 3.

Fig. 2
figure 2

Artificial immune model

6.2 Algorithm formulation

Fig. 3
figure 3

Flowchart of MEIOA

Inspired by Fig. 2 and immune metaphors, this section formulates our new approach (MEIOA) in detail. As associated with the above problem SDA in Sect. 3 and the racing ranking approach ARRA in Sect. 5.1, a real-coded candidate solution is regarded as an antigen-specific B cell, and meanwhile the problem itself is viewed as the antigen; those best solutions found until now correspond to memory cells. Based on the above ARRA and bio-immune inspirations, MEIOA is shown in Fig. 3 above. It divides the current population into superior and inferior subpopulations by ARRA. On the one hand, the superior subpopulation consists of \(\beta \)-empirical nondominated B cells which also update the memory pool so that those elitist B cells can be reserved in the process of evolution. One such subpopulation also transforms its elements toward the desired regions through proliferation and mutation. On the other hand, the inferior subpopulation is composed of \(\beta \)-empirical dominated B cells, in which each B cell makes change only through uniform mutation. In the whole process of solution search, the operator of immune regulation is used to regulate the sizes of the two subpopulations. Once the number of \(\beta \)-empirical nondominated B cells is beyond a limited bound, some of them are regarded as inferior B cells to enter the inferior subpopulation. MEIOA’s detailed descriptions are as follows:

  1. Step 1:

    Input parameters: population size N, memory pool size M, initial sample size \(m_{0}\), tolerance factor \(\varepsilon \), significance level \(\delta \), high-quality solution ratio \(\eta \%\), maximal clonal size \(C_{\max }\), mutation weight \(\mu \) and maximal importance level \(R_{\max }\);

  2. Step 2:

    Set \(n \leftarrow 1\). Generate an initial population \(\hbox {A}_{n}\) of N random B cells, and initialize memory pool \(\hbox {M}_{set}\);

  3. Step 3:

    Evaluate all the elements in \(\hbox {A}_{n}\) by ARRA, and acquire the importance levels of B cells, and r(x) with \({{\varvec{x}}}\in \hbox {A}_{n}\); divide \(\hbox {A}_{n}\) into two subpopulations \(\hbox {B}_{n}\) and \(\hbox {C}_{n}\), where \(\hbox {B}_{n}\) consists of \(\beta \)-empirical nondominated B cells (in other words, their importance levels are the highest) and \(\hbox {C}_{n}\) is composed of elements in \(A_{n}\) but not in \(\hbox {B}_{n}\);

  4. Step 4:

    Copy elements in \(\hbox {B}_{n}\) into \(\hbox {M}_{set}\), and eliminate those \(\varepsilon \)-empirical dominated B cells; if \({\vert }\hbox {M}_{set}{\vert }> M\), the conventional crowding distance method is used to remove those redundant cells;

  5. Step 5:

    Calculate PDM values of B cells in \(\hbox {A}_{n}\) through Eq. (12);

  6. Step 6:

    If \(\hbox {B}_{n}\) includes at least \(\eta \%\) of elements in \(\hbox {A}_{n}\), those elements in \(\hbox {B}_{n }\) with small PDM values move from \(\hbox {B}_{n}\) to C\(_{n}\); subsequently, those B cells in C\(_{n}\) with large PDM values are kept so that \(\mathrm{C}_{n}\) includes only r B cells with \(r=N-{\vert }\hbox {B}_{n} {\vert }\);

  7. Step 7:

    B cell x in \(\hbox {B}_{n}\) proliferates clonal subpopulation \(Cl({{\varvec{x}}})\) with size \(cl({{\varvec{x}}})= \min (r({{\varvec{x}}})\), \(C _{\max })\); the clones of all the B cells in \(\hbox {B}_{n}\) constitute clonal population \(\hbox {D}_{n}\);

  8. Step 8:

    B cell x in \(\hbox {D}_{n}\) changes its genes through polynomial mutation with mutation rate \(p _{m}({{\varvec{x}}})={\mu }+(1-{\mu })/ (r({{\varvec{x}}})+r _{\max })\); B cell y in \(\hbox {C}_{n}\) mutates its genes through uniform mutation with mutation rate \(p _{m}({{\varvec{x}}})={\mu }+(1-{\mu })/(r({{\varvec{x}}})+1)\); all those mutated B cells form \(\hbox {E}_{n}\);

  9. Step 9:

    Evaluate all elements in \(E_{n}\) by ARRA;

  10. Step 10:

    Execute comparison between the mutated cells and their parents. For each \(\varvec{x}\) in A\(_n\), if \(\varvec{x}\) is inferior to \(\varvec{y}\) by virtue of the version of \(\beta \)-dominance, \(\varvec{x}\) is replaced by \(\varvec{y}\), by which an updated population A\(_{n+1}\) is created, where \(\varvec{y}\) denotes the mutated cell of \(\varvec{x}\) or the best of the mutated clones of it; if the termination criterion is not satisfied, then set \(n\leftarrow n+1\) and go to Step 3; otherwise, output the cells in M\(_{\textit{set}}\).

In the above approach, Step 3 estimates the objective values of B cells and divides \(\hbox {A}_{n}\) into \(\hbox {B}_{n}\) and \(\hbox {C}_{n}\). In Step 6, once the size of \(\hbox {B}_{n}\) is beyond a limited bound, some elements in \(\hbox {B}_{n}\) with small PDM values enter \(\hbox {C}_{n}\), due to the requirement of the diversity of population. Such two subpopulations evolve toward different directions by Step 8. All the mutated B cells are evaluated in Step 9 through ARRA.

7 Computational complexity and performance criteria

7.1 Computational complexity

Based on the above approach description, MEIOA’s computational complexity is decided by Steps 4, 5, 8 and 9. We acquire the following conclusion.

Theorem 3

MEIOA’s computational complexity in the worst case is

$$\begin{aligned} O_c= & {} O((N+M)\log (N+M)\nonumber \\&+\,NC_{\max } (M_2 +p+NC_{\max })). \end{aligned}$$
(13)

Through Eq. (13), MEIOA’s complexity depends mainly on N, M, \(M _{2}\), \(C _{\max }\) and p. We know that \(M_2\) helps the ARRA suppress the noisy influence on high-quality individuals; M is used to ensure that the solutions acquired by MEIOA have satisfactory distributions. Consequently, in order to consider MEIOA’s efficiency, N and \(C_{\max }\) take values as small as possible, e.g., \(N=30\) and \(C_{\max }=2\).

7.2 Performance criteria

We here cite four reported criteria (Gong et al. 2008; Hu 2010) to execute algorithm comparison. Assume that two Algorithms A and B execute only once on some optimization problem, respectively. Correspondingly, they acquire \(\varepsilon \)-empirical nondominated sets P and Q in sequence.

  1. (A1)

    Coverage rate (CR). This can measure the difference of the optimized qualities for A and B, defined as

    $$\begin{aligned} { CR}({ A},\,{ B})=\frac{\left| {\left\{ {{{\varvec{x}}}\in Q|\exists {{\varvec{y}}}\in P, s.t.{{\varvec{y}}}\prec _{\hat{{\varepsilon }}} {{\varvec{x}}}} \right\} } \right| }{\left| Q \right| }. \end{aligned}$$
    (14)

    Obviously, \(0\le CR(A,B)\le 1.\) If \(CR(A,B)> CR(B,A)\), A can achieve better solution search than B. If \(CR(A,B)=1\), the solution quality gained by A is absolutely superior to that obtained by B.

  2. (A2)

    Coverage density (CD). This can formulate the distribution performance of solutions in P, given by

    $$\begin{aligned} \,\,\,\,\,CD= & {} \frac{1}{|P|-1}\sum _{j=1}^{|P|} {(d_j -{\bar{{d}}})^{2}} ,\\ d_j= & {} {\mathop {\min }\limits _{j\ne i,1\le k\le |P|}} \left\{ \left\| {{{\varvec{x}}}_j -{{\varvec{x}}}_k } \right\| ,{{\varvec{x}}}_\mathbf{j} ,{{\varvec{x}}}_k \in P \right\} ,\nonumber \\ {\bar{{d}}}= & {} \frac{1}{\left| P \right| }\sum _{j=1}^{\left| P \right| } {d_j }.\nonumber \end{aligned}$$
    (15)

    Equation (15) indicates that CD takes values within 0 and 1. If \(CD = 0\), the solution set P is totally with uniform distribution, and thus if CD is smaller, the distribution of the solutions in P is better.

  3. (A3)

    Coverage scope (CS). This denotes the coverage width of solutions in P, given by

    $$\begin{aligned}&CS=\mathop {\max }\limits _{1\le j,k\le |P|}\{ {\left\| {{{\varvec{x}}}_j -{{\varvec{x}}}_k } \right\| ,{{\varvec{x}}}_j ,{{\varvec{x}}}_k \in P}\}. \end{aligned}$$
    (16)
  4. (A4)

    Convergence metric (CM). This is utilized to measure how close the solution set P approaches \(P^{*}\), where \(P^{*}\) is the theoretical Pareto optimal solution set. CM is defined by

    (17)

    where dist(xP) represents the Hausdorff distance from x to P. It is clear that A is convergent if \(CM(P ^{*}, P)=0\). Therefore, the smaller the \(CM(P^{*}, P)\), the better the convergence.

Fig. 4
figure 4

Comparison of results on CM based on box plots for problems MOP1 to MOP4. a MOP1, b MOP2, c MOP3, d MOP4

8 Experimental study

All experiments are executed on a Windows XP system with CPU/3.50 GHz and 2.98 GB RAM by means of Visual C++ platform. In order to analyze MEIOA’s intrinsic characteristics, two evolutionary approaches [HEA (Zhang et al. 2015) and FPGA (Eskandari and Geiger 2009)] and three immune optimization approaches [PDMIOA (Zhang and Tu 2007b), NNIA (Gong et al. 2008) and MAMMOIA (Hu 2010)] are selected to participate in comparison by means of two benchmark test suites and one engineering example. It is pointed out that, although NNIA and MAMMOIA are not specially designed to solve stochastic programming problems, they are still efficient and competitive in stochastic environments. On the other hand, FPGA, HEA and PDMIOA with static sampling strategies are developed to deal with MEVP problems, and their individuals in this experiment take the same reported sampling size 300. After manual experimental tuning, MEIOA takes population size 30 but 100 for the compared approaches. In order to be fair, each approach is required to stop its solution search procedure when the total of evaluations in a run is beyond T, while executing 100 times on each test problem. It is emphasized that T is defined according to different kinds of problems below. Other parameter settings of the compared approaches are the same as those in their corresponding literature. In particular, in order to avoid noisy influence on solution comparison between the above approaches, all the best solutions found by them are required to re-evaluate \(10^{5}\) times, since their empirical objective values are used to execute statistical analysis.

8.1 MEIOA’s parameter settings

In MEIOA, \(\varepsilon \) and \(\delta \) do not depend on any prior information on problems solving. The parameter \(\varepsilon \) as a factor of tolerance takes values within 0.01 and 0.1 usually. \(\delta \) used in deciding \(M_{2}\) and the Hoeffding’s bound is the significance level presented in Corollary 2 above, changing within 0.05 and 0.2. In addition, M, \(C_{\max }\) and \(R_{\max }\) influence the optimized quality and efficiency. In order to acquire a set of solutions with uniform distribution and wide coverage scope, M takes values within 60 and 100. \(C_{\max }\) is an important parameter to strengthen the ability of population exploitation. We think that it is rational to define \(C_{\max }\) within 1 and 3. \(\eta \) helps for improving the quality of solution search, taking values within 60 and 80 generally. \(R_{\max }\) is used to decide the maximal sorting level of individuals in the process of evolution. If \(R_{\max }\) is large, ARRA spends much more runtime to decide competitive individuals, and conversely those excellent individuals are identified difficultly. We designate it as 5. Further, \(\mu \) as a minimal mutation rate influences the solution quality to be acquired; it takes values within 0 and 0.3. \(m_{0}\) is an initial sample size of individual included in ARRA, taking values within 10 and 60. After experimental trials, we define \(M=100\), \(m_{0}=30\), \(\eta =80\), \(C_{\max }=2\), \(\varepsilon =0.05\), \(\delta =0.1\) and \(\mu =0.1\).

8.2 Theoretical test example design

In this experiment, nine static multi-objective benchmark problems, proposed originally by Van Veldhuizen and Deb et al. (1999, 2000), i.e., five test problems [MOP1, MOP2, MOP3, MOP4 and MOP6 (Van Veldhuizen 1999)] and four test problems [ZDT1, ZDT2, ZDT3 and ZDT4 (Zitzler et al. 2000)], are picked up to transform into MEVP problems (see “Appendices 2 and 3”) by, respectively, adding a Gaussian noise with the standard normal distribution to their respective subobjective functions. Such MEVP problems are still named by the same versions as their original ones and divided into two sections. The first section, namely the first test suite, includes the above former five test problems, while the other test problems, i.e., the latter four test problems, constitute the second test suite.

It is worth pointing out that in the first test suite, besides MOP1 and MOP3 involving in at most two decision variables, other three examples are all three-dimensional problems. Whereas MOP1 is easily solved in that its subobjective functions are quadratic, other test problems are handled difficultly because of multimodality. These examples are chosen to check if the above approaches can all effectively solve low-dimensional uni- and multi-modal expected value problems. Additionally, the test problems in the second test suite are still difficult, since their Pareto fronts are composed of disjoint curve segments. In order that the approaches can present clear differences for one such test suite, our experiment is executed on these problems according to different dimensional settings, i.e., \(n=30\) and 80.

8.3 Experimental analysis on the first test suite

All the above approaches take the same maximal evaluation number \(2\times 10^{5 }\), among which each of the compared approaches directly executes 100 single executions on the sample average approximation model (SAA) of each test example, and so does MEIOA on the sample-dependent approximation model (SDA) of it. Based on the performance criteria in Sect. 7.2, their statistical results are given in Table 1. Additionally, we only take four examples, MOP1 to MOP4, for example, to display the corresponding box plots (see Fig. 4) on CM, because of the limit of space.

Table 1 Statistical comparison of \(\varepsilon \)-nondominated sets acquired for the first test suite
Table 2 Statistical comparison of \(\varepsilon \)-nondominated sets found for the second test suite with dimension 30

Table 1 presents the performance characteristics of the above four approaches and also their main differences with respect to evolution, noise handling, diversity, search stability, solution distribution, execution efficiency and so forth. By the statistical results on ACR in columns 3–8 and Fig. 4, MEIOA can clearly acquire the best solution qualities for the above examples, since there are always the following coverage relationships for any given test example:

Fig. 5
figure 5

Comparison of convergence based on box plots for the second test suite with dimension 30. a ZDT1, b ZDT2, c ZDT3, d ZDT4

$$\begin{aligned}&ACR\text { (MEIOA, PDMIOA)}\ge ACR\text { (PDMIOA, MEIOA)},\\&ACR\text { (MEIOA, FPGA)}\ge ACR\text { (FPGA, MEIOA)},\\&ACR\text { (MEIOA, HEA)}\ge ACR\text { (HEA, MEIOA)},\\&ACR\text { (MEIOA, MAMMOIA)} \ge ACR\text { (MAMMOIA, MEIOA)}, \\&\text {and}\,ACR\text { (MEIOA, NNIA)} \ge ACR\text { (NNIA, MEIOA)}. \end{aligned}$$

These inequalities illustrate that MEIOA, relying upon the proposed ARRA, can work well over the compared methods and effectively overcome noise interference. For instance, for MOP2 the nondominated sets found by MEIOA cover averagely in order 41, 50, 43, 6 and 46% of those obtained by NNIA, MAMMOIA, HEA, FPGA and PDMIOA; conversely, the latter five approaches can only gain smaller ACR values in comparison with MEIOA. On the other hand, the values on CR and Fig. 4 can ensure that PDMIOA behaves well over the other compared approaches, since it only performs worse for MOP2 and MOP6. We also notice that the compared approaches except PDMIOA have similar solution qualities. As associated with the values on CS, we see that the two evolutionary approaches FPGA and in particular HEA indeed get easily into local search, whereas the immune optimization approaches are opposite. In particular, although NNIA presents the phenomenon of search instability, it is still competitive for the above examples. Totally, with respect to solution quality we claim that MEIOA is best, PDMIOA is secondary, and NNIA and MAMMOIA are better than FPGA and HEA.

Through the results on CD given in columns 9 to 12 in Table 1, we note that the best solutions gained by the above six approaches present different distribution characteristics. For each test example, the solutions obtained by MEIOA can almost uniformly and stably distribute over the theoretical Pareto front, whereas all the compared approaches except NNIA cause poor solution distributions for at least one test problem. We also observe that those best solutions acquired by NNIA are with relatively uniform and stable distributions by comparison against the other compared approaches. Additionally, with respect to solution coverage scope, the values on CS hint that HEA and in particular FPGA can only find extremely crowding solutions with narrow coverage scopes and thus get easily into local search. The four immune optimization approaches, however, can make their individuals evolve along multiple directions and thus present their strong population diversity. With respect to performance efficiency, the values on AT illustrate that, although the above six approaches are with the same termination criterion when solving the same test example, their performance efficiencies are different. The compared approaches spend twice runtime to solve each test problem in comparison with MEIOA. We also know that these compared approaches have similar performance efficiencies, and relatively PDMIOA is more efficient.

As a whole, all the above comments can draw the conclusions: (i) When solving the above first test suite, MEIOA can averagely perform well with the aspects of solution quality, solution distribution and coverage scope as well as performance efficiency, (ii) PDMIOA behaves well over the other compared approaches, while NNIA is also a still competitive approach, and (iii) compared by MEIOA, each of the compared approaches spends much more than runtime to solve each test problem above.

8.4 Experimental analysis on the second test suite

Let the test problems in such test suite share the same dimension 30 or 80. This is done to examine sufficiently whether the above experimental conclusions are true when some hard problems with higher dimensions are taken into account. After manual trials, in the case of \(n=30\) we define the total of evaluations for each approach as \(2\times 10^{5 }\) for problems ZDT1 to ZDT3 but \(2\times 10^{6}\) for ZDT4 because of difficulty. When \(n=80\), however, we set \(T=2\times 10^{6}\) for the former three problems, but \(T=2\times 10^{7}\) for the latter one problem. Like the above experiment, each compared approach directly runs 100 times on the sample average approximation model (SAA) of each of those MEVP problems, and MEIOA executes 100 times on the sample-dependent approximation model (SDA) of it.

Case 1: 30-dimensional experimental analysis

The results acquired by the above approaches are listed in Table 2, and correspondingly, their box plots on convergence metric CM are displayed in Fig. 5.

The values on ACR in Table 2 illustrate that, besides the inequalities in Sect. 8.3 still being true, the solutions obtained by MEIOA for each test problem cover greatly those obtained by other approaches. For example, when solving the hard ZDT4, we can clearly see that the solutions found by MEIOA dominate averagely in order 87, 79, 25, 25 and 82% of those acquired by NNIA, MAMMOIA, HEA, FPGA and PDMIOA; conversely, these latter five approaches can only gain smaller ACR values by comparison against MEIOA. This indicates that ARRA presented in MEIOA can effectively distinguish between individuals in noisy environments. Additionally, FPGA behaves well over NNIA, MAMMOIA and HEA, while PDMIOA and HEA are superior to NNIA and MAMMOIA. Whereas the compared approaches can acquire some nondominated solutions with relatively stable distributions for problems ZDT1 to ZDT3, they fail to solve ZDT4. In particular, it seems to be true that HEA is better than either NNIA or MAMMOIA. In fact, the values on CS hint that it is of poor population diversity and gets easily into local search. In addition, although NNIA is not suitable for ZDT4, we think that it is a still competitive optimizer.

The statistical results on Mean in CD andCS listed in columns 9 and 12 in Table 2 follow that the solutions acquired by MEIOA are with almost uniform distributions and wide coverage scopes; the values on SD demonstrate sufficiently that such approach can perform stable search for the above problems. Other approaches, however, expose some weak search performances when solving ZDT4. Namely, whereas they can acquire some nondominated solutions with wide coverage scopes in comparison with MEIOA, their values on Mean and SD clearly show that their search performances are extremely instable, which can also be found in Fig. 5. In addition, the values on AT, listed in Table 2 derive that, when solving each test problem, the compared approaches spend twice runtime in comparison with MEIOA; MAMMOIA needs the most average runtime; the other compared approaches have similar performance efficiencies.

Graphically, Fig. 5 displays some important and different properties for the approaches. When solving each problem above, those solutions obtained by MEIOA are very close to the Pareto optimal front, as its box plot is very narrow and near the horizontal line. Thus, MEIOA can present the best effects for all the above test problems, whereas other approaches are opposite. Most notably, by the experimental results in Sect. 8.3 we know that PDMIOA is globally superior to the other compared approaches. This, however, is not true in this experiment.

Case 2: 80-dimensional experimental analysis

The statistical results are listed in Table 3, and meanwhile, the corresponding box plots are shown in Fig. 6.

Table 3 Statistical comparison of \(\varepsilon \)-nondominated sets found for the second test suite with dimension 80

The results in Table 3 and Fig. 6 present some distinct differences for the above approaches and expose some prominent drawbacks for the compared approaches. MEIOA can still perform well over other approaches with the aspects of solution quality, distribution, efficiency and so on. Although the compared approaches can exhibit their stable search performances for problems ZDT1 to ZDT3, they cause poor solution search. Particularly, HEA still gets easily into local search. Relatively, FPGA is more suitable for the above high-dimensional problems in comparison with the other compared approaches. NNIA and MAMMOIA are superior to PDMIOA and HEA. On the other hand, the values on AT presented in Table 3 illustrate that the same conclusion on performance efficiency as that in Sect. 8.3 is also true. Consequently, in comparison with the above experimental results in Table 2, we can assert that MEIOA is effective and efficient for the second test suite and FPGA is secondary. The other compared approaches cannot present stable search performances, when the dimensions make a change.

Summarily, when solving the above two test suites, the six approaches exhibit their respective performance characteristics. For the simple uni-modal MEVP problem (MOP1), whereas they can all find some approximate Pareto optimal solutions in a single run, their nondominated sets present different distribution characteristics. MEIOA can effectively solve all the above problems with high efficiency, but other approaches are difficult. Relatively, PDMIOA can perform well over the other compared approaches for the first test suite but fails to solve the second test suite. FPGA can work well over the other compared approaches for the second test suite but cannot do so for the first test suite; HEA can only find local Pareto optimal solutions.

Fig. 6
figure 6

Comparison of convergence based on box plots for the second test suite with dimension 80. a ZDT1, b ZDT2, c ZDT3, d ZDT4

8.5 Experimental analysis on significant difference

Table 4 Statistical comparison of CM based on the Friedman rank and Friedman aligned rank tests

We next detect whether MEIOA shows a significant improvement over the above five compared approaches for the above two test suites. By means of the average of convergence metric values acquired by each approach for the thirteen test problems, i.e., MOP1 to MOP4, MOP6, and ZDT1 to ZDT4 with dimensions 30 and 80, we obtain the statistical values of the Friedman rank and Friedman aligned rank tests given in Table 4. As the table formulates, with respect to solution quality MEIOA shows a significant improvement over the compared approaches with significance level \(\alpha =5\%\); therefore, we can reject the null hypothesis of error convergence metric for MEIOA versus the other approaches. Table 4 also highlights that MEIOA is the most effective algorithm with a rank of 1.00 and 20.77 for the Friedman and Friedman aligned tests, respectively. In addition, since the statistical values 35.42 and 30.63 are larger than 5.99 \((\hbox {X}_{1-\alpha /2}^2 =5.99)\), the above six approaches have significant difference with respect to solution quality.

8.6 Sensitivity analysis

We conduct the current experiment to check if MEIOA’s solution quality is sensitive to the settings of parameters. Here, only two parameters of \(m_{0 }\) and \(\mu \) are used to analyze MEIOA’s sensitivity. \(m_{0}\) as an initial sample size of individual included in ARRA influences the efficiency of solution search, while \(\mu \) as a crucial parameter can strengthen the diversity of population and accelerate population evolution. We define (\(m_{0}\), \(\mu \)) as \((10k,0.1(k-1))\) or \((10(7-k)\), \(0.1(k-1))\) with \(1\le k \le 6\). Other settings of parameters are the same as those in Sect. 8.1. After executing 100 runs on MOP1 and MOP3 as well as ZDT1 and ZDT3 with dimension 30, MEIOA acquires the following statistical results on CM in Table 5.

Table 5 Comparison of ACM values with different settings of (\(m _{0}\), \(\mu \)); ACM denotes the average of 100 results on CM

Table 5 illustrates that the settings of the two parameters influence MEIOA’s solution quality. When \(m_{0}\) is directly proportional to \(\mu \) with \(m_0 >10\) and \(\mu >0\), MEIOA are not totally sensitive to the parameter settings in that its solution quality is not influenced seriously. This phenomenon also takes place in the case where \(m_{0}\) is conversely proportional to \(\mu \) with \(m_0 < 60\) and \(\mu >0\). We also notice that, when \(\mu \) is set as 0, MEIOA’s solution quality is poor, due to large values on ACM for MOP1 and MOP3. Summarily, MEIOA is not sensitive to the settings of \(m_0 \) and \(\mu \), provided that \(m_0 \ge 20\) and \(0.1\le \mu \le 0.4.\) However, in order to take a trade-off between effect and efficiency, we suggest that the two parameters take values with \(30\le m _{0} \le 40\) and \(0.1\le \mu \le 0.3\).

8.7 Engineering application example

We examine further whether the above approaches can effectively solve the machine tool spindle design optimization problem with noise. The original problem includes four decision variables, described by a static bi-objective model (Tan et al. 2001). In order to examine the above approaches, we transform such static model into the following MEVP model:

$$\begin{aligned}&\min E\left[ {f_1 \left( {x,\xi } \right) ,f_2 \left( {x,\xi } \right) } \right] \\&f_1 \left( {{{\varvec{x}}},\xi } \right) =\frac{\pi }{4}\left[ {a\left( {d_a^2 -d_0^2 } \right) +l\left( {d_b^2 -d_0^2 } \right) } \right] +\xi _1 , \\&f_2 \left( {{{\varvec{x}}},\xi } \right) =\frac{Fa^{3}}{3EI_a }\left( {1+\frac{lI_a }{aI_b }} \right) +\frac{F}{c_a }\left[ \left( {1+\frac{a}{l}} \right) ^{2}\right. \\&\qquad \qquad \quad \left. +\frac{c_a a^{2}}{c_b l^{2}} \right] +\xi _2 , \\&x=\left\{ {d_a ,d_b ,d_0 ,l} \right\} ,180\,\mathrm{mm}\le l\le 200\,\mathrm{mm}, \\&d_0 \le \frac{d_b }{1.25}, d_b \le \frac{d_a }{1.05}, d_a \in \left\{ {80,85,90,95} \right\} , \\&d_b \in \left\{ {75,80,85,90} \right\} , c_a =35400\left| {\delta _{ra} } \right| ^{\frac{1}{9}}d_a ^{\frac{10}{9}}, \\&c_b =35400\left| {\delta _{rb} } \right| ^{\frac{1}{9}}d_b ^{\frac{10}{9}},I_a =0.049\left( {d_a^4 -d_0^4 } \right) , \\&I_b =0.049\left( {d_b^4 -d_0^4 } \right) , \\&E=210000\,\mathrm{N/mm}^{2}, F=10000\,\mathrm{N}, \\&\delta _{ra} =0.001, \delta _{rb} =-0.001, \\&\xi _1 ,\xi _2 \sim N(0,\sigma ^{2}). \end{aligned}$$

This is a difficult optimization problem, as both continuous and discrete decision variables as well as some constraints are included. In this experiment, the six approaches take the same maximal evaluation number \(10^{5}\); we examine their ability of noise suppression by considering different settings of variance \(\sigma \). Here, take \(\sigma =0, 0.5, 1\) and 1.5. Like the above experiments, for a given value on \(\sigma \) each approach executes 100 runs. Thereafter, the statistical values are shown in Table 6 below. Take \(\sigma =0.5\) and 1.5 for example. We only display the acquired nondominated fronts in Fig. 7, due to the unknown Pareto front of such problem.

Table 6 Statistical comparison of \(\varepsilon \)-nondominated sets for the above engineering problem
Fig. 7
figure 7

Comparison of nondominated fronts found by the six approaches for the above engineering problem. a \(\sigma =0.5\), b \(\sigma =1.5\)

Through Table 6, it is obvious that the similar conclusions to those in Sects. 8.3 and 8.4 can be drawn. More precisely, whatever \(\sigma \) takes within 0 and 1.5, MEIOA can stably seek some approximate solutions for each run, while Fig. 7 hints that the nondominated fronts acquired by it are with relatively satisfactory distributions and wide coverage scopes. Thus, it can effectively suppress noisy influence on solution search quality. However, other approaches expose their weak ability of noise handling and the poor diversity of population, as the solutions obtained by them depend greatly on the amplitude of \(\sigma \). It seems that their solutions found are with uniform distributions, due to their small values on Mean in CD. In fact, when \(\sigma \) takes a large value, the values on Mean in CS show that the compared approaches, in particular HEA, can only find some solutions which cover narrow regions, because their static sampling schemes result that those high-quality individuals in a given evolving population cannot be effectively discriminated. In addition, we easily see that MEIOA spends the least runtime to solve the above problem and PDMIOA is secondary. We can clearly emphasize that MEIOA is the best for the above problem.

9 Conclusions and further work

This work aims to investigate one multi-objective immune optimization approach (MEIOA) solving a challenging topic in the context of optimization-nonlinear multi-objective expected value programming with unknown noise distribution. We first transform one such kind of multi-objective programming into a sample average approximation model with the same sample size for each candidate solution. Second, this approximation model is extended into a sample-dependent approximation model which the sample size of each candidate depends on its quality. One such model is used to seek the desired approximate solutions. In order to reduce MEIOA’s computational complexity, we study the relationship between the theoretically best individual set and the empirically best individual set included in a given population. This derives out a useful sample bound estimate for each individual. By means of one such estimate, a reported Hoeffding’s inequality and an adaptive sampling rule, an adaptive racing ranking approach (ARRA) is designed to help MEIOA determine those empirically best individuals presented in the current population. Additionally, a novel metric criterion model, based on the well-known crowding distance approach and the version of vector inner product is constructed to pick up some excellent and diverse individuals. Thereafter, as related to the principle of the immune response, some bio-immune inspirations are adopted to develop an immune optimization mechanism which consists of ARRA, proliferation, adaptive mutation and immune selection. The theoretical analysis demonstrates that MEIOA’s computational cost is decided mainly by Mand \(M_{2}\). By two test suites of benchmark problems and an engineering example, comparative experiments have illustrated that MEIOA is a competitive optimizer and also performs well over the compared approaches. The sensitivity analysis has indicated that MEIOA is robust for nonlinear MEVP problems. However, whereas we make some studies on how to explore immune optimization to solve nonlinear MEVP, some issues will be further studied. For example, its structures needs to be optimized in the precondition of improving the solution quality, while its engineering applications are to be further discussed.