1 Introduction

Parameter uncertainty widely exists in engineering optimization problems. Typical uncertainties can be found in geometric dimensions, material properties, loads and boundary conditions, etc. Various methods have been developed for design optimization under uncertainty based on probabilistic method known as reliability-based optimization or robust optimization, and works in this field include: first/second order reliability method (Zhao and Ono 1999), first order second moment (Jung and Lee 2002), adaptive importance sampling technique (Au and Beck 1999), advanced mean value (Wu et al. 1990) and its hybrid variant (Youn et al. 2003), sequential optimization and reliability assessment (Du and Chen 2004), and single-loop design method (Liang et al. 2004), etc. Lee and Chen (2009) conduct an in-depth examination of several widely used probabilistic techniques. In these methods, uncertain parameters are treated as random variables, and their precise probability distribution functions should be predefined based on a large amount of information. However, the fact is that the process of getting adequate uncertainty information is generally expensive and time-consuming, and sometimes even impossible. Additionally, the literature (Ben-Haim and Elishakoff 1990) has shown that small deviations of the parameter probability distributions from the real distributions may cause large errors. Therefore, for many engineering problems without enough uncertainty information, the probabilistic methods will encounter difficulty inevitably.

Different strategies have been studied to overcome the aforementioned limitations of the probabilistic models. For many practical design problems, the bounds of the uncertain parameters can be much more easily identified compared with creating the precise probability distributions. Thus, non-probabilistic models have been considered as attractive supplements to the conventional probabilistic models in practical engineering design problems (Möller and Beer 2008). The uncertain-but-bounded parameters can be treated as convex models such as ellipsoid and interval set (Ben-Haim and Elishakoff 1990). This concept of convex model has been applied to reliability-based design optimization (Ben-Haim 1994; Elishakoff 1995). Since then, optimization under uncertainty using interval models has been addressed by a number of studies, which include: interval arithmetic (Rao and Cao 2002), tolerance box (Parkinson 1995), multi-point approximation (Penmetsa and Grandhi 2002), anti-optimization (Lombardi and Haftka 1998), sensitivity region (Gunawan and Azarm 2005) and multi-ellipsoid convex model (Luo et al. 2008).

Interval models are applicable to problems with incomplete information. It considers only the lower and upper bounds of the uncertain parameters without assuming their precise probability distribution functions. For a specific design point, the responses of objective function and constraints construct intervals due to the interval uncertainty. The bounds of these intervals are the minimal and maximal responses of objective and constraints under the uncertainty. Combined with perturbation technique, interval mathematics was used by Qiu and Elishakoff (1998), Elishakoff et al. (1994) to treat uncertain-but-bounded parameters in the sizing optimization of truss structures. Jiang et al. (2007a) proposed an efficient nonlinear interval number programming method using interval analysis based on the Taylor expansion. These local expansion-based methods, such as Taylor series expansion and perturbation method are difficult to work well when the large variation of uncertain parameters and high nonlinearity of the system are involved. Recently, progresses (Ma 2002) have been made in NIP problem. Attempt has been done to integrate the probabilistic model and interval model together to describe the uncertainties in reliability based design (Du and Sudjianto 2005). Jiang et al. (2008) proposed a general NIP method in which not only the uncertain objective function but also the uncertain constraints are nonlinear, and the variation of the uncertain parameters can be relatively large. In that method, searching for minimal and maximal responses of the objective and constraints is integrated with the main optimization, thus a nested optimization is formed. Besides, most of practical design problems are based on time-consuming simulation models such as FEM, and hence such a nested optimization using actual simulation model will lead to extremely low efficiency.

How to quantify the uncertainty correctly is only one side of the problem, and how to solve the corresponding uncertain optimization efficiently seems a more important side of the problem (Beyer and Sendhoff 2007). Most of the current works focus on modeling the interval-based optimization problems, while this paper places emphasis on solving the NIP problem efficiently and accurately. Currently, approximation models are often used in the computationally intensive optimization problems, which can provide fast estimations of the objective and constraints at new design points. The NIP methods using approximation techniques, named approximation nonlinear interval-based programming (ANIP), are capable to deal with complex engineering problems (Jiang et al. 2007c). The commonly used approximation techniques are: polynomial response surface methodology (Myers and Montgomery 2002), kriging (Sacks et al. 1989; Martin and Simpson 2005), neural networks (Smith 1993), radial basis function (Dyn et al. 1986; Mullur and Messac 2004) and support vector regression (Clarke et al. 2005). In general, there are no clear conclusions on which approximation model is superior among others for every problem (Jin 2005; Acar and Solanki 2009). In this work, RBF approximation model is used because of its fine performance on computational efficiency, numerical stability and capacity of capturing nonlinear behavior (Jin et al. 2001). The accuracy of the approximation is one of the most important issues in all kinds of approximation assisted optimization methods, but in ANIP methods, we only care about the accuracy of approximations in some local regions which have key effects on optimal results. However, design of experiment (DOE) methods such as Latin hypercube design and orthogonal design generate samples using uniform distributions in the entire sampled space. Consequently, if we want to obtain more precise approximations at the concerned regions, we have to increase sample size over the whole sampled space. In fact, the approximation requirement in unconcerned regions can be relaxed for the sake of saving computational cost.

To realize the abovementioned concept, a local-densifying method combined with RBF approximation model is proposed. In the local-densifying method, the current best combinations consisted of the current best design and its corresponding boundary uncertain parameters are sequentially added to the local regions where the minimal and maximal responses of current approximation models take place. Then the RBF approximation models are reconstructed using these densified samples for next iteration until the stop criteria are reached. From the view point of DOE techniques, this proposed local-densifying method is an adaptive DOE method or an on-line approximation method since the new design points are selected according to some specific criteria. However, most of the adaptive DOE methods are developed for deterministic optimization problems (Wang 2003; Li et al. 2007, 2008; Jin 2005). The research works on developing adaptive approximation optimization techniques for NIP is still sparse. In this paper, we attempt to address this topic.

The rest of this paper is organized as follows. A general NIP model is firstly given and the conventional ANIP method is introduced. A new ANIP method with local-densifying method based on RBF models is then proposed, which is the core of this paper. Finally, two numerical examples and a practical engineering application are provided to demonstrate the effectiveness of the present method.

2 Statement of the problem

A NIP problem is given as follows:

$$ \mathop {\min }\limits_{\rm X} f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right) $$

subject to

$$ \label{eq1} {\begin{array}{l} {\kern-6pt} {g_i \left( {{\rm {\bf X}},{\rm {\bf U}}} \right)\le _{\emph{mw}} V_i^I =\left[ {V_i^L ,V_i^R } \right],\quad i=1,...,l} \hfill \\[6pt] {\kern-6pt} {{\rm {\bf U}}\!\in \!{\rm {\bf U}}^I\!=\!\left[ {{\rm {\bf U}}^L,{\rm {\bf U}}^R} \right],\!U_i \in U_i^I \!=\!\left[ {U_i^L ,U_i^R } \right],\,\,\, i=1,2,...,q} \hfill \\[6pt] {\kern-6pt} {{\rm {\bf X}}_l \le {\rm {\bf X}}\le {\rm {\bf X}}_r } \hfill \\ \end{array} } $$
(1)

where X is an n-dimensional design vector, and X l and X r denote the allowable lower and upper boundary vectors of X, respectively. U is a q-dimensional uncertain vector which contains all the uncertain parameters in the system and the uncertain interval vector U I denotes their uncertainty. The superscript I indicates an interval, and L and R denote the lower and upper bounds of the interval, respectively. Uncertainty space is defined as all U ∈ \(\mathbb{R}\) q that satisfy U \(^{L} \le \textbf{U} \le \textbf{U}^{R}\). \(V_i^I\) represents an allowable interval of the ith constraint. l is the number of the constraints. f and g i represent the objective function and the ith constraint, respectively. f and g i are usually obtained from the simulation models in practical applications. They are required to be both nonlinear functions of X and U, and continuous with respect to U in our formulation. Note that the symbol \(\le _{\emph{mw}}\) represents that an interval is better than another but not that one is less than another (Ishibuchi and Tanaka 1990).

3 Treatment of the conventional ANIPs

Using the order relation in interval mathematics (Chanas and Kuchta 1996), the uncertain objective function can be transformed into two deterministic objective functions as follows (Ishibuchi and Tanaka 1990):

$$\begin{array}{lll} \label{eq2} &\mathop {\min }\limits_{\rm X} \left[ {m\left( {f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right)} \right),w\left( {f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right)} \right)} \right] \hfill \\[6pt] &m\left( {f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right)} \right)=\dfrac{1}{2}\left( {f^L\left( {\rm {\bf X}} \right)+f^R\left( {\rm {\bf X}} \right)} \right) \hfill \\[6pt] &\emph{w}\left( {f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right)} \right)=\dfrac{1}{2}\left( {f^R\left( {\rm {\bf X}} \right)-f^L\left( {\rm {\bf X}} \right)} \right) \hfill \end{array} $$
(2)

where m(f(X, U)) and \(\emph{w}(f\)(X, U)) are middle point and radius of the response interval of the objective function, respectively. For a specific design vector X , the possible values of f caused by the uncertainty will form an interval:

$$ \label{eq3} f\left( {{\rm {\bf X}}^\prime ,{\rm {\bf U}}} \right)\in \left[ {f^L\left( {{\rm {\bf X}}^\prime } \right),f^R\left( {{\rm {\bf X}}^\prime } \right)} \right] $$
(3)

the lower bound and upper bound can be obtained using optimization methods:

$$\begin{array}{lll} \label{eq4} f^L\left( {{\rm {\bf X}}^\prime } \right)&=&\mathop {\min }\limits_{\rm {\bf U}} f\left( {{\rm {\bf X}}^\prime ,{\rm {\bf U}}} \right),\quad f^R\left( {{\rm {\bf X}}^\prime } \right)=\mathop {\max }\limits_{\rm {\bf U}} f\left( {{\rm {\bf X}}^\prime ,{\rm {\bf U}}} \right) \hfill \\[6pt] {\rm {\bf U}}^L\le {\rm {\bf U}}&\le& {\rm {\bf U}}^R \hfill \end{array} $$
(4)

Jiang proposed a modified satisfaction degree of interval number which represents a certain degree that one interval number is better (or worse) than another.

$$\begin{array}{lll} \label{eq5} &P\left( {A_{_i }^I \le _{\emph{mw}} B_{_i }^I } \right)\ge \lambda _i \;\left( {P,\lambda _i \in \left[ {0,1} \right]} \right) \hfill \\[6pt] &A_{_i }^I =g_i \left( {{\rm {\bf X}},{\rm {\bf U}}} \right)=\left[ {g_i^L \left( {\rm {\bf X}} \right),g_i^R \left( {\rm {\bf X}} \right)} \right],\;B_{_i }^I =\left[ {\emph{v}_i^L ,\emph{v}_i^R } \right] \hfill \\[6pt] &g^L\left( {\rm {\bf X}} \right)=\mathop {\min }\limits_{\rm {\bf U}} g\left( {{\rm {\bf X}},{\rm {\bf U}}} \right),\quad {\rm g}^R\left( {\rm {\bf X}} \right)=\mathop {\max }\limits_{\rm {\bf U}} g\left( {{\rm {\bf X}},{\rm {\bf U}}} \right) \hfill \end{array} $$
(5)

where the satisfaction degree \(P(A^{I} \le_{\emph{mw}} B^{I})\) can be computed through the following formula (Jiang et al. 2008):

$$\begin{array}{lll} &P\left( {A^I\le _{\emph{mw}} B^I} \right)\\ &=\left\{ {{\small\begin{array}{lll} &{\kern-15pt} {0,} & {A^L\ge B^R} \\[6pt] &{\kern-15pt} {0.5\cdot \dfrac{B^R-A^L}{A^R-A^L}\cdot \dfrac{B^R-A^L}{B^R-B^L},} & {B^L\le A^L<B^R\le A^R} \\[6pt] &{\kern-15pt} {\dfrac{B^L-A^L}{A^R-A^L}+0.5\cdot \dfrac{B^R-B^L}{A^R-A^L},} &{A^L<B^L<B^R\le A^R} \\[6pt] &{\kern-15pt} {\dfrac{B^L-A^L}{A^R-A^L}+\dfrac{A^R-B^L}{A^R-A^L}\cdot \dfrac{B^R-A^R}{B^R-B^L}} & {A^L<B^L\le A^R<B^R} \\[6pt] &{\kern6pt} {+\ 0.5\cdot \dfrac{A^R-B^L}{A^R-A^L}\cdot \dfrac{A^R-B^L}{B^R-B^L},}&\\[6pt] &{\kern-15pt} {\dfrac{B^R-A^R}{B^R-B^L}+0.5\cdot \dfrac{A^R-A^L}{B^R-B^L},} & {B^L\le A^L<A^R<B^R} \\[6pt] &{\kern-15pt} {1,} & {A^R<B^L} \\ \end{array} }} \right. \end{array}$$

where \(A_i^I \) is the interval of the ith constraint at X and \(B_i^I \) is the ith allowable interval. P represents the satisfaction degree that the constraint is satisfied. λ i is a predetermined satisfaction level. The feasible field of the constraints (5) can be controlled by adjusting the value of λ i , namely a greater λ i indicates a smaller feasible field.

Linear combination method is then employed to integrate the deterministic two objective functions, and the penalty function method is used to deal with the constraints. Thus a single-objective optimization problem in terms of a penalty function is finally constructed:

$$\begin{array}{lll} \mathop {\min }\limits_{\rm X} f_p =& \left( {1-\beta } \right)\frac{\left( {m\left( {f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right)} \right)+\xi } \right)}{\phi }\\ &+\ \beta \frac{\left( {\emph{w}\left( {f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right)} \right)+\xi } \right)}{\psi }\\ &+\ \sigma \sum\limits_{i=1}^l {\varphi \left( {P\left( {A_i \le _{\emph{mw}} B_i } \right)-\lambda _i } \right)} \end{array}$$

Subject to

$$ \label{eq6} {\rm {\bf X}}_l \le {\rm {\bf X}}\le {\rm {\bf X}}_r $$
(6)

where 0 ≤ β ≤ 1 is a weighting factor of the two objective functions. ξ is a number making m(f(X, U)) + ξ and \(\emph{w}(f\)(X, U)) + ξ non-negative. ϕ and ψ are two normalization factors. σ is a penalty factor which is usually specified as a large value. φ is a function in the following form:

$$\begin{array}{lll} \label{eq7} &\varphi \left( {P\left( {A_i \le _{\emph{mw}} B_i } \right)-\lambda _i } \right)\\[12pt] &{\kern4pt}=\max \left[ {0,-\left( {P\left( {A_i \le _{\emph{mw}} B_i } \right)-\lambda _i } \right)} \right]^2 \end{array} $$
(7)

Equation 6 can be solved by traditional optimization methods. It can be found from (4) and (5) that the inner optimizations are involved in each main optimization step. The computational cost will be very expensive when the simulation models are time-consuming. Approximation models thus can be used to improve the computation efficiency, and then the general NIP problem (1) can be transformed into the following ANIP problem:

$$ \mathop {\min }\limits_{\rm X} \tilde {f}\left( {{\rm {\bf X}},{\rm {\bf U}}} \right) $$

subject to

$$\begin{array}{lll} \label{eq8} &\tilde {g}_i \left( {{\rm {\bf X}},{\rm {\bf U}}} \right)\le _{\emph{mw}} V_i^I =\left[ {V_i^L ,V_i^R } \right],\quad i=1,...,l \hfill \\[6pt] &{\rm {\bf U}}\!\in\! {\rm {\bf U}}^I\!=\!\left[ {{\rm {\bf U}}^L,{\rm {\bf U}}^R} \right],\!U_i \in U_i^I \!=\!\left[ {U_i^L ,U_i^R } \right],\,\,\, i\!=\!1,2,...,q \hfill \\[6pt] &{\rm {\bf X}}_l \le {\rm {\bf X}}\le {\rm {\bf X}}_r \hfill \end{array} $$
(8)

where \(\tilde {f}\) and \(\tilde {g}_i \) represent the approximate objective function and constraints, respectively. A conventional solving process for the ANIP optimization can be outlined in Fig. 1. At each iteration, uncertain parameters are treated the same as design variables, thus a n + q dimensional hybrid space is formed. DOE methods are used to select the samples within the hybrid space. Then the approximation models can be built using these samples. Consequently, the computational efficiency can be greatly improved, but the accuracy of the optimal results obtained by ANIP methods cannot be guaranteed through this one-step approximation.

Fig. 1
figure 1

Optimization process of a NIP problem based on approximation models

4 Local-densifying method based on RBF approximation models

4.1 Radial basis functions approximation

Radial basis functions have been developed for the interpolation of scattered multivariate data (Queipo et al. 2005). The method uses linear combinations of radially symmetric functions based on Euclidean distance or other such metric. A RBF model can be expressed as:

$$ \label{eq9} \tilde {f}\left( {{\rm {\bf x}}_p } \right)=\sum\limits_i^m {w_i h_p \left( {r_i } \right)} $$
(9)

where w i represents the coefficient of the linear combinations. m is the number of samples. x p represents the vector of estimation point p. h p (r i ) is the radial basis function with respect to the Euclidean distance between the estimation point p and ith sample where \(r_i =\left\| {{\rm {\bf x}}_p -{\rm {\bf x}}_i } \right\|_2 \). The main feature of radial basis functions is that their response decreases or increases monotonically with the variable r. Some commonly used radial basis functions are multi-quadratics, thin plate spline and Gaussian. In our work, the Gaussian radial basis function is adopted:

$$ \label{eq10} h_p \left( {r_i } \right)=\exp \left( {-r_i^2 /c^2} \right) $$
(10)

The parameter c is the average spacing of all the samples. It controls the decay rate of the function. A set of m samples are needed to define the coefficients of RBF model which can be expressed in matrix form as:

$$ \label{eq11} {\rm {\bf f}}={\rm {\bf Hw}} $$
(11)

where H is a matrix:

$$ \label{eq12} {\rm {\bf H}}=\left( {{\begin{array}{*{20}c} {h_1 \left( {r_1 } \right)} \hfill & {h_1 \left( {r_2 } \right)} \hfill &\cdots \hfill & {h_1 \left( {r_m } \right)} \hfill \\ {h_2 \left( {r_1 } \right)} \hfill & {h_2 \left( {r_2 } \right)} \hfill &\cdots \hfill & {h_2 \left( {r_m } \right)} \hfill \\ \vdots \hfill & \vdots \hfill & \ddots \hfill & \vdots \hfill \\ {h_m \left( {r_1 } \right)} \hfill & {h_m \left( {r_2 } \right)} \hfill &\cdots \hfill & {h_m \left( {r_m } \right)} \hfill \\ \end{array} }} \right) $$
(12)

If the inverse of H exists, the coefficient vector can be obtained:

$$ \label{eq13} {\rm {\bf w}}={\rm {\bf H}}^{-1}{\rm {\bf f}} $$
(13)

It has been proved that the matrix H is always invertible for arbitrary scattered samples (Liu 2003).

In conventional ANIP method, large approximate errors in the bounds of response intervals will make the results untrustworthy. To improve the precision, a large sample size is needed to construct high-fidelity approximation models. There is a trade-off between accuracy and efficiency. To pursue high efficiency, sometimes we have to sacrifice the high accuracy. In conventional ANIP methods, the samples are uniformly selected by DOE methods. It is necessary to arrange the samples more appropriately because the result of NIP method is only significantly affected by some key local regions.

4.2 Iterative mechanism

Local-densifying method is an updating strategy of sampling method focusing the limited sample resources on the concerned local regions. The RBF approximation models are reconstructed using the local densified samples in each iteration step. Therefore, it ensures that the approximations have small approximate errors on the bounds of response intervals so that the result can be more reliable. However, local-densifying may lead to overlap of some samples, and thus the matrix H would be singular or ill-conditioned. Therefore, a treatment is needed to exclude the added points which are too close to the previous samples.

The iterative process of local-densifying method based on RBF approximation models for ANIP is outlined as follows:

  1. (a)

    Obtain initial samples by Latin hypercube design method within the hybrid space, and give allowable errors ε 1 > 0 and ε 2 > 0, and make j = 1.

  2. (b)

    Create RBF approximation models with the samples, and construct the ANIP problem by replacing the objective function and constraints with the approximation models. Then, the current best design vector X (j) is obtained by solving the ANIP problem.

  3. (c)

    The response interval \(\left[ {\tilde {f}^L\left( {{\rm {\bf X}}^{\left( j \right)}} \right),\tilde {f}^R\left( {{\rm {\bf X}}^{\left( j \right)}} \right)} \right]\) of approximate objective and corresponding uncertain parameters \({\rm {\bf U}}_f^L \)and \({\rm {\bf U}}_f^R \) can be obtained:

    $$\begin{array}{lll} \label{eq14} \tilde {f}^L\left( {{\rm {\bf X}}^{\left( j \right)}} \right)&=\tilde {f}\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^L } \right),\\ \tilde {f}^R\left( {{\rm {\bf X}}^{\left( j \right)}} \right)&=\tilde {f}\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^R } \right) \end{array} $$
    (14)

    That means the approximate objective function achieves the minimum and maximum at the combinations \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^L } \right)\) and \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^R } \right)\), respectively. Similarly, the response intervals of the approximate constraints \(\left[ {\tilde {g}_z^L \left( {{\rm {\bf X}}^{\left( j \right)}} \right),\tilde {g}_z^R \left( {{\rm {\bf X}}^{\left( j \right)}} \right)} \right]\), z = 1 ···l and corresponding uncertain parameters \({\rm {\bf U}}_{g_z }^L ,{\rm {\bf U}}_{g_z }^R \), z = 1 ···l are obtained:

    $$\begin{array}{lll} \label{eq15} \tilde {g}_z^L \left( {{\rm {\bf X}}^{\left( j \right)}} \right)&=\tilde {g}_z \left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_{g_z }^L } \right),\\ \tilde {g}_z^R \left( {{\rm {\bf X}}^{\left( j \right)}} \right)&=\tilde {g}_z \left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_{g_z }^R } \right),\;z=1\cdots l. \end{array} $$
    (15)

    The zth approximate constraint achieves the lower and upper bounds at the combinations \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_{g_z }^L } \right)\) and \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_{g_z }^R } \right)\), respectively.

  4. (d)

    Compute the responses of actual objective \(\big[f^{L}\)(X (j)), f R(X \(^{(j)})\big]\) and actual constraint \(\left[ {g_z^L \!\left( {{\rm {\bf X}}^{\left( j \right)}} \right),g_z^R \!\left( {{\rm {\bf X}}^{\left( j \right)}} \right)} \right]\), z = 1 ···l at the combinations obtained from step (c).

  5. (e)

    Compute the deviation e max :

    $$\begin{array}{lll} \label{eq16} e_{\max } &=\max \left\{ {\left| {\frac{f^L-\tilde {f}^L}{f^L}} \right|+\left| {\frac{f^R-\tilde {f}^R}{f^R}} \right|,\left| {\frac{g_z^L -\tilde {g}_z^L }{g_z^L }} \right|}\right.\\ &{\kern36pt}\left.{ +\ \left| {\frac{g_z^R -\tilde {g}_z^R }{g_z^R }} \right|,z=1\cdots l} \right\} \end{array} $$
    (16)

    If e max  < ε 1, then X (j) is selected as the final optimal design vector and the iteration terminates, otherwise, turn to step (f).

  6. (f)

    Calculate the distances between \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^L } \right)\) and all the other samples and obtain the minimal distance d min . If d min  > ε 2, add \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^L } \right)\) to the sample set; calculate the distances between the combination \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^R } \right)\) and all the samples and obtain the minimal distance d min . If d min  > ε 2, then add \(\left( {{\rm {\bf X}}^{\left( j \right)},{\rm {\bf U}}_f^R } \right)\) to the samples. Similarly, the boundary combinations of each constraint are all treated with the above process, and the combinations, which do not satisfy d min  > ε 2, are excluded. If all the boundary combinations fail to satisfy d min  > ε 2, the iteration terminates, and X (j) is selected as the final optimal design vector; otherwise, set j = j + 1 and turn to step (b).

In step (a), the initial sample size can be relatively small because the initial approximation models, which are used as guides for the subsequent densifying, do not need to be highly accurate. The aim of this step is to construct approximation models capturing the major characteristics of the actual objective function and constraints. They can help to identify the regions that should be densified. However, the sample size can not be too small. A too small sample size may result in the loss of key information. For such case, the optimization may converge to the local optimum. In step (c), the main optimization and the inner optimizations are both solved by intergeneration projection genetic algorithm (IP-GA; Xu et al. 2001), which is a modification based on micro GA (Krishnakumar 1989).

In step (e), the stop criterion e max  < ε 1 is used, whose satisfaction indicates that the approximation models of objective and constraints have a fine approximate accuracy at the neighborhoods of the lower and upper bounds. That can ensure the trustworthiness of the optimal result, since the computation of ANIP method greatly depends on these response bounds.

Step (f) is the process of sample densifying. Current best combinations are added to the samples at each iteration step, and then the approximation accuracy in the boundary regions can be improved gradually. From (12), we can find that if two samples overlap or locate too close, the H matrix will be singular or ill-conditioned and the RBF approximation models will fail to be constructed. Additionally, we can see from the decreasing feature of Gaussian radial function that the estimations within the neighborhood of a sample have high approximate accuracy. In other words, it does not help to promote the approximate accuracy when the distance between two samples becomes very close. They will, in contrast, make the computation unstable. Therefore, an allowable error ε 2 is set to determine which boundary combinations should be excluded. A high approximate accuracy in the key regions are considered to be attained when all the boundary points satisfy the condition d min  < ε 2, and then the iteration terminates.

The flowchart of the present method is shown in Fig. 2. It can be found that the computation time is spent in three parts. The first part is to compute the samples for constructing the approximation models. For complex simulation models, it would dominate the process. The second one is to search the bounds of objective and constraints. As the approximation models are used, the optimization efficiency can be very high. The third one is to validate the approximation responses, in which the actual responses of the objective and constraints need to be computed.

Fig. 2
figure 2

Uncertain optimization method combined A-NIP with local-densifying method

5 Numerical examples and discussions

Two numerical examples are given to test the proposed method. The first simply numerical test is given for the purpose of intuitively understanding how this method processes. The second complex numerical test is present to compare the proposed method with the conventional method with respect to the computational efficiency and accuracy. Finally, a practical engineering application is provided to demonstrate the effectiveness of the present method in real-world engineering.

5.1 Numerical test 1

A numerical test problem is given as:

$$\begin{array}{lll} &\mathop {\min }\limits_x f\left( {x,u} \right)\\ &=-24\sin \left(\left. {\sqrt {\left( {\frac{x-40}{2}} \right)^2+\left( {u-5} \right)^2} } \right)\right/\\ &{\kern14pt} \sqrt {\left( {\frac{x-40}{2}} \right)^2+\left( {u-5} \right)^2} +46 \end{array}$$

subject to

$$\begin{array}{lll} \label{eq17} 24\le x&\le 43 \hfill \\ u\in U^I&=\left[ {3.5,16.5} \right] \hfill \end{array} $$
(17)

This is a simple unidimensional optimization problem. The uncertainty level is given as a relatively large value ±65% off from the midpoint. The normalization factors φ and ψ are set to 5.0, respectively. The weighting factor β is specified as 0.4. ξ and σ are set to 0.0 and 100,000, respectively. The allowable errors ε 1 and ε 2 are set to 0.003 and 0.15, respectively. The number of objective function evaluations is used to examine the efficiency of this method.

Additionally, the actual response of the objective function is showed in Fig. 3. The optimal design variable 39.99 is obtained from the NIP method with the actual model, in which IP-GA is employed to solve the nested optimization. The maximal generations for IP-GA are both specified as 100. The optimal interval of objective function is [22.00, 51.21], and the corresponding boundary combinations are (39.99, 5.00) and (39.99, 9.50), respectively. The value of penalty function is 5.562. This result is used as an accurate solution to test the accuracy of the present method.

Fig. 3
figure 3

Actual response of test function 1

Four cases with 4 different initial sample sizes are investigated, and the optimization results are listed in Table 1, where ISN stands for the initial sample size, and x* is the optimal design value. NI and Nobj stand for the total number of iterations and the number of objective function evaluations, respectively.

Table 1 Results with different initial sample number (test function 1)

It can be found from Table 1 that the results obtained from the present method are almost the same as the accurate solution when the initial sample sizes are 10, 20 and 30. We can also see that the total number of iterations decreases as the number of initial samples increases. It needs only two iterations when the initial sample size is 30, which implies that 30 initial samples are sufficient to construct a high quality approximation model for this simple problem and there is no need to densify more samples. In the fourth case, the number of objective function evaluations is less than the second and third cases even though the largest initial sample size is used. The results indicate that blindly decreasing the initial sample size would not help to reduce the computational cost, since fewer initial samples may lose the key information of the objective function. In the first case, five initial samples are used and the optimization process converges after eight iterations. The optimal design point is far from the accurate solution. The penalty function value, however, is very close to the accurate one. Two reasons lead to this result: Firstly, the NIP problem is solved by transforming the single-objective uncertain problem into two-objective deterministic problem. One objective minimizes the midpoint of the response interval and another minimizes the radius of the response interval. In such case, either the design point that causes the minimal midpoint or the one causes the minimal radius could be the optimal design. Secondly, if the key characteristics of the objective function are missed because of insufficiency of initial samples, the optimization will be hard to find the global optimum, and the local-densifying method is also difficult to refine the approximation model.

In order to interpret the details of this process better, Table 2 lists each iterative results of the third case. From the table, we can see that the approximate response interval \(\big[ {\tilde {f}^L,\tilde {f}^R} \big]\) is far from the accurate response interval \(\big[f^{L}\), \(f^{R}\big]\) under the same design variable at the beginning. The difference of \(\big[ {\tilde {f}^L,\tilde {f}^R} \big]\) and \(\big[f^{L}\), \(f^{R}\big]\) is getting smaller and smaller as the iteration proceeds. It shows that the initial approximation model has a globally low accuracy, and approximations in the boundary regions could achieve a high accuracy after several local-densifying steps. Figure 4 shows the transformation of the approximation model. The section curves of approximation model and actual model under the current best design in each iteration step can be compared visually.

Table 2 Optimization results of the third case
Fig. 4
figure 4figure 4

Approximation models and section comparisons

We can learn from Fig. 4 that the section curves of approximation model and actual model are prone to coincidence on the lower and upper bounds as the iteration processes. At the final step, the two curves nearly overlap in the key regions marked with circles. As the lower and upper bounds of objective are only needed for the transformation of the uncertain optimization problem in (2), the poor approximation accuracy in the region marked with rectangle does not affect the optimal result. This feature shows the advantage of local-densifying method that focuses the approximation accuracy only on the key regions and alleviates the requirement of approximation in the rest space. The distributions of samples at initial step and final step are shown in Fig. 5. We can see that the added points are densified in the local regions marked with circles, where the function has minimal and maximal responses under the optimal design variable, namely the neighborhoods of combinations (39.99, 5.00) and (39.99, 9.50). If the conventional ANIP method is used, the entire space should have the same sample density as these two local regions for constructing a high quality approximation model.

Fig. 5
figure 5

Sample distributions of initial and last steps

5.2 Numerical test 2

A more complex three-dimensional testing problem with two inequality constraints is given as follows:

$$ \mathop {\min }\limits_{\rm {\bf X}} f\left( {{\rm {\bf X}},{\rm {\bf U}}} \right)=U_1 \left( {X_1 -2} \right)^2+U_2 \left( {X_2 -1} \right)^2+U_3 X_3 $$

subject to

$$\begin{array}{lll} \label{eq18} {\kern-12pt}&U_1 X_1^2 -U_2^2 X_2 +U_3 X_3 \ge _{\emph{mw}} \left[ {6.5,7.0} \right] \hfill \\ &U_1 X_1 +U_2 X_2 +U_3^2 X_3^2 +1\ge _{\emph{mw}} \left[ {10.0,15.0} \right] \hfill \\ &-2\le X_1 \le 6,\;-4\le X_2 \le 7,\;-3\le X_3 \le 8 \hfill \\ &U_1 \in \left[ {0.6,1.8} \right],\;U_2 \in \left[ {0.5,1.5} \right],\;U_3 \in \left[ {0.6,2.0} \right] \hfill \end{array} $$
(18)

X is the design variable vector and U represents the uncertain parameter vector. The uncertainty level are given to relatively large values of 50%, 50% and 53.8%, respectively. The normalization factors φ and ψ are set to 1.4 and 2.0 respectively. The weighting factor β is specified as 0.5. ξ and σ are set to 4.0 and 100000, respectively. The satisfaction levels of the two constraints λ 1 and λ 2 are both specified as 0.8. The allowable errors ε 1 and ε 2 are set to 0.05 and 0.15, respectively. The approximation models of objective and constraints are built simultaneously, and the present method is applied.

Firstly, the accurate solution X = (4.01, 0.87, −3.00)T is obtained from the NIP method with actual models, in which IP-GA is employed to solve the nested optimization. The maximal number of generations for outer IP-GA is specified as 1,000 and the inner is specified as 300. The optimal response interval of objective function is [−3.49, 5.49]. The value of penalty function is 3.91.

Four cases with different initial sample sizes are investigated. The initial sample sizes are 30, 40, 50 and 60, and the results are listed in Table 3. As the initial sample size increases, the total iterations decrease gradually and the number of objective evaluations tends to increase. All the satisfaction degree at the optimal design points are satisfied (great than or equal to the satisfaction levels λ 1 = 0.8 and λ 2 = 0.8). Besides, the optimal results of the four cases slightly differ from the accurate solution X = (4.01, 0.87, −3.00)T, though all the penalty function values are close to the accurate one f p = 3.91. It means that almost the same objective value is achieved with different optimal design variables, and all of them are acceptable for optimal design.

Table 3 Results of different initial sample numbers (test function 2)

In the above four cases, the total numbers of the samples are 70, 78, 86 and 90, respectively. The conventional ANIP method is applied to solve this problem with the same sample size as the four cases finally use. The samples are distributed uniformly in the entire hybrid space by Latin hypercube design method. The outer and inner optimizations are both solved by IP-GA algorithm, and Table 4 lists the results. We can see that the errors on the bounds of response intervals between the approximation models and the actual models are large. In addition, the penalty function values at the optimal design are 7.98, 7.75, 7.83 and 7.74, while the optimal values are 3.98, 3.99, 3.92 and 4.03 in Table 3. It demonstrates the effectiveness of the present method and also indicates that a much more reliable and accurate result could be found with the same sample size compared to the conventional ANIP method. In such a six dimensional hybrid space, high accuracy approximation models are very difficult to be constructed by using limited samples with uniform distribution; nevertheless a fine result can be obtained by supplementing with local-densifying method.

Table 4 Results of uniform distributional samples (test function 2)

5.3 Application for structural crashworthiness

An automobile thin-walled beam model as shown in Fig. 6 is investigated. Closed-hat beam, a typical thin-walled beam, is constructed by a hat-shaped part and a plate, connected by spot welds. It is one of the important structures of a vehicle body for load-support and energy-absorption for crashworthiness. Optimizing the thin-walled beam structures to improve their crashworthiness performance is very important to the vehicle security design. The application of structural optimization for a closed-hat beam impacting a rigid wall is investigated. Based on the reference (Kurtaran et al. 2002), the closed-hat beam will be optimized to maximize the absorbed energy subjected to an average normal impact force on the rigid wall. The research (Wang 2002) indicated that the plate thickness t, round radius R of the hat beam, and the space length d of each two neighboring spot-welding points have prominent effects on the crashworthiness performance of a closed-hat beam, and hence these three parameters are used as design variables in this application.

Fig. 6
figure 6

A closed-hat beam impacting the rigid wall and its cross-section (millimeter)

The FEM model of the closed-hat beam impacting a rigid wall with the initial velocity of 10 m/s is shown in Fig. 7. The simulation duration of the impact process is set as 20 ms. An elasto-plasticity material model of bilinear kinematic hardening is used for the closed-hat beam, as given in Table 5.

Fig. 7
figure 7

FEM mesh of the closed-hat beam impacting a rigid wall

Table 5 Material properties of the closed-hat beam

The Belytschko–Tsay shell element is used to generate the FEM mesh, and the total number of the elements is 4,200. In order to achieve enough impact energy, a 250-kg mass is attached to the end of the closed-hat beam. The nominal values of the yield stress σ s and tangent modulus E t are 310 and 763 Mpa, respectively. σ s and E t are treated as the uncertain parameters because of the manufacturing and measuring errors. The uncertain level is 5% off from the normal values, namely σ s  ∈ [294.5 Mpa, 325.5 Mpa] and E t  ∈ [724.85 Mpa, 801.15 Mpa]. As a result, this optimization problem can be expressed as the following form:

$$ \mathop {\max }\limits_{t,R,d} f_e \left( {t,R,d,\sigma _s ,E_t } \right) $$

subject to

$$\begin{array}{lll} \label{eq19} &g_f \left( {t,R,d,\sigma _s ,E_t } \right)\le _{\emph{mw}} \left[ {65{\rm KN},70{\rm KN}} \right] \hfill \\[6pt] &\sigma _{\rm s} \in \left[ {294.5\,{\rm Mpa},325.5\,{\rm Mpa}} \right],\\[6pt] &E_t \in \left[ {724.85\,{\rm Mpa},801.15\,{\rm Mpa}} \right] \hfill \\[6pt] &0.5\,{\rm mm}\le t\le 2.5\,{\rm mm},\;1{\rm mm}\le R\le 8\,{\rm mm},\\[6pt] &10{\rm mm}\le d\le 60\,{\rm mm} \hfill \end{array} $$
(19)

where the objective function f e and constraint g f represent the absorbed energy of the closed-hat beam and the axial impact force respectively, and they are both obtained through the FEM simulation. The maximal generation numbers for outer and inner IP-GA are both specified as 300. The normalization factors φ and ψ are set to 1.9, 2.4 respectively. The allowable errors ε 1 = 0.5 and ε 2 = 0.15. The satisfaction degree level λ and the penalty factor σ are set to 0.8 and 100,000 respectively.

50 initial samples are used to create the approximation models for the uncertain objective and constraint. The result is listed in Table 6. It can be found that the optimal design vector (2.00 mm, 2.75 mm, 34.98 mm)T is achieved with the penalty function value f p = 2.32 after 23 iterations. The interval for energy absorption due to the uncertain parameters is [8.30 kJ, 9.28 kJ], and the interval for average normal impact force is [51.10 kN, 69.74 kN]. Initial samples for objective and constraints can be obtained simultaneously by carrying out the same FEM model, while densified samples during the process should be computed independently. The total number of calling the FEM model is 96.

Table 6 Optimization results of closed-hat beam

6 Remarks and conclusions

In this paper, a new ANIP method based on a local-densifying method is suggested to promote the efficiency and accuracy of the conventional ANIP method. Under the same or even less computation cost, this new ANIP method has higher accuracy than the conventional ANIP methods. Consequently, high accuracy and low computational cost can be achieved simultaneously. This superiority is more prominent especially for complex problems. Besides, as we only care about the accuracy of approximations in some local regions influencing the solution, more samples are expected to spend there to obtain precise approximations. But on the other hand, if some main characteristics are out of capturing due to the relaxation of approximation quality in the unconcerned regions, the proposed method will reach the local optimum, which is also the main effect of initial sample size, although the local optimal design is often acceptable for practical engineering problems, especially when the systems are complicated.

For large-scale problems, the approximation errors tend to increase exponentially as the dimension increases. The approximation quality can not be guaranteed with a limited sample size. To pursue high accuracy of the approximation for high dimensional problems, a relatively large sample size should be used, which is sometimes unacceptable for practical applications. Nevertheless, a significant improvement of the efficiency and accuracy can still be achieved by the proposed method. Furthermore, if the nonlinearity of the function becomes too high, the local densifying method may have difficulty to identify the key local regions and the algorithm has to make great efforts to revise the approximation models. As a result, this makes this approach quite low efficient. In order to improve the present algorithm, we will consider these abovementioned problems in our future work.