1 Introduction

The multi-objective optimization (MOO) problems widely exist in engineering design (Zarchi and Attaran 2019; Tian et al. 2018; Wang et al. 2011; Jaouadi et al. 2020), and a large number of optimization schemes have been provided by different MOO methods. A series of prominent work in this filed has been carried out and reported. Omkar et al. (2011) proposed a vector evaluated artificial bee colony (VEABC) algorithm to deal with the multi-objective design optimization of the laminated composite components. Bui et al. (2019) suggested a multi-objective optimization method based on the mixed integer linear programming (MILP) to determine trade-off between conflicting operation objectives of wind farm systems. Posteljnik et al. (2016) proposed a multi-objective optimization method for the wind turbine structure by adopting the particle swarm (PSO) algorithm. In the abovementioned works, the multi-objective optimization problems are mostly formulated by explicit mathematical expressions. However, the objective functions and constraints are usually black-box functions for most MOO problems, which are evaluated by complex simulation model with high computational costs and poor efficiency (Liu et al. 2015, 2017a, b; Hu et al. 2020). Therefore, an efficient method for solving the MOO problems should be developed.

To achieve better efficiency of the existing MOO method in dealing with the complex engineering problems, many scholars use the approximation model instead of the actual numerical simulation model such as Kriging (Choi et al. 2018), radial basis function (RBF) (Liew et al. 2004), and support vector regression (Lu and Roychowdhury 2008). Furthermore, a series of optimization methods based on the approximation model are further developed, which are extensively utilized in the optimal design of complex engineering problems. Kiani and Yildiz (2016) adopted the radial basis function to optimize the crashworthiness and NVH of the vehicle. Bu et al. (2018) employed the adaptive Kriging model to optimize the physical properties of the flywheel motor. Manshadi and Jamalinasab (2017) used response surface model to optimize the flap shape of unmanned aerial vehicle. In the reliability analysis (Liu et al. 2019a, 2020; Peng et al. 2018), the approximation model is also commonly used. Liu et al. (2019b) applied the radial basis function to perform the reliability analysis of complex structures. Zhang et al. (2017) adopted the response surface to evaluate the time-dependent reliability of uncertain structures. In order to further improve the accuracy of multi-objective optimization algorithm, the adaptive approximation model which contains updating management frameworks is also employed for the MOO problems. Yang et al. (2002) solved the multi-objective optimization problem of the I-beam structure by adopting the adaptive approximation models which are sequentially updated during the iterative optimization process. Fang et al. (2014) employed the multi-objective particle swarm optimization (MOPSO) algorithm for the crashworthiness design by using the adaptive approximation models, in which sequential sampling points are generated over the design space and the Kriging models are refitted in an iterative fashion. Jang et al. (2009) proposed an adaptive approximation framework and a convergence criterion of the adaptive approximation model to perform the multi-objective optimization for the full stochastic fatigue design problem. In the above works, the approximation model is widely used to solve engineering problems, but it is only the engineering application of approximation model and the method of constructing approximation model is not detailed discussed. Considering the accuracy of approximation model directly affects the optimization results of engineering problems, hence the method of constructing approximation model should be further researched.

In recent years, many studies have been proposed on how to achieve better efficiency and accuracy of the approximation models. Wang et al. (Wang et al. 2001; Wang 2003) developed the inherited Latin hypercube sampling (LHS) approach to strengthen the prediction ability and stability of the adaptive response surface. Cheng et al. (2015) proposed a trust region peak tracking sampling approach to enhance the ability of approximation model when dealing with the high-dimensional problems. Garud et al. (2017) suggested a united sampling criterion to enhance the overall accuracy of the global optimum in the approximation-based problems. As mentioned above works, the accuracy of approximation model is improved effectively based on the selection and layout of samples. It is noteworthy that the shape parameters of the approximation model, nonlinearity of objective functions, and constraints also affect the accuracy of the approximation model besides the samples. Although some scholars (Stolbunov and Nair 2018; Fasshauer and Zhang 2007; Long et al. 2016; Lee et al. 2008; Koupaei et al. 2018; Sarra and Sturgill 2009; Fornberg and Piret 2008) have studied the selection methods of the shape parameters in the constructing approximation model process, these methods determine the shape parameters by samples and use the same shape parameters to construct approximation models of different nonlinear optimization objective functions and constraints, which will inevitably cause the error of approximation model. Considering the existence of different nonlinear optimization objective functions and constraints, therefore, the shape parameters should be determined according to specific optimization objective functions and constraints rather than only according to samples. Meanwhile, the prediction ability of the approximation model cannot be guaranteed by the construction of the primary approximation model. Therefore, further studies for updating approximation model are necessary. This method should determine the location and direction of the distribution samples through the model management technology, so as to continuously update the samples and modify the approximation model until the accuracy requirements are reached. Hence, for promoting the MOO method into practical applications, the high-efficiency MOO methods with approximation model need to be developed.

For the above reasons, an efficient MOO method based on the adaptive approximation model is developed and applied to the practical engineering applications in this paper. The frame of this paper is organized as follows: the MOO problem and the related definitions are first introduced in Section 2. Then, an efficient MOO method based on the adaptive approximation model is detailed and discussed in Section 3. In Section 4, two numerical examples and two engineering examples are investigated to demonstrate the effectiveness of the present method. Finally, some conclusions are summarized in Section 5.

2 Statement of the multi-objective optimization problem

A typical description of multi-objective optimization (MOO) problem is as follows:

$$ {\displaystyle \begin{array}{c}\underset{\mathrm{X}}{\min }\ {f}_{\mathrm{l}}\left(\mathbf{X}\right)\kern0.33em ,\kern0.75em \mathrm{l}=1,2,\dots, \mathrm{a}\\ {}\mathrm{s}.\mathrm{t}.\kern0.5em {h}_u\left(\mathbf{X}\right)=0,\mathrm{u}=1,\dots, \mathrm{b}\\ {}{g}_v\left(\mathbf{X}\right)\le 0,\mathrm{v}=1,\dots, \mathrm{c}\\ {}\mathrm{X}={\left({X}_1,{X}_d,\dots, {X}_n\right)}^{\mathbf{T}},d=1,\dots, n\end{array}} $$
(1)

where X = (X1, Xd, …, Xn)T stands for the n-dimensional vector which consists of design variables. fl (X) denotes the objective function. hu (X) and gv (X) represent the equality constraint and the inequality constraint, respectively. During the process of solving the abovementioned multi-objective problems, a series of feasible solutions can be solved. Xξ and Xψ are individual solutions in the feasible set ΩS, namely, Xξ, Xψ ∈ ΩS. If the objectives corresponding to the solution Xξ are partially superior to the objectives corresponding to the solution Xψ, it is defined that the solution Xξ dominates the solution Xψ, written as \( {\mathbf{X}}_{\xi}\underset{\hbox{---} }{\prec }{\mathbf{X}}_{\psi } \). If the objectives corresponding to the solution Xξ are all superior to the objectives corresponding to the solution Xψ, it is defined that the solution Xξ strongly dominates the solution Xψ, written as Xξ ≺ Xψ. If X ∈ ΩS and there is no X ∈ ΩS dominating X, the solution X is defined as the Pareto optimal solution or non-dominated solution of the multi-objective optimization problem.

The Pareto optimal solutions in the feasible set ΩS are considered to be equally important unless further information is provided from the decision-maker. These Pareto optimal solutions are defined as the Pareto optimal set or non-dominated set. The Pareto optimal set converge into a curve in the feasible design space which is defined as the Pareto optimal frontier. The aim of the multi-objective optimization algorithm is to find the Pareto optimal set to form the Pareto frontier. However, the challenge to the most MOO problems is that the objective functions and constraints are usually implicitly functions, which are evaluated by complex simulation model with high computational costs and low efficiency. Therefore, a multi-objective optimization method with high efficiency should be developed.

3 The multi-objective optimization method based on the adaptive approximation model

As mentioned above, during the process of solving multi-objective optimization problems, complex function evaluations may cause a high computational cost and low efficiency. Thus, an efficient MOO method based on the adaptive approximation model will be developed and discussed in the following contents.

3.1 Radial basis function (RBF)

Radial basis function is a linear combination of radial functions with respect to the Euclidean distance between the predict points and the sample points (Amouzgar and Strömberg 2017; Zhang et al. 2020), which can be used to approximate complex simulation model or black-box function. The RBF can be defined as:

$$ \tilde{f}\left(\mathbf{X}\right)=\sum \limits_{i=1}^N{w}_ih\left({r}_i\right) $$
(2)

where \( \tilde{f}\left(\mathbf{X}\right) \) denotes the response value of approximate function. N is the number of samples which are generated by the experimental design such as the Latin hypercube design (LHD) (Park 1994; Joseph et al. 2015; Husslage et al. 2011). wi represents the coefficient of the linear combinations. h(ri) is the basis function, and ri = ‖X − Xi‖is the Euclidean distance between the predict point and the samples.

When N samples are obtained, the corresponding RBF model can be expressed as:

$$ \mathbf{f}=\mathbf{hw} $$
(3)

where f denotes the N-dimensional response vector of samples. w is the coefficient vector. h represents N × N dimensional matrix, which is described as follows:

$$ \mathbf{h}=\left(\begin{array}{ccc}h\left(\left\Vert {\mathbf{X}}_1-{\mathbf{X}}_1\right\Vert \right)& \cdots & h\left(\left\Vert {\mathbf{X}}_1-{\mathbf{X}}_N\right\Vert \right)\\ {}\vdots & \ddots & \vdots \\ {}h\left(\left\Vert {\mathbf{X}}_N-{\mathbf{X}}_1\right\Vert \right)& \cdots & h\left(\left\Vert {\mathbf{X}}_N-{\mathbf{X}}_N\right\Vert \right)\end{array}\right) $$
(4)

And the coefficient vector w can be calculated by Eq. (5):

$$ \mathbf{w}={\mathbf{h}}^{-1}\mathbf{f} $$
(5)

When the coefficient vector w obtained is substituted into Eq. (2), the approximation model based on RBF is finally constructed. It is noteworthy that the prediction accuracy of the approximation model is significantly affected by the selection of the basic function which is listed in Table 1.

Table 1 Different basis functions

Due to the strong applicability of Gaussian function and its wide application in approximation models, the Gaussian function is chosen for a representative of the basis functions in this paper. The Gaussian function is expressed as follows:

$$ h\left({r}_i\right)={e}^{-\varepsilon {r_i}^2},{r}_i=\left\Vert \mathbf{X}-{\mathbf{X}}_i\right\Vert $$
(6)

where ε denotes the shape parameter. In the process of constructing approximation model, there is usually no obvious formula to describe the intrinsic connections between the shape parameter and the sample points. Moreover, the accuracy of approximation model will be directly affected by the setting of shape parameters. Thus, it is necessary to develop a more effective method for setting shape parameter to ensure the prediction ability and accuracy of the approximation model.

3.2 Reverse shape parameter analysis method (RSPAM)

According to the construction principle of the above approximation model, the approximation response value at the sample is equal to the actual response value. However, there is a certain error of the approximation model at the predict point. When the initial samples are fixed, the error is related to the selected basis function and its shape parameters. Therefore, a reverse shape parameter analysis method based on the RBF is developed in this section. Based on this method, the accuracy and efficiency of the approximation models will be greatly enhanced. The following is discussed in detail.

Firstly, N samples ΩX = {X1, …, Xj, …, XN}, j = 1, 2, …, N containing the key point Xj are given by Latin hypercube design (LHD). Meanwhile, M adjacent sample points \( {\Omega}_{{\mathbf{X}}^K}=\left\{{\mathbf{X}}_1^K,\dots, {\mathbf{X}}_p^K,\dots, {\mathbf{X}}_M^K\right\} \),p = 1, 2, …, M and \( {\Omega}_{{\mathbf{X}}^K}\subseteq {\Omega}_{\mathbf{X}} \), near the key point Xj are used for constructing the mini RBF:

$$ \tilde{f}\left(\mathbf{X}\right)=\sum \limits_{p=1}^Mh\left({r}_p\right)\ {w}_p=\sum \limits_{p=1}^M{e}^{-\varepsilon {\left(\left\Vert \mathbf{X}-{\mathbf{X}}_p^K\right\Vert \right)}^2}{w}_p $$
(7)

In Eq. (7), the coefficient vector w = (w1, …, wM)Tand the shape parameter ε are two unknown variables which need to be solved. Based on Eq. (5), the coefficient vector w = (w1, …, wM)T can be expressed as Eq. (8):

$$ \mathbf{w}={\left({w}_1,\dots, {w}_M\right)}^{\mathrm{T}}={\left(\begin{array}{ccc}h\left(\ \left\Vert {\mathbf{X}}_1^K-{\mathbf{X}}_1^K\right\Vert\ \right)& \cdots & h\left(\ \left\Vert {\mathbf{X}}_1^K-{\mathbf{X}}_M^K\right\Vert\ \right)\\ {}\vdots & \ddots & \vdots \\ {}h\left(\ \left\Vert {\mathbf{X}}_M^K-{\mathbf{X}}_1^K\right\Vert\ \right)& \dots & h\left(\ \left\Vert {\mathbf{X}}_M^K-{\mathbf{X}}_M^K\right\Vert\ \right)\end{array}\right)}^{-1}\left(\begin{array}{c}f\left({\mathbf{X}}_1^K\right)\\ {}\vdots \\ {}f\left({\mathbf{X}}_M^K\right)\end{array}\right) $$
(8)

Secondly, substituting the key point Xj into the above mini RBF Eq. (7) and assuming that the approximate function value \( \tilde{f}\left({\mathbf{X}}_j\right) \)at the key point Xj is equal to the actual valuef (Xj), namely:

$$ {\displaystyle \begin{array}{l}f\left({\mathbf{X}}_j\right)=\tilde{f}\left({\mathbf{X}}_j\right)\\ {}=\sum \limits_{p=1}^Mh\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_p^K\right\Vert \right)\ {w}_p=\sum \limits_{p=1}^M{e}^{-\varepsilon {\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_p^K\right\Vert \right)}^2}{w}_p\\ {}=\left(h\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_1^K\right\Vert \right)\kern0.5em \cdots \kern0.5em h\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_M^K\right\Vert \right)\right)\bullet {\left({w}_1,\dots, {w}_M\right)}^{\mathrm{T}}\end{array}} $$
(9)

Then, the equation for solving the shape parameter can be obtained by substituting Eq. (8) into Eq. (9) as follows:

$$ {\displaystyle \begin{array}{c}f\left({\mathbf{X}}_j\right)=\overset{\sim }{f}\left({\mathbf{X}}_j\right)\\ {}=\left(h\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_1^K\right\Vert \right)\kern0.5em \cdots \kern0.5em h\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_M^K\right\Vert \right)\right)\bullet {\left({w}_1,\dots, {w}_M\right)}^{\mathrm{T}}\\ {}=\left(\begin{array}{ccc}h\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_1^K\right\Vert \right)& \cdots & h\left(\left\Vert {\mathbf{X}}_j-{\mathbf{X}}_M^K\right\Vert \right)\end{array}\right)\bullet {\left(\begin{array}{ccc}h\left(\left\Vert {\mathbf{X}}_1^K-{\mathbf{X}}_1^K\right\Vert \right)& \cdots & h\left(\left\Vert {\mathbf{X}}_1^K-{\mathbf{X}}_M^K\right\Vert \right)\\ {}\vdots & \ddots & \vdots \\ {}h\left(\left\Vert {\mathbf{X}}_M^K-{\mathbf{X}}_1^K\right\Vert \right)& \dots & h\left(\left\Vert {\mathbf{X}}_M^K-{\mathbf{X}}_M^K\right\Vert \right)\end{array}\right)}^{-1}\left(\begin{array}{c}f\left({\mathbf{X}}_1^K\right)\\ {}\vdots \\ {}f\left({\mathbf{X}}_M^K\right)\end{array}\right)\end{array}} $$
(10)

Based on the above analysis, only one unknown variable of shape parameter ε is included, which can be solved by Eq. (10). It is worth noting that a key point and its adjacent samples can be used to obtain a corresponding shape parameter value by solving Eq. (10). Hence, taking N samples ΩX = {X1, …, Xj, …, XN}, j = 1, 2, …, N as key points in turn, Ncorresponding shape parameters Ωε = {ε1, …, εj, …, εN}, j = 1, 2, …, Ncan be obtained by solving the above steps. In this paper, these shape parameters Ωε = {ε1, …, εj, …, εN} are called “candidate shape parameters.”

Thirdly, N candidate shape parameters Ωε = {ε1, …, εj, …, εN} are used to construct N test approximation models, respectively, and the error analysis is also performed for every test approximation model. The candidate shape parameter with the minimum error is identified as the optimal shape parameter for constructing the approximation model. The detailed process is as follows:

The samples ΩX = {X1, …, Xj, …, XN}are divided into two parts according to the sample number. The first half of the samples {X1, …, XN/2} are used as the “the construction points,” and the second half of the samples \( \left\{{\mathbf{X}}_{\frac{N}{2}+1},\dots, {\mathbf{X}}_N\right\} \) are chosen as the “the test points.” The construction points are combined with each candidate shape parameter εj to construct corresponding approximation model, and the test points are adopted to evaluate the prediction ability of the approximation model through the error criterion σj defined by Eq. (11).

$$ {\sigma}_j=\frac{\sum \limits_{t=\frac{N}{2}+1}^N\left|\frac{f\left({\mathbf{X}}_t\right)-\overset{\sim }{f}\left({\mathbf{X}}_t\right)}{f\left({\mathbf{X}}_t\right)}\right|}{\frac{N}{2}}\kern0.75em ,\kern0.75em j=1,2,\dots, N,t=\frac{N}{2}+1,\dots, N $$
(11)

wheref (Xt) denotes the actual value at the test point and \( \tilde{f}\left({\mathbf{X}}_t\right) \) represents the approximate value of the function at the test point. After all the errors corresponding to candidate shape parameters are obtained, the shape parameter with the minimum error is chosen for the optimal shape parameter.

Finally, the samples ΩX = {X1, …, XN} are combined with the optimal shape parameter to construct the final approximation model.

3.3 Local-densifying approximation method

When the approximation models of objective functions and its corresponding constraints are constructed based on the RBF, the MOO problem in Eq. (1) becomes the following approximate MOO problem:

$$ {\displaystyle \begin{array}{c}\underset{\mathbf{X}}{\min }\ {\tilde{f}}_l\left(\mathbf{X}\right),l=1,2,\dots, a\\ {}s.t.{\tilde{h}}_u\left(\mathbf{X}\right)=0,u=1,\dots, b\\ {}{\tilde{g}}_v\left(\mathbf{X}\right)\le 0,v=1,\dots, c\\ {}\mathbf{X}={\left({X}_1,{X}_d,\dots, {X}_n\right)}^{\mathbf{T}},d=1,\dots, n\end{array}} $$
(12)

where \( {\tilde{f}}_l\left(\mathbf{X}\right) \), \( {\tilde{h}}_u\left(\mathbf{X}\right) \), and \( {\tilde{g}}_v\left(\mathbf{X}\right) \) represent approximation models of the objective function, the equality constraint, and the inequality constraint, which are explicit functions with respect to the design vector X.

For the approximation MOO problem in Eq. (12), the Pareto optimal set of objective functions and their corresponding constraint function values are primarily concerned. Considering the accuracy of the abovementioned approximate models depend largely on the samples in the experimental design space, hence, it is necessary to arrange the samples more reasonably so that the limited samples are more distributed near the region where the Pareto optimal set are found. When the prediction ability and accuracy of the approximation model are ensured in the local area where the Pareto optimal set are found, more reliable optimization results can be obtained.

For these reasons, the local-densifying approximation method is applied to enhance the local precision of approximation models in the Pareto optimal set, which is the main mechanism of the adaptive approximation model. The local-densifying approximation method is an updating strategy of sampling method which concentrates limited sample resources to the local regions we concerned, namely, more samples are expected to be densified into the local regions where the Pareto optimal set are found. In the local-densifying approximation method, the Pareto optimal set are obtained and evaluated in each iteration step. According to the approximation values at the optimal Pareto set, the approximation models are updated by adding local-densified samples to the sample space for next iteration until the stop criteria are satisfied. Therefore, it ensures that the approximation models have small approximate errors in the Pareto optimal set and the result can be more reliable.

3.4 Computational procedure

In this method, the micro multi-objective genetic algorithm (μMOGA) (Liu et al. 2012) is applied to solve the MOO problem in Eq. (12) due to the μMOGA has an outstanding performance in searching the Pareto optimal set. The computational procedure of the proposed MOO method is given in Fig. 1 and the details are outlined as:

  1. (1)

    Sampling in the design vector space by the LHD. Set the allowable error δ and initial iteration step s = 1.

  2. (2)

    According to the above initial samples, computing the actual response values of the objective function fl(X)and constraints hu(X) and gv(X) evaluated by the actual simulation model.

  3. (3)

    Obtaining the optimal shape parameter by using Reverse shape parameter analysis method.

  4. (4)

    Constructing approximation models of the objective function and constraints and obtaining the approximate MOO problem as Eq. (12).

  5. (5)

    Solving the approximate MOO problem Eq. (12) by μMOGA and obtaining the Pareto optimal set of the approximation objective function \( \left\{{\tilde{f}}_l^{(s)}\left({\mathbf{X}}^{(z)}\right)\right\} \), z = 1, 2, …, t, where \( {\tilde{f}}_l^{(s)}\left({\mathbf{X}}^{(z)}\right) \) stands for the zth Pareto solution of the Pareto optimal set in step s; Similarly, their corresponding approximation constraint function values \( {\tilde{h}}_u^{(s)}\left({\mathbf{X}}^{(z)}\right) \) and \( {\tilde{g}}_v^{(s)}\left({\mathbf{X}}^{(z)}\right) \) at optimal solution X(z) also can be obtained.

  6. (6)

    Based on the simulation model, computing the actual response values of the objective function \( \left\{{f}_l^{(s)}\left({\mathbf{X}}^{(z)}\right)\right\} \), z = 1, 2, …, t, and their corresponding actual constraint function values \( {h}_u^{(s)}\left({\mathbf{X}}^{(z)}\right) \) and \( {g}_v^{(s)}\left({\mathbf{X}}^{(z)}\right) \).

  7. (7)

    Calculating the maximum error Δmax:

Fig. 1
figure 1

Flowchart of the MOO method based on the adaptive approximation model

$$ {\Delta}_{\mathrm{max}}=\max \left\{{\Delta}_z\right\},\kern0.5em z=1,2,\dots, t $$
$$ {\boldsymbol{\varDelta}}_z=\max \left\{\left|\frac{f_l^{(s)}\left({\mathbf{X}}^{(z)}\right)-\tilde{f}_{l}^{(s)}\left({\mathbf{X}}^{(z)}\right)}{f_l^{(s)}\left({\mathbf{X}}^{(z)}\right)}\right|,\left|\frac{h_u^{(s)}\left({\mathbf{X}}^{(z)}\right)-\tilde{h}_{u}^{(s)}\left({\mathbf{X}}^{(z)}\right)}{h_u^{(s)}\left({\mathbf{X}}^{(z)}\right)}\right|,\left|\frac{g_v^{(s)}\left({\mathbf{X}}^{(z)}\right)-\tilde{g}_{v}^{(s)}\left({\mathbf{X}}^{(z)}\right)}{g_v^{(s)}\left({\mathbf{X}}^{(z)}\right)}\right|\right\} $$

If Δmax ≤ δ, then \( \left\{{f}_l^{(s)}\left({\mathbf{X}}^{(z)}\right)\right\} \) and z = 1, 2, …, t are determined as the final solution and the iteration terminates. Otherwise, updating the approximation models by adding the optimal set {X(z)} with error more than δ to the sample space. Set s = s + 1 and move to step (4).

4 Numerical examples and discussion

In this section, the accuracy of optimization results and the efficiency of the algorithm are tested by two numerical examples. Meanwhile, the proposed method is also performed to demonstrate its practical engineering applicability through two actual engineering examples.

4.1 Numerical example A

In the numerical example A, a two-objective optimization problem with two constraints is as follows:

$$ {\displaystyle \begin{array}{c}\underset{X}{\min }{f}_1\left(\mathbf{X}\right)=-{X}_1{X}_2+{X}_3{X}_4/2\\ {}{f}_2\left(\mathbf{X}\right)=2{X}_1{X}_2/{X}_4^2+3{X}_2{X}_4/{X}_3^2\\ {}s.t.{g}_1\left(\mathbf{X}\right)=10-2{X}_1-{X}_2/{X}_3+{X}_4\le 0\\ {}{g}_2\left(\mathbf{X}\right)=65-{X}_2+{X}_1/{X}_4-5{X}_3\le 0\\ {}\mathbf{X}={\left({X}_1,{X}_2,{X}_3,{X}_4\right)}^T,10\le {X}_1,{X}_2,{X}_3,{X}_4\le 20\end{array}} $$
(13)

where f1(X) and f2(X) represent the objective function 1 and the objective function 2, respectively. g1(X) and g2(X) are the inequality constraints. The X stands for the design vector. The allowable error δ is set to 0.1. The μMOGA and local-densifying approximation method are also applied to solve the Pareto optimal set. The preset parameters for the μMOGA are as follows: the maximum generation is 200, the population size is 5.0, and the probability of crossover is 0.5. To analyze the influence of the shape parameters for the optimization results, the abovementioned problem is investigated for two cases.

4.1.1 The influence of the number of adjacent sample points

In this case, the influence of the number of adjacent sample points M on the optimization results is discussed. Meanwhile, the local-densifying approximation method is adopted to improve the accuracy of Pareto optimal set. The fifty initial samples are generated by the LHD and the number of adjacent sample points M is varied between 4 and 6 in steps of 1. The proposed RSPAM is used to solve the shape parameters of the RBFs, and the corresponding approximation models of the objective functions and constraints are constructed based on the initial samples and these shape parameters. The obtained shape parameters of the approximation models under different M values are listed in Table 2, and the iterative process with optimization results are presented in Figs. 2, 3, and 4. Simultaneously, the numbers of samples in the iterative process and the corresponding maximum error are detailed in Table 3.

Table 2 The shape parameters under different M values (numerical example A)
Fig. 2
figure 2

The iterative process and optimization results under M = 4 (numerical example A)

Fig. 3
figure 3

The iterative process and optimization results under M = 5 (numerical example A)

Fig. 4
figure 4

The iterative process and optimization results under M = 6 (numerical example A)

Table 3 The count of samples in the iterative process and the corresponding maximum error (numerical example A)

According to Fig. 2, when M = 4, the maximum error of the functions at the Pareto optimal set is 13.74% in step 1. The iterative process indicates that the prediction ability of the approximation model is improved through adding local-densified samples to the sample space, and in step 3, the corresponding maximal error of the objective functions and constraint functions is found to be 4.65%. The result implies that the Pareto optimal set satisfying the allowable error are obtained by three iterative steps. While the number of adjacent sample points M is set to M = 5 and M = 6, the final Pareto optimal set satisfying the allowable error are obtained by one iterative step which are presented in Figs. 3 and 4. The corresponding maximum errors are 3.0% and 7.97%, respectively. According to the above computational results, when M = 4, more iterative steps are needed to obtain the final Pareto optimal set because the insufficient sample information is available to solve shape parameters. Meanwhile, the maximum error of M = 6 is more than that of M = 5 because the ill-conditioned matrix appears many times in the solving process of shape parameters. It is obvious that the selection of the number of adjacent sample points M will directly affect the ability of the algorithm to obtain the expected optimization results. According to the above optimization results, when the number of adjacent sample points M is varied between 4 and 6, five adjacent sample points (M = 5) will be helpful to improve the accuracy of optimization results and the efficiency of the algorithm in this representative case.

4.1.2 The computational efficiency and optimization accuracy

This case is the verification of the proposed MOO method in the computational efficiency and optimization accuracy. The RSPAM is compared with Long’s method (Long et al. 2016) and Koupaei’s method (Koupaei et al. 2018) for calculating the shape parameters of the RBFs. In the proposed RSPAM, five adjacent sample points (M = 5) could be recommended to solve the optimal shape parameter according to the above optimization results. In the Long’s method, the formula of solving shape parameters εL is expressed as:

$$ {\varepsilon}_L={\left(\left(\max \left(\mathbf{X}\right)-\min \left(\mathbf{X}\right)\right)/N\right)}^{\frac{1}{n}} $$
(14)

where (max(X) − min(X)) represents the longest distance in the sample points. N and n stand for the number of sample points and the dimension of design variables, respectively.

In Koupaei’s method, the shape parameter εk can be solved by the following optimization problem:

$$ \underset{\varepsilon_k}{\min } RMS\left({\varepsilon}_k\right)=\sqrt{\frac{1}{\varpi}\sum \limits_{i=1}^{\varpi }{\left|f\left({\mathbf{X}}_i\right)-\tilde{f}\left({\mathbf{X}}_i,{\varepsilon}_k\right)\right|}^2} $$
(15)

where RMS(εk) represents the root mean square (RMS) error with respect to the shape parameter εk. f(Xi) and \( \tilde{f}\left({\mathbf{X}}_i,{\varepsilon}_k\right) \) denote the actual function and approximate function, respectively. Xi, i = 1, …, ϖ is the test point, and ϖ stands for the number of the test points. The number of initial samples is 50, and the proposed RSPAM (M = 5), Long’s method, and Koupaei’s method are used to solve the shape parameters of the RBFs as listed in Table 4, respectively. Then, the corresponding approximation models of the objective functions and constraint are constructed based on these samples and shape parameters.

Table 4 The shape parameters based on different methods (numerical example A)

The iterative process and optimization results under the methods of Long and Koupaei are shown in Figs. 5 and 6. The count of samples in the iterative process and the corresponding maximum error are detailed in Table 5. It is obvious that the proposed method and Koupaei’s method can obtain the Pareto optimal set satisfying the allowable error in step 1. The corresponding maximum errors of the Pareto optimal set are 3.0% and 6.46%, respectively. However, the errors of the Pareto optimal set in step 1 are not less than allowable error 10% by using the Long’s method. The corresponding maximal error of the objective functions and constraint functions is found to be 28.90%, which suggests the approximation models are relatively coarse by using the methods of the Long’s method. Through adding local-densified samples to the sample space in the iterative process, the maximum error is 8.80% in step 4 for the Long’s method, which demonstrates the optimization results have reached the allowable error requirement after 4 steps. The optimization results show that the Pareto optimal set obtained by the proposed MOO method in step 1 are obviously better than that obtained by the methods of Koupaei and Long. The results also indicate that the optimal shape parameters obtained by the RSPAM could ensure the accuracy of the approximation model at the initial step.

Fig. 5
figure 5

The iterative process and optimization results based on the Long’s method (numerical example A)

Fig. 6
figure 6

The iterative process and optimization results based on the Koupaei’s method (numerical example A)

Table 5 The count of samples in the iterative process and the corresponding maximum error (numerical example A)

From the calculation results of three different methods, the proposed method could obtain the optimal solutions more efficiently. This is because the Koupaei’s method mainly focuses on the minimum RMS and ignores the internal relationship of the initial sample points. Simultaneously, the Long’s method only solve the shape parameters through the number of sample points and design variables, which uses the same shape parameters to construct approximation models, namely, the characteristics of the nonlinear objective functions and constraints are not considered in the Long’s method. However, the proposed method solves the corresponding shape parameters for different nonlinear objective functions and constraints based on the internal relation of the initial sample points, namely, the different shape parameters are used to construct the approximation models of the nonlinear objective functions and constraints than only the number of sample points and design variables are considered.

4.2 Numerical example B

An optimization design problem of I-beam is illustrated in Fig. 7. The length L of the I-beam is 200 cm. Two loads are applied to the I-beam as shown in the Fig. 7, which are P = 600kN and Q = 50kN, respectively. The Young’s modulus of the I-beam is E = 200GPa. The above MOO problem can be defined as minimizing the cross-sectional area and vertical deflection of the I-beam under the constraint of principal stress below the allowable stress σ = 6 kN/cm2. The parameters of the I-beam are considered as the design variables which areX1,X2,X3, andX4. Therefore, the above MOO problem is formulated as:

$$ {\displaystyle \begin{array}{c}\underset{X}{\min}\left({f}_1\left(\mathbf{X}\right),{f}_2\left(\mathbf{X}\right)\right)\\ {}{f}_1\left(\mathbf{X}\right)=2{X}_2{X}_4+{X}_3\left({X}_1-2{X}_4\right)\\ {}{f}_2\left(\mathbf{X}\right)=\frac{P{L}^3}{48 EI}=\frac{5000}{\frac{1}{12}{X}_3{\left({X}_1-2{X}_4\right)}^3+\frac{1}{6}{X}_2{X}_4^3+2{X}_2{X}_4{\left(\frac{X_1-{X}_4}{2}\right)}^2}\\ {}\begin{array}{c}s.t.g\left(\mathbf{X}\right)=\frac{180000{X}_1}{X_3{\left({X}_1-2{X}_4\right)}^3+2{X}_2{X}_4\left[4{X}_4^2+3{X}_1\left({X}_1-2{X}_4\right)\right]}\\ {}+\frac{15000{X}_2}{\left({X}_1-2{X}_4\right){X}_3^3+2{X}_4{X}_2^3}\le 6\end{array}\\ {}\mathbf{X}={\left({X}_1,{X}_2,{X}_3,{X}_4\right)}^T\\ {}70\mathrm{cm}\le {X}_1\le 90\mathrm{cm},40\mathrm{cm}\le {X}_2\le 60\mathrm{cm},0.5\mathrm{cm}\le {X}_3\le 2\mathrm{cm},1\mathrm{cm}\le {X}_4\le 3\mathrm{cm}\end{array}} $$
(16)

where f1(X) and f2(X) represent the cross-sectional area and vertical deflection of the I-beam, respectively. g(X) is the principal stress. In the optimization process, the allowable error δ is set to 0.1. The Pareto optimal set is found by using the μMOGA and local-densifying approximation method. The parameters for μMOGA are preset as follows: the maximum generation is 200, the population size is 5.0, and the probability of crossover is 0.5, respectively. To discuss the influence of the shape parameters for the optimization results, the abovementioned MOO problem is investigated for two cases.

Fig. 7
figure 7

The structure of I-beam

4.2.1 The influence of the number of adjacent sample points

In this case, the shape parameters of the RBFs are solved by the proposed RSPAM and the accuracy of Pareto optimal set is ensured by using the local-densifying approximation method. The number of adjacent sample points M is set to 4, 5, and 6, respectively. The shape parameters under different M values are presented in Table 6 and the iterative process with optimization results are displayed in Figs. 8, 9, and 10. For better illustration of the influence on the optimization results under different number of adjacent sample points M, the count of samples in the iterative process and the maximum error in each step are detailed in Table 7.

Table 6 The shape parameters under different M values (numerical example B)
Fig. 8
figure 8

The iterative process and optimization results under M = 4 (numerical example B)

Fig. 9
figure 9

The iterative process and optimization results under M = 5 (numerical example B)

Fig. 10
figure 10

The iterative process and optimization results under M = 6 (numerical example B)

Table 7 The count of samples in the iterative process and the corresponding maximum error (numerical example B)

In step 1, the approximation models are constructed by using 50 initial samples which are generated by the LHD. The corresponding maximal error of the objective functions and constraint functions are found to be 38.30%, 11.96%, and 18.15% under three different number of adjacent sample points, respectively. The results indicate that the local-densifying approximation method has to be employed for strengthening the accuracy of approximation models. In step 2, The Pareto optimal set with 5 adjacent sample points has reached the allowable error requirement and the maximal error of the objective functions and constraints is found to be 7.22%. However, the Pareto optimal set of the other two adjacent sample points still fails to meet the allowable error requirement. Through adding local-densified samples to the sample space, both M = 4 and M = 6 reach the allowable error requirements in step 3. The corresponding maximum error is 6.60% and 7.24%, respectively.

According to the above results, the accuracy and efficiency of the optimization are greatly influenced by the number of adjacent sample points M. Generally, the accuracy of optimization results will be relatively low when the fewer adjacent sample points are used to solve Eq. (10). When more adjacent sample points are selected, the accuracy of optimization results may be improved but ill-conditioned matrices may appear in the process of solving Eq. (10) which will lead to the failure of solution and low efficiency. In the above cases, M = 4 to 6 are investigated and the optimization results demonstrate that the accuracy and efficiency of the proposed method can be well satisfied when M = 5. It is noteworthy that the number of adjacent sample points M could also be determined by a trade-off between the accuracy and efficiency from the operator. Meanwhile, more cases of different M values have been included in the future work.

4.2.2 The computational efficiency and optimization accuracy

The reverse shape parameter analysis method (RSPAM, M = 5), Long’s method, and Koupaei’s method are used again to solve the shape parameters of the RBFs, respectively. The shape parameters obtained by the different methods are listed in Table 8. Based on these shape parameters and the initial samples, the corresponding approximation models of the objective functions and constraints are constructed. The other preset parameters keep unchanged. The iterative process and optimization results under the methods of Long and Koupaei are presented in Figs. 11 and 12. Meanwhile, the count of samples in the iterative process and the corresponding maximum error are detailed in Table 9.

Table 8 The shape parameters based on different methods (numerical example B)
Fig. 11
figure 11

The iterative process and optimization results based on the Long’s method (numerical example B)

Fig. 12
figure 12

The iterative process and optimization results based on the Koupaei’s method (numerical example B)

Table 9 The count of samples in the iterative process and the corresponding maximum error (numerical example B)

Based on the above optimization results, the maximal error of the objective functions and constraint functions based on the proposed method is 11.96% in step 1, which is very close to the allowable error value. It indicates that the optimal shape parameters obtained by the RSPAM could ensure the accuracy of the approximation model at the beginning of iteration. With the increasing of sample points by employing local-densifying approximation method, the Pareto optimal set satisfying the allowable error are obtained by the proposed MOO method in step 2 and the corresponding maximal error of the objective functions and constraint functions is found to be 7.22%, which is obviously better than other methods. While the errors of the other methods are not less than the allowable error 10% in step 2, which shows that the methods of Long and Koupaei need more iterative steps to obtain the Pareto optimal set for meeting the accuracy requirements. By using the local-densifying approximation method, the maximal error of the objective functions and constraint functions for the Koupaei’s method is 6.59% in step 3, which indicates the optimization results have reached the allowable error requirement after 3 steps. However, the corresponding maximal error based on the Long’s method is 22.26% in step 5, which demonstrates the Pareto optimal set still fail to meet the allowable error requirement. This is because the Long’s method only consider the influence of the number of sample points and design variables, so that the shape parameters of three approximation models \( {\tilde{f}}_1\left(\mathbf{X}\right) \), \( {\tilde{f}}_2\left(\mathbf{X}\right) \), and \( \tilde{g}\left(\mathbf{X}\right) \)are the same. It induces the accuracy of the approximation models needs to be improved by more iterative steps, while the RSPAM fully considers the influence of different objective functions and constraints based on the internal relation of the initial sample points, and the corresponding approximation models are constructed by using the different shape parameters. Therefore, the proposed method could get more accurate approximation model and solve the optimal solutions more quickly than other two methods.

4.3 Engineering example A

Gear transmission is one of the most common drive forms of mechanical transmission system. The gear transmission structure will directly affect the transmission efficiency of mechanical system. Hence, the proposed method is used to design a practical gear transmission structure. As shown in Fig. 13, the numerical model of the gear transmission structure is constructed, which consists of main gear and pinion. Minimizing the spokes volume and the bending stress of the main gear are two objectives for this design problem. During the transmission process, the spoke thickness t, spoke width W, and spoke radius r will play important roles in the bending stress of the main gear. Thus, the spoke thickness t, spoke width W, and spoke radius r are considered as the design variables. The other parameters in the gear system are as follows: the modulus of the gear is m = 4, the diameter of pinion is 108 mm, the diameter of gear is 324 mm, the material of the gear is 45# steel, the permissible bending stress is 325.08 MPa, the torque of pinion gear is T = 286500 N ⋅ mm, and the periphery force of the tooth surface is Ft = 5305.5 N. The optimization problem can be described as follows:

Fig. 13
figure 13

The structure of gear

$$ {\displaystyle \begin{array}{c}\underset{t,W,r}{\min}\left(P\left(t,W,r\right),V\Big(t,W,r\Big)\right)\\ {}5\mathrm{mm}\le t\le 10\mathrm{mm},20\mathrm{mm}\le W\le 80\mathrm{mm},40\mathrm{mm}\le r\le 65\mathrm{mm}\end{array}} $$
(17)

where P(t, W, r) stands for the bending stress of the main gear and V(t, W, r) is the spoke volume.

In the optimization process, the allowable error δ is set to 0.1. The μMOGA and local-densifying approximation method are also used to solve the Pareto optimal set. The preset parameters for μMOGA are as follows: the maximum generation is 200, and the population size and the probability of crossover are 5.0 and 0.5. The approximation models of functions P(t, W, r) and V(t, W, r)are constructed by using 30 initial samples which are generated by the LHD. Based on the proposed RSPAM (M = 5), the optimal shape parameters are solved for the approximation models, which are 0.5344 and 0.1827 for the approximation models \( \tilde{P}\left(t,W,r\right) \) and \( \tilde{V}\left(t,W,r\right) \), respectively. Under the help of the proposed method, the iterative process is displayed in Fig. 14 and the count of samples in the iterative steps are detailed in Table 10. It can be found in step 1, the Pareto optimal set of approximate functions\( \tilde{P}\left(t,W,r\right) \) and \( \tilde{V}\left(t,W,r\right) \) are far from the actual function P(t, W, r) and V(t, W, r), which obviously shows that the approximation models are not accurate enough. Through the local-densifying approximation method for three iterative steps, the maximal error of the objective functions has been reduced to 8.61%, which indicates the Pareto optimal sets have satisfied the allowable error. Hence, ten solutions are chosen from the Pareto solutions given in Table 11. The engineers will select the No.10 solution if the bending stress of the main gear is the primary concern. Alternatively, if more attention to the spoke volume is paid, the No.1 solution would be considered.

Fig. 14
figure 14

The iterative process and optimization results (engineering example A)

Table 10 The count of samples in the iterative process and the corresponding maximum error (engineering example A)
Table 11 The Pareto solutions satisfying the requirement (engineering example A)

4.4 Engineering example B

The forearm structure is one of the most important parts of the bucket wheel stacker-reclaimer. When the bucket wheel stacker-reclaimer is in operation, the 9600-kg bucket wheel is installed at the front end of the forearm structure. To enhance the safety performance of forearm structure, and reduce the weight to satisfy the requirements of lightweight design, the proposed MOO method is performed for optimizing the forearm design. The numerical model of the forearm structure is displayed in Fig. 15. In this optimization design problem, the maximum stress and the weight of the forearm structure are adopted as two optimization objectives. The maximum displacement of the forearm structure in the vertical direction is considered as the constraint. Four thickness parameters of the forearm structure are identified as the design variables. The material properties of the forearm structure are as follows: the density is 7.85g/cm3, the Young’s modulus is 200Gpa, the Poisson’s ratio is 0.2, and the permissible maximum stress is 230 MPa. Therefore, the optimization design problem can be described as follows:

Fig. 15
figure 15

The forearm structure of the bucket wheel stacker-reclaimer

$$ {\displaystyle \begin{array}{c}\underset{\mathbf{X}}{\min}\left(P\left(\mathbf{X}\right),M\left(\mathbf{X}\right)\right)\\ {}\mathrm{s}.\mathrm{t}.D\left(\mathbf{X}\right)\le 25\mathrm{mm}\\ {}\mathbf{X}={\left({X}_1,{X}_2,{X}_3,{X}_4\right)}^T\\ {}8\mathrm{mm}\le {X}_1,{X}_2,{X}_3,{X}_4\le 14\mathrm{mm}\end{array}} $$
(18)

where P(X) stands for the maximum stress of the forearm structure and M(X) is the total weight of the forearm. D(X) is the maximum displacement of the forearm in the vertical direction. X stands for the design vector: X1 represents the thickness of the girder web, X2 is the thickness of the diagonal beam, X3 is the thickness of the flange plate on the cross beam, and X4stands for the thickness of the flange plate on the girder.

In the optimization process, the allowable error δ is set to 0.05. The μMOGA and local-densifying approximation method are also used to solve the Pareto optimal set. The preset parameters for μMOGA are as follows: the maximum generation is 200, and the population size and the probability of crossover are 5.0 and 0.5. The approximation models of the objective functions P(X) and M(X) and constraint D(X)are constructed by using thirty initial samples which are generated by the LHD. Based on the proposed RSPAM (M = 5), The optimal shape parameters of approximation models \( \tilde{P}\left(\mathbf{X}\right) \), \( \tilde{M}\left(\mathbf{X}\right) \), and \( \tilde{D}\left(\mathbf{X}\right) \) are obtained by RSPAM, which are 0.1020, 0.1295, and 0.6390, respectively. The corresponding approximation models are constructed based on these samples and shape parameters. Under the help of the present method, the iterative process and the optimization results are illustrated in Fig. 16 and the count of samples in iterative steps are given in Table 12. Summing up the above results, the final Pareto optimal set satisfying the allowable error can be obtained in step 2, in which the maximal error of the objective functions and constraints is 1.08%. Therefore, eleven solutions are obtained as listed in Table 13. The engineer’s different attention to the optimization objectives will determine the choice of the Pareto solution. If the maximum stress of the forearm structure is the primary concern, the No.1 solution could be selected; if the minimal weight of the forearm structure is the most important, the No.2 solution could be selected. Depending on the optimization results, the engineers could choose the appropriate optimization solution to ensure the safety of the forearm structure.

Fig. 16
figure 16

The iterative process and optimization results (engineering example B)

Table 12 The count of samples in the iterative process and the corresponding maximum error (engineering example B)
Table 13 The Pareto solutions satisfying the requirement (engineering example B)

5 Conclusion

This paper provides an efficient multi-objective optimization (MOO) method based on the adaptive approximation model. In this method, the approximation models based on radial basis function (RBF) is used to instead of the actual objective functions and constraint functions of the MOO problem. For improving the prediction ability and accuracy of the approximation model, the reverse shape parameter analysis method (RSPAM) is proposed. By using the RSPAM, the equation for calculating the shape parameter is established based on initial samples and the optimal shape parameter is solved according to the error criterion. Compared to the traditional approximation model technology, the optimal shape parameter can be obtained effectively without additional sampling, which effectively reduces the model error caused by the operator’s experience. Moreover, the local-densifying approximation method is employed to densify more samples into the local area where the Pareto optimal set of objective functions and their corresponding constraint functions. Because the prediction ability and accuracy of approximation model in local area we concerned is guaranteed, the Pareto optimal set can be further improved. The computational results of two numerical examples demonstrate the outstanding performance of the proposed method in solving the optimization solutions. This method is also performed for the optimization design problems of the gear transmission structure and the forearm of the bucket wheel stacker-reclaimer. The satisfactory optimization results further demonstrate the suitability of the proposed method in practical applications. It is noteworthy that the accurate optimization results are ensured by the RSPAM and local-densifying approximation method. Additionally, when the high-dimensional and strong simulation models are involved, the more initial samples should be adopted to construct the RBF approximation model and more local-densifying iterations should be processed to obtain the accurate optimization solutions. Hence, how to greatly improve the abovementioned problems for better efficiency has been included in our future work.