1 Introduction

As an important aspect in engineering application and scientific research, optimization is indispensable. In practical problems it is usual to optimize several objectives at the same time. These problems are multi-objective optimization problem (MOP). For any MOP, there are several points in decision space called Pareto optimal solutions (PS). Each objective of PS cannot be better without making deterioration of other objectives. The objective vectors of PS are Pareto front (PF). The purpose of an algorithm for solving MOP can be summarized as the following aspects: (1) searching solutions whose objective vectors are on or near PF; (2) searching solutions whose objective vectors distribute along PF as widely as possible.

Evolutionary computing has been widely used to solve MOP and is a great success solving many problems. Literatures Deb et al. (2002) and Zitzler et al. (2002) are some well-known multi-objective evolutionary algorithms (MOEAs). In Deb et al. (2002) a fast nondominated sorting approach is proposed. The solutions are ranked by their levels of nondomination firstly. Then with comparison of crowding distance, the population can be updated given convergence and diversity. In Zitzler et al. (2002) both dominating and dominated solutions are used to assign the fitness of one solution. In the updating of archive, the number of individuals is constant and the boundary solutions remain. Both algorithms treat the MOP as a whole and improve the quality of population by the domination relationship. Many advanced versions have been proposed to improve the performance in this framework. Elhossini et al. (2010) combine evolutionary algorithm (EA) and particle swarm optimization (PSO) to solve MOPs. Both EA operator and PSO operator are used. In Zhang et al. (2008), a regularity model is built to predict the promising area of optimal solutions in decision space.

Instead of treating the MOP as a whole, many traditional mathematical programming methods solve MOP by decomposing it into a number of single objective problems. Based on this idea, Zhang and Li (2007) proposed a new framework of MOEA, multi-objective optimization evolutionary algorithm based on decomposition (MOEA/D). It uses a set of single objective problems to solve MOP. This is a very different work from domination-based algorithms. The population of MOEA/D is composed by different single objective problems. The purpose of searching the PS for a MOP is transformed to search the best solutions of these single objective problems. In another words, the best solutions of these single objective problems are on the PS of the MOP. Once all the global optimal solutions of the single objective problems are obtained, the MOP is solved. In recent years, MOEA/D attracts more and more research attention. Zhang et al. (2010), Li and Landa-Silva (2011), Martínez and Coello Coello (2011), Sindhya et al. (2011), Ishibuchi and Nojima (2011), Cheng et al. (2012) are some of the studies.

In optimization of a single objective problem, avoiding the solutions trapping into local optimum is an important issue. It is also a key point in MOEA/D. The local optimum will lower the performance of MOEA/D from both the convergence and diversity of the solutions. From the view of convergence, it is obvious that the local optimal solution is not on the PS. From the view of diversity, the local optimum will assimilate the solutions of different subproblems. It will be described in part III in detail. In order to make the evolution of subproblems in MOEA/D more effective, some effort has been made in the last several years. Li and Zhang (2009), Zhang et al. (2009), Montaño et al. (2010) are some of them. In Li and Zhang (2009) a version combining DE operator and MOEA/D is proposed. A new mating pool is formed for generating of new solutions and updating of the subproblems. In Zhang et al. (2009), a strategy of allocation of computational resource is designed to make the algorithm focus more attention on more promising subproblems. In Montaño et al. (2010), the neighbors of one solution are determined with the Euclidean distance between solutions. The nondominated solution among its neighbors is defined as locally nondominated solution. These solutions are selected to generate new ones for maintaining the diversity.

Quantum-inspired evolutionary computing (QIEC) is a method based on concept and principle from quantum mechanics. Narayanan and Moore (1996) firstly combine evolutionary computation (EA) and quantum-inspired computation to solve the traveling salesman problem. After that a series of EA inspired by quantum computation appear, such as Han and Kim (2002), Abs da Cruz et al. (2004), Jiao et al. (2008). They are a kind of algorithm characterized by quantum bits and quantum gates. The quantum gates are used to change the quantum bits and generate new solutions through observing. Another kind of quantum-inspired computation is proposed by Sun et al. (2004). It is an idea inspired by the behavior of particles in a potential field. The particles are bounded by the attractor. Meanwhile, they will appear in anyplace of the space with different probability density. Via setting potential well and solving Schrödinger equation, a new style to search space is built. Based on this point, Li et al. (2012) proposed an improved cooperative quantum-behaved particle swarm optimization method for solving real parameter optimization and obtained a good performance.

In this paper, we combine the attractor-based quantum method with MOEA/D to solve multi-objective optimization problems. It takes efforts on two sides: (1) define a so-called global solution (GS) and a local solution (LS) for each subproblem. The GS is the solution with the best scalar function value for the subproblem. The LS is the best solution among the solutions fitting the subproblem; (2) a new method inspired from quantum behavior for generating new solutions is designed for MOEA/D. New solutions are generated around LS to ensure wide searching in space. And with maximum likelihood estimation, GS keeps appearing probability as high as possible in the next generation. Through the new style of generating solutions, the GS and LS guide the evolution cooperatively. The whole paper is arranged as follow: Sect. 2 shows the background and basic methods related to our work. In Sect. 3, some features of MOEA/D are shown and the motivation and process of embedding quantum method in MOEA/D are explained. In Sect. 4, some experimental results are shown and analyzed. In Sect. 5, some conclusions and prospect are summarized.

2 Background and basic method

2.1 MOP decomposition and MOEA/D

In order to solve a MOP, many methods have been proposed to decompose a MOP into a number of scalar optimization problems. In Marler and Arora (2004) many styles can be found. In this chapter, three approaches mentioned in Zhang and Li (2007) are introduced.

The first one is the weighted sum approach which is effective for convex problems. \({\mathop \lambda \limits ^{\rightharpoonup }}=(\lambda _1 ,\lambda _2 ,\ldots \lambda _m )^T\) is a weight vector, i.e., \(\lambda _{i} \ge 0(i = 1, 2, \ldots ,m)\) and \(\sum _{i=1}^{m} \lambda _{i} = 1\). A subproblem constructed by a scalar function can be stated as:

$$\begin{aligned}&\mathrm{Minimize}\;\;{g^{ws}}\left( {X|}{\mathop \lambda \limits ^{\rightharpoonup }} \right) = \sum \limits _{i = 1}^m {{\lambda _i}} {f_i}\left( {{x_i}} \right) \nonumber \\&\mathrm{subject}\;\;\mathrm{to}\;\;\;X \in {\Omega ^n} \end{aligned}$$
(1)

Here \({\mathop \lambda \limits ^{\rightharpoonup }}\) is a coefficient of this scalar function. \(X\) is the variable vector to be optimized.

Another approach is Tchebycheff Approach. The form of the scalar function can be expressed as:

$$\begin{aligned}&\mathrm{Minimize}\;\;{g^{te}}\;\;\left( {X\left| {{\mathop \lambda \limits ^{\rightharpoonup }} ,{\mathop {Z^{*}}\limits ^{\rightharpoonup }}}\right. } \right) = \mathop {\max }\limits _{1 \le i \le m} \left\{ {{\lambda _i}\left| {{f_i}({x_i}) - z} \right. _i^*} \right\} \nonumber \\&\mathrm{subject}\;\;\;\mathrm{to}\;\;\;X \in {\Omega ^n} \end{aligned}$$
(2)

Here \({\mathop {Z^{*}}\limits ^{\rightharpoonup }}\) is called reference point, i.e., \(z_i^*=\mathop {\min }\nolimits _{X\in \Omega ^n} f_i (X)\).

The third method is penalty-based boundary intersection (PBI) approach. The form of the scalar function can be described as:

$$\begin{aligned}&\mathrm{Minimize}\;\;{g^\mathrm{pbi}}\;\;\left( {X\left| {\mathop \lambda \limits ^{\rightharpoonup }} {,}{\mathop {Z^{*}}\limits ^{\rightharpoonup }} \right. } \right) = {d_1} + \theta {d_2}\nonumber \\&\mathrm{subject}\;\;\mathrm{to}\;\;X \in {\Omega ^n} \nonumber \\&\mathrm{where}\;\;\;{d_1} = \frac{\left\| {(\mathop {Z^{*}}\limits ^{\rightharpoonup }- F(X))}^{T} \mathop \lambda \limits ^{\rightharpoonup }) \right\| }{{\left\| {{\mathop {\lambda }\limits ^{\rightharpoonup }}} \right\| }} \nonumber \\&\mathrm{and}\;\;\;{d_2} = \left\| {F(X) - (\mathop {Z^{*}}\limits ^{\rightharpoonup }- {d_1}\mathop \lambda \limits ^{\rightharpoonup })} \right\| \end{aligned}$$
(3)

\({\mathop {Z^{*}}\limits ^{\rightharpoonup }}\) is reference point as in Tchebycheff Approach. \(\theta >\) 0 is a penalty parameter. The details of this approach can be found in Zhang and Li (2007).

MOEA/D is a framework combining the population of subproblems with EA to solve a MOP. Because the subproblems are single objective problems, the values of scalar functions are used to evaluate solutions instead of using Pareto domination comparison. The process of MOEA/D is shown in Algorithm 1:

figure a

The framework of MOEA/D has some natural advantages than the non-domination-based MOEAs. The first one is its computational complexity. As well known, the computational complexity of fast nondominated sorting (Deb et al. 2002) is as high as \(O( {MN^2})\) (where \(M\) is the number of objectives and \(N\) is the population size). But in the frame of MOEA/D, for the nondominated sorting is avoided, the computation cost is low. Another advantage of MOEA/D is its ability of maintaining diversity. Each subproblem responds a region in PF. A uniform and broad searching can be offered with different subproblems. No special operators for diversity of solutions are needed.

2.2 Quantum-inspired method in EC

Quantum-inspired methods used in EC are a series of approaches coming from the concepts and equations of quantum mechanics. Usually these methods are associated with quantum gates, quantum chromosome, quantum bits, potential well, etc. In this paper, we mainly focus on the method based on attractor and characteristic length which is proposed by Sun et al. (2012).

In order to illustrate these approaches, some background knowledge is required. In classical physics, the state of motion of a particle is described by Newton equation based on Newton three laws of motion. The motion is deterministic, i.e., once the coordinate and momentum of a particle are given, its trajectory can be grasped. But in quantum space, there is wave particle duality. The motion trajectory of a particle is uncertain even if the coordinate and momentum is given. The particle can appear at any location of the space at the next time point. Instead of predicting the exact location of the particle, we just have the ability to give the probability density of appearing in the space. This prediction is achieved through wave function \(\Psi (X,t)\), the square of whose modulus is probability density. Here a three-dimensional space acts as an example. For a given time point, some equations can be described as follows:

$$\begin{aligned}&Q = {\left| {\Psi (x,y,z)} \right| ^2}\end{aligned}$$
(4)
$$\begin{aligned}&\int \limits _{ - \infty }^{ + \infty } {{{\left| \Psi \right| }^2}{d_x}{d_y}{d_z} = } \int \limits _{ - \infty }^{ + \infty } {Q{d_x}{d_y}{d_z} = 1} \end{aligned}$$
(5)

\(x,y,z\) are coordinates describing the space. \(Q\) is the probability density function of appearing on any location for a particle.

In order to get the probability density function \(Q\), Schrödinger equation should be solved. The form of the equation can be described as:

$$\begin{aligned} ih\frac{\partial }{{\partial t}}\psi (X,t) = \mathop H\limits ^ \wedge \psi (X,t) \end{aligned}$$
(6)

Here \(h\) is Planck constant. \(\mathop H\limits ^\wedge \) is Hamiltonian operator:

$$\begin{aligned} \mathop H\limits ^ \wedge = - \frac{{{h^2}}}{{2m}}{\nabla ^2} + V(X) \end{aligned}$$
(7)

\(m\) is the mass of the particle and \(V( X)\) is the potential field where the particle is. \(\nabla ^{2}\) is the Laplacian operator.

As shown in Fig. 1, in the quantum space, a point \(A\) and any point \(B\) have the low and high energy, respectively. This form of potential field is called potential well. Assuming \(V(X)\) is one dimensional \(\delta \) potential well:

$$\begin{aligned} V(X) = - \gamma \delta (X - p) = - \gamma \delta (Y) \end{aligned}$$
(8)

Here \(Y=X-p\). \(p\) is the coordinate of \(A\). \(\gamma \) is an intensity coefficient. Then assuming the particle is in a certain energy state, the Schrödinger equation is equivalent to the following form:

$$\begin{aligned} \frac{d^2\phi (Y)}{dY^2}+\frac{2m}{h^2}[E+\gamma \delta (Y)]\phi (Y)=0 \end{aligned}$$
(9)

Here the probability density can be given as:

$$\begin{aligned} Q(Y)=\left| {\phi (Y)} \right| ^2=\frac{1}{\sqrt{L} }e^{{{-2\left| Y \right| }} /{L}} \end{aligned}$$
(10)

\(L={h^2} / {m\gamma }.\) Based on the probability density function, the location of a particle can be determined by a stochastic equation as:

$$\begin{aligned} X=p\pm \frac{L}{2}\ln (1 / \mu ) \end{aligned}$$
(11)

\(\mu \sim U( {0,1}).\) The details of above derivation of the equations can be referenced in Sun et al. (2004).

Fig. 1
figure 1

Particles in quantum space

According to Eq. 11, it can be found that a particle will appear around \(p\), the bottom of the potential well, with high probability. And in other locations, the particle will appear with low probability. This point will be discussed more clearly in Sect 3. Here \(p\) and \(L\) are called as attractor and characteristic length, respectively. In Sun et al. (2012), a series of choice are given for the attractor and characteristic length. In the following section, special \(p\) and \(L\) are designed for MOEA/D according to its characters.

3 Motivation and quantum-inspired MOEA/D

3.1 Motivation

Assuming the set of subproblems in MOEA/D is {\(S_{1}\)\(S_{2}\), \(S_{3},\ldots S_{N}\)}, the desired result is that every global optimum of the subproblems can be obtained to form the PF of the MOP. In the original framework of MOEA/D, a best solution is stored for each subproblem and it can be updated by new solutions generated from its neighbors. It means a new solution can be used on different subproblems. Or it can be said the solutions of several subproblems can be replaced by the same one. Obviously this cooperation of subproblems can improve the convergence of the solutions because more calculated resources are spend on several best solutions found so far. But on the other side, this cooperation blocks searching the global optimums of subproblems at some cases. Once a good solution is obtained by a subproblem, the solutions stored by its neighbors may be updated by the good one. This will lead the best solution of the neighbor to lose its trajectory. A visual example is shown in Fig. 2. For each subproblem, the best solution stored is represented as the hollow point. The ideal state is that the hollow point move along its arrow and reach a black point on the PF. Thus when all the hollow points reach the black points, the whole PF is obtained. But during the evolution, once a good solution like the red point appears, the best solutions of \(S_k ,S_{k+1} \) will be updated by the red point. The moving strategy of these two hollow points will deviate from the intended track. The immediate result is that the black point responding these hollow points cannot be reached at the end of the evolution. From the point of PF, it will be observed that some parts of the PF cannot be found. Some studies have shown similar shortcomings. Liu et al. (2014) points out the simple replacement may miss some search regions. Li and Zhang (2009) indicates the loss of diversity at the early stage of search on some problems. And in Sect. 4, some visual results of the original MOEA/D can be seen.

Fig. 2
figure 2

The updated hollow points in oval with the red point (color figure online)

Based on above characters of the original MOEA/D, a special quantum-inspired method is designed to relieve the side effects of the cooperation among subproblems. It focuses on the design of the attractor and the characteristic length used in quantum method. The next two subsections will show the details.

3.2 The attractor and characteristic length for QMOEA/D

In order to construct the attractor and characteristic length of QMOEA/D, two solutions are stored for each subproblem. One is marked as global solution (GS) and the other is marked as local solution (LS). When a new solution is generated from one subproblem, the GSs of its neighbors will be updated. Then a most suitable subproblem for the new solution will be determined and the LSs of its neighbors will be updated. The visual difference between GS and LS is shown in Fig. 3. A new solution represented by a black point is generated from subproblem \(S_n \). The GSs of the neighbor subproblems of \(S_n \) will be updated if the black point is better than the stored GSs. Obviously the black point is closer to the subproblem \(S_m \). Then we will determine whether the LSs of the neighbors of \(S_m \) should be replaced by the new point according to the scalar function value.

Fig. 3
figure 3

The update of the GS and LS with a newly generated solution

Why are these two solutions stored for one subproblem? As what have been discussed, during the moving of one solution, if the objective vector deviates from the intended region, this solution is not suitable for the corresponding subproblem. It will mislead the subproblem to spend calculate resources on the regions focused by other subproblems, which is not helpful to obtain the global optimum. For a new solution which has a good scalar function value, if it deviates from the region of responding subproblem, it cannot update the LS of the subproblem. So the LSs are used to keep the solutions moving along the intended region. The GS is the same as the solutions in original MOEA/D. Except the scalar function value, no special information is used to guide the updating of GS. It is obvious that the solutions with best fitness should remain so to promote the evolution of populations. So with the use of LSs and GSs of these subproblems, the solutions are expected to move along the suitable direction to obtain the whole PF.

In quantum-inspired algorithms, the probability density function of the new solution deduced from Eq. 10 can be described as:

$$\begin{aligned} Q(X)=\frac{1}{\sqrt{L} }\mathrm{e}^{{-2\left| {X-p} \right| } /L} \end{aligned}$$
(12)

When \(p=0,L=1\) and \(p=0,L=5\), the probability density function of \(X\) is shown in Figs. 4 and 5, respectively. It is easy to find that solutions near attractor \(p\) are more likely to be generated. But the characteristic length \(L\) can be used to modify the appearing probability of one location. When \(L=1\) and \(L=5\), the distributions are significantly different.

Fig. 4
figure 4

The probability density with \(p = 0\), \(L = 1\)

Fig. 5
figure 5

The probability density with \(p = 0\), \(L = 5\)

Taking both MOEA/D and quantum-inspired methods into account, the LS of each subproblem is selected as the attractor. Because the LS is the best one among the solutions fitting for the subproblem, more calculate resources should be spent around LS in generating more suitable solutions. Thus, Eq. 12 becomes the following form:

$$\begin{aligned} Q(X)=\frac{1}{\sqrt{L} }\mathrm{e}^{{-2\left| {X-{LS}} \right| } /L} \end{aligned}$$
(13)

In evolutionary algorithm, it is reasonable to remain the best solution. So for each subproblem, the solutions around GS which is the best solution obtained so far should have a high probability to appear. From Eq. 13 and maximum likelihood estimation, \(L\) can be obtained as:

$$\begin{aligned} L=2\left| {{GS}-{LS}} \right| \end{aligned}$$
(14)

Thus the equation for QMOEA/D to generate new solution Ns can be described as:

$$\begin{aligned} {Ns}={LS}\pm ({GS}-{LS})*\ln (1 / \mu ) \end{aligned}$$
(15)

3.3 Framework of QMOEA/D

In the above subsection, two main elements in quantum-inspired method are constructed for MOEA/D. Based on the previous attractor and characteristic length, the framework of QMOEA/D is shown as Algorithm 2:

figure b

In the above framework, step 5 would be discussed in detail. For each subproblem \(S_i \) whose weight vector is \(\left( {\lambda _i^1 ,\lambda _i^2 ,\ldots ,\lambda _i^m }\right) \), we normalize its weight vector to be \(\left( {1,\frac{\lambda _i^2 }{\lambda _i^1 },\ldots ,\frac{\lambda _i^m }{\lambda _i^1 }}\right) \). And the objective vector \(F({Ns})=\left( {f_{Ns}^1 ,f_{Ns}^2 ,\ldots ,f_{Ns}^m }\right) \) is also normalized as \(\Big (\!{1,\frac{f_{Ns}^2 }{f_{Ns}^1 },\ldots ,\frac{f_{Ns}^m }{f_{Ns}^1 }}\!\Big )\). The Euclidean distance between these two normalized vectors is recorded as the distance between \(Ns\) and \(S_i \). The subproblem with the shortest distance to \(Ns\) is selected as the most suitable one.

4 Comparison and discussion

In the following, the performance metric and test suits used in our experiments are introduced firstly. Then the comparison between QMOEA/D and nondominated sorting approach NSGA-II, QMOEA/D and original MOEA/D (MOEA/D-SBX), QMOEA/D and some other advanced MOEA/D are shown, respectively. And some discussion about QMOEA/D is made in the last subsection.

4.1 Performance metric

In the experiments, inverted generational distance (IGD) (Sierra and Coello Coello 2005) is used as the metric to measure the quality of solution sets. \(P\)* is an objective vector set whose members distribute on the true PF uniformly. \(P\) is another objective vector set whose members are obtained by a multi-objective optimization algorithm. Then the IGD of \(P\) can be calculated as:

$$\begin{aligned} \mathrm{IGD}(P) = \frac{{\sum \limits _{\nu \in {P^*}} {d(\nu ,P)} }}{{\left| {{P^*}} \right| }} \end{aligned}$$
(16)

Here \(\vert P^{*}\vert \) is the number of members in \(P\)* and set 500 in our experiments. \(d\)(\(\nu \)\(P)\) is the minimum distance from \(\nu \) to the set \(P\).

It is obvious that the IGD is able to measure the convergence of solutions. Meanwhile, the metric can reflect the diversity of solutions like Fig. 6. When the objective vectors, represented by the hollow circles, are not distributed along PF widely, no solutions in set \(P\) is near with set \(A\). So the sum of the distance in Eq. (16) will be large. And the IGD of the hollow circles could not be small.

Fig. 6
figure 6

A PF losing diversity

For the comparison experiments in this section, several independent runs are performed with each algorithm. The average value, best value and standard deviation are shown and compared, respectively. To be more rigorous, the Wilcoxon signed-ranks test is performed on each problem. The significance level is set to be 0.05 (García et al. 2009). The sign (+) indicates the comparison algorithm is significantly better than QMOEA/D. The sign (\(-\)) indicates the comparison algorithm is significantly worse than QMOEA/D. The sign (=) indicates no significant difference.

4.2 Multi-objective suites

Eight test problems are used in the experiments. All the test suites are minimization problems. ZDT1 and ZDT2 come from Zitzler et al. (2000). ZZJ1, ZZJ2, ZZJ5 and ZZJ6 come from Zhang et al. (2008). MOP1 and MOP2 come from Liu et al. (2014). The details of the functions are as follows:

ZDT1:

$$\begin{aligned} \begin{array}{l} f_1 (x)=x_1 \\ f_2 (x)=g(x)\left[ 1-\sqrt{\frac{f_1 (x)}{g(x)}}\right] \\ \mathrm{where} \quad g(x)=1+\frac{9\left( \sum \limits _{i=2}^n {x_i }\right) }{n-1} \\ \end{array} \end{aligned}$$

and \(x_i \in [0,1]\;\;\;i=1,\ldots ,n\), \(n = 30\).

ZDT2:

$$\begin{aligned} \begin{array}{l} f_1 (x)=x_1 \\ f_2 (x)=g(x)\left[ 1-\left( \frac{f_1 (x)}{g(x)}\right) ^2\right] \\ \mathrm{where} \quad g(x)=1+\frac{9\left( \sum \limits _{i=2}^n {x_i }\right) }{n-1} \\ \end{array} \end{aligned}$$

and \(x_i \in [0,1]\;\;\;i=1,\ldots ,n\), \(n = 30\).

ZZJ1:

$$\begin{aligned} \begin{array}{l} f_1 (x)=x_1 \\ f_2 (x)=g(x)\left[ 1-\sqrt{\frac{f_1 (x)}{g(x)}}\right] \\ \mathrm{where} \quad g(x)=1+\frac{9\left( \sum \limits _{i=2}^n {(x_i -x_1 )^2}\right) }{n-1} \\ \end{array} \end{aligned}$$

and \(x_i \in [0,1] \quad i=1,\ldots ,n\). \(n = 30\).

ZZJ2:

$$\begin{aligned} \begin{array}{l} f_1 (x)=x_1 \\ f_2 (x)=g(x)\left[ 1-\left( \frac{f_1 (x)}{g(x)}\right) ^2\right] \\ \mathrm{where} \quad (x)=1+\frac{9\left( \sum \limits _{i=2}^n {(x_i -x_1 )^2}\right) }{n-1} \\ \end{array} \end{aligned}$$

and \(x_i \in [0,1] \quad i=1,\ldots ,n\), \(n = 30\).

ZZJ3:

$$\begin{aligned} \begin{array}{l} f_1 (x)=x_1 \\ f_2 (x)=g(x)\left[ 1-\sqrt{\frac{f_1 (x)}{g(x)}}\right] \\ \mathrm{where} \quad g(x)=1+\frac{9\left( \sum \limits _{i=2}^n {(x_i^2 -x_1 )^2}\right) }{n-1} \\ \end{array} \end{aligned}$$

and \(x_i \in [0,1] \quad i=1,\ldots ,n\), \(n = 30\).

ZZJ4:

$$\begin{aligned} \begin{array}{l} f_1 (x)=x_1 \\ f_2 (x)=g(x)\left[ 1-\left( \frac{f_1 (x)}{g(x)}\right) ^2\right] \\ \mathrm{where} \quad g(x)=1+\frac{9\left( \sum \limits _{i=2}^n {(x_i^2 -x_1 )^2}\right) }{n-1} \\ \end{array} \end{aligned}$$

and \(x_i \in [0,1] \quad i=1,\ldots ,n\), \(n = 30\).

MOP1:

$$\begin{aligned} f_1 (x)= & {} (1+g(x))x_1\\ f_2 (x)= & {} (1+g(x))(1-\sqrt{x_1 } ) \end{aligned}$$

where \(g(x) = 2\) sin (\(\pi x\) \(_{1})\) \(\sum _{i=2}^{n}\)(\(-\)0.9\(t_{i}^{2}\) + \(\vert t_{i}\vert ^{0.6})\),

$$\begin{aligned} t_i =x_i -\sin (0.5\pi x_1 ) \end{aligned}$$

and \(x_i \in [0,1] \quad {i=1,\ldots ,n} \), \(n = 10\).

MOP2:

$$\begin{aligned} \begin{array}{l} f_1 (x)=(1+g(x))x_1 \\ f_2 (x)=(1+g(x))(1-x_1^2 ) \\ \end{array} \end{aligned}$$

where \(g(x)=10\sin (\pi x_1 )\sum \nolimits _{i=2}^n {\left( \frac{\left| {t_i } \right| }{1+e^{5\left| {t_i } \right| }}\right) } \)

$$\begin{aligned} t_i =x_i -\sin (0.5\pi x_1 ) \end{aligned}$$

and \(x_i \in [0,1] \quad {i=1,\ldots ,n}\), \(n = 10\).

4.3 Comparison betwwen QMOEA/D and NSGA-II

In this subsection, a comparison is made between QMOEA/D and NSGA-II which is the classical evolutionary algorithm for multi-objective optimization. The population size is 100 for both the algorithms like Deb et al. (2002). 3 \(\times \) 10\(^{5}\) fitness evaluations or 3000 generations are used as the termination condition. The neighbor size for QMOEA/D is 20. The parameters used for crossover and mutation keep the same with Deb et al. (2002). 30 independent experiments are run with each algorithm on each problem.

The PFs found in the 30 runs are plotted in Fig. 7. For each problem, all the PFs obtained with one algorithm are overlaid in the figure. The same style is used for plotting other figures of PFs in the following subsections.

Fig. 7
figure 7

The PFs from the 30 independent runs on the 8 test problems. The left are the results of QMOEA/D. The right are the results of NSGA-II

Table 1 shows the average, best and std of IGDs on different problems, respectively. For ZDT1–ZZJ6, both the algorithms can obtain the whole PFs. But the quality of these PFs is quite different. QMOEA/D outperforms NSGA-II on all these six problems on both the average IGD and best IGD. Especially on ZDT2 and ZZJ6, the advantage is huge from the view of average IGD. And the std shows QMOEA/D has a more steady performance than NSGA-II. On MOP1 and MOP2, the PFs obtained by QMOEA/D are wider than the ones from NSGA-II which can be observed from Fig. 7.

Table 1 The average, best and std of IGDs among the 30 independent runs with QMOEA/D and NSGA-II

With signed-ranks test, it can be observed that on all the eight problems QMOEA/D outperforms NSGA-II significantly. The different frameworks of the two algorithms may be a main reason for the different performance. Methods based on decomposition performs better than the methods based on domination ranking on many test problems, which has been shown in Zhang and Li (2007). The quantum-inspired operators for MOEA/D may be another important factor. In the next subsection, the influence of quantum-inspired operator will be shown through the comparison between QMOEA/D and the original MOEA/D.

4.4 Comparison betwwen QMOEA/D and the original MOEA/D

The performance of QMOEA/D and the original MOEA/D (MOEA/D-SBX) are shown in this subsection. For fairness, the common parameters keep the same in both the algorithms. The population size is 200. The size of neighbor is 20. The decomposition approach used here is Tchebycheff. And 3 \(\times \) 10\(^{5}\) fitness evaluations are used as the termination condition. The distribution index for polynomial mutation is 20 and the mutation rate is 1/\(n\). \(n\) is the dimension of the decision space. In MOEA/D, the distribution index for simulated binary crossover (SBX) is set to 20. And the SBX rate is 1.

Both MOEA/D-SBX and QMOEA/D have been run 30 times independently on the 8 test problems, respectively. Figure 8 shows the Pareto fronts obtained from the 30 independent runs.

Fig. 8
figure 8

The PFs from the 30 independent runs on the 8 test problems. The left is the results of QMOEA/D. The right is the results of MOEA/D-SBX

From the shape of the obtained PFs, an intuitive conclusion is that QMOEA/D can find more complete PFs. Both algorithms have good performance on ZDT1 and ZDT2. But on ZZJ1, ZZJ2, ZZJ5, ZZJ6 and MOP1, MOEA/D-SBX has little effect. Only a few points or parts near the PF are obtained. However QMOEA/D has better performance on these five problems. Except MOP1, the PFs from QMOEA/D are complete and smooth. For MOP1, although the convergence of the obtained PF is not good, the PF is relatively complete. On MOP2, a good PF is not obtained with both the algorithms. MOEA/D-SBX only obtains several points near the boundary of the PF. Meanwhile, a part with bad convergence of PF is found with QMOEA/D on this problem.

Generally speaking, the original MOEA/D (MOEA/D-SBX) is easy to lose the diversity of the solutions which has been proved by the shape of PF. Among the eight problems, the whole PF is obtained only on two problems. The results also tally with the motivation of our work. Some reasons for the poor performance on diversity have been shown in Part III. The solution cannot move along its own intended track in the framework of MOEA/D-SBX. Thus some parts of the PF cannot be found.

Compared with MOEA/D-SBX, QMOEA/D has a significant improvement. In order to show the results more precisely, Table 2 presents the average, best and std of IGDs with the 30 independent runs on the 8 problems. The better one is marked with bold style.

Table 2 The average, best and std of IGDs among the 30 independent runs with QMOEA/D and MOEA/D-SBX

From the view of signed-ranks test, QMOEA/D outperforms MOEA/D-SBX on all the problems except ZDT2. Especially for ZZJ1, ZZJ2, ZZJ5, ZZJ6 and MOP1, the IGD values have great improvement than the original MOEA/D. Because IGD is an index evaluate the solutions from both the convergence and diversity. Generally speaking, the huge difference can be attributed to the failure of MOEA/D-SBX in obtaining the complete PF.

4.5 Comparison between QMOEA/D and some advanced versions of MOEA/D

In this subsection, two advanced versions of MOEA/D are used as the comparison algorithms. One is MOEA/D-DE (Li and Zhang 2009) and the other is MOEA/D-DRA (Zhang et al. 2009).

In MOEA/D-DE, three aspects are modified based on MOEA/D-SBX. (1) The differential evolution (DE) operator, instead of SBX, is used to generate new solutions; (2) the neighbors are used as the mating/update range with probability \(\delta \). Otherwise, the whole subproblems are used as the mating/update range; (3) the new generated solutions can update \(n_{r}\) solutions stored in mating/update range at most. These are the differences between MOEA/D-DE and MOEA/D-SBX. The details of MOEA/D-DE can be found in Li and Zhang (2009).

In MOEA/D-DRA, except the three aspects mentioned above, a utility value is stored for each subproblem. The utility value is calculated based on the relative decrease of the scalar function value during several generations. It is used to allocate the compute resources to different subproblems. The details of MOEA/D-DRA can be found in Zhang et al. (2009).

In this subsection, the common parameters are the same as above subsection. The population size is 200. The neighbor size is 20. Tchebycheff approach is also used here. 3 \(\times \) 10\(^{5}\) fitness evaluations are used as the termination condition. The distribution index for polynomial mutation is 20 and the mutation rate is 1/\(n\). \(n\) is the dimension of decision space. Here the \(F\) and CR in DE is set to 0.5 and 1.0, respectively, like Li and Zhang (2009). \(n_{r}\) and \(\delta \) used in MOEA/D-DE and MOEA/D-DRA are set to 2 and 0.9, respectively. In MOEA/D-DRA, the utility value is updated every 50 generations.

For these three algorithms, the experiments are run 30 times independently. Figure 9 shows all the PFs obtained in the 30 runs. From the left to the right, the figures respond to QMOEA/D, MOEA/D-DE and MOEA/D-DRA, respectively.

Fig. 9
figure 9

The PFs from the 30 independent runs on the 8 test problems. The left is the results of QMOEA/D. The middle is the results of MOEA/D-DE. The right is the results of MOEA/D-DRA

From Fig. 9, it can be found that all these three algorithms have better performance than MOEA/D-SBX. On the first six problems, all these three algorithms can obtain the complete PF. But on MOP1, only QMOEA/D obtains the relatively complete PF. On MOP2, the solutions from MOEAD-DRA focus on the two ends of PF. From the view of PF shape, QMOEA/D and MOEA/D-DE have a similar performance on this problem.

In MOEA/D-DE, the new generated solution can update only \(n_{r}\) solutions instead of updating \(T\) solutions. This strategy may reduce the negative influence caused by other subproblems. It will help the solution move along its own track. The same mechanism also exists in MOEA/D-DRA. This is the reason why these two algorithms have a good performance on the first six problems. For the last two problems, there are some deception regions in their decision space. It is a challenge for many algorithms.

Some more precise values are given in the following. Table 3 shows the average, best and std IGDs of the 30 independent runs on each problems. The best one is marked with bold style.

Table 3 The average, best and std of IGDs among the 30 independent runs with QMOEA/D, MOEA/D-DE and MOEA/D-DRA

For the average IGDs, QMOEA/D obtains 6 best results among the 8 ones. On ZZJ5 and ZZJ6, the IGD values of QMOEA/D are slightly worse than the best ones. On MOP1 and MOP2, the best ones com from QMOEA/D. Especially on MOP1, the advantage is greater than the results of the other two algorithms. For the best IGDs, QMOEA/D obtains 5 best results among the 8 ones. On ZZJ5, ZZJ6, and MOP2, the results of QMOEA/D are worse than the best one. But the shortage is slight. On MOP1, the best value also belongs to QMOEA/D with huge advantage.

Generally speaking, QMOEA/D has good performance on most of the test problems. Although not all the best results are from QMOEA/D, QMOEA/D can offer competitive results among these advanced versions of MOEA/D.

4.6 Some further discussion of the operators in QMOEA/D

As described in the above section, two changes are proposed in QMOEA/D. One is the stored GS and LS. The other is the quantum-inspired generator. In this subsection, the different effect of the two changes in QMOEA/D will be discussed based on some results of the comparison experiments.

A comparison algorithm called SMOEA/D is designed based on QMOEA/D. In SMOEA/D, GS and LS are also stored and used as the parent solutions for each subproblem. But the new solutions are generated with simulated binary crossover (SBX), which means step 1 in Algorithm 2 is replaced with SBX.

The distribution index in SBX is set to be 20. For both QMOEA/D and SMOEA/D, the population size is 200. And the stopping condition is 3 \(\times \) 10\(^{5}\) fitness evaluations. 30 independent runs are also performed.

The PF shape of QMOEA/D and SMOEA/D are plotted in Fig. 10. Table 4 shows the average, best and std of IGDs on different problems, respectively. The better one is marked with bold style.

Fig. 10
figure 10

The PFs from the 30 independent runs on the 8 test problems. The left is the results of QMOEA/D. The right is the results of SMOEA/D

Table 4 The average, best and std of IGDs among the 30 independent runs with QMOEA/D and SMOEA/D

Firstly, the discussion is about the use of GS and LS. From Figs. 8 and 10, it can be found that the PFs obtained by SMOEA/D and QMOEA/D are more complete than the PFs obtained by the original MOEA/D. Although the generators are different in QMOEA/D and SMOEA/D, in these two algorithms the GS and LS are stored and used for one subproblem. So it can be concluded that the use of LS and GS is helpful to retain the diversity of solutions in the framework of decomposition.

Secondly, the discussion is around the quantum-inspired generator. The main difference between QMOEA/D and SMOEA/D focuses on the different generators. In Table 4, different performance on the value of IGDs are shown. For ZDT1, ZDT2, ZZJ1, ZZJ2 and ZZJ6, the IGDs are similar for both the methods. With signed-rank test, it can be found that QMOEA/D is still better than SMOEA/D on two of them. For ZZJ5, MOP1 and MOP2, QMOEA/D is advantageous to SMOEA/D on both the average and best value. From the view of signed-rank test, QMOEA/D also outperforms SMOEA/D significantly on these three problems. Although the SMOEA/D has the ability to retain the relatively complete shape of PF. The quality of the solutions is not as good as the results obtained by QMOEA/D. As described above, the only difference between these two algorithms is the generator. With the same parents, GS and LS, the solutions obtained by QMOEA/D have better performance than the solutions offered by SMOEA/D. The advantages should be contributed to the generator used in QMOEA/D. It means the quantum-inspired generator is more effective to improve the performance of original MOEA/D with GS and LS.

5 Summary

MOEA/D decomposes a MOP into a number of single objective optimization problems. This framework provides us great convenience to the use of some methods which are widely applied in single objective optimization. But subproblems in MOEA/D are not totally the same as the simple single objective problems. In order to use the approaches in this framework effectively, more special features of MOEA/D should be mined.

In this paper, we combine the quantum-inspired method with MOEA/D because of the poor performance on the diversity. The subproblems in MOEA/D are some associated problems. The neighboring weight vectors respond to the similar subproblems. The information from the neighbor problems may make the solution deviate from its own track and reduce the diversity of the solutions. So we store LS and GS for each subproblem. Besides, special attractor and characteristic length are designed for MOEA/D.

And the experimental results show that the QMOEA/D is an effective and competitive algorithm for solving MOPs. But on some test problems such as MOP2, no excellent results are obtained. More study about MOEA/D are needed. Some more special behavior characters should be discovered for designing more reasonable approach.