1 Introduction

Many real-world problems are dynamic multiobjective optimization problems (DMOPs), with not only the conflict among multiple objectives but also the objective function changing over time (Farina et al. 2004). Dealing with a complex dynamic system, static optimization method has obvious limitations. On these issues, the research objective is changing intricately. Traditional evolutionary algorithm’s goal is to make the population gradually converge to ultimately get a satisfactory solution set, but this would make the population lose diversity; especially in the later stages of the evolution, population will gradually lose ability to adapt to environmental change, which is the challenge of traditional evolutionary algorithm in a dynamic environment (Branke 2002; Jin and Branke 2005; Goh and Tan 2009; Nguyen et al. 2012). To track the Pareto optimal solution in a timely manner after change, researchers need to make adjustments to traditional static multiobjective algorithm (Coello Coello 2006, 2007; Yang et al. 2013; Li et al. 2013), so that it can quickly respond to environmental changes.

In recent years, researchers have proposed some dynamic multiobjective optimization algorithms to adapt DMOPs in artificial immune algorithm, co-evolutionary algorithm, particle swarm optimization algorithm, and memetic computing, which are based on the natural computation methods (Avdagi’c et al. 2009; Shang et al. 2013; Greeff and Engelbrecht 2008; Huang et al. 2011; Isaacs et al. 2008; Liu et al. 2011; Wei and Wang 2012; Azevedo and Araujo 2011; Manriquez et al. 2010; Cámara et al. 2010; Zheng 2007). These methods are mostly concentrated on maintaining population diversity and adopting the strategy of multi-population. However, these methods are slightly blind to solve DMOPs, because the population still relies on the ability to evolve independently to find the optimal solution after the diversity of the population is maintained, and the rate of convergence is the main problem. As to the way to post for objectives after change which provides a guiding direction for population evolution, it will play a vital role to improve the rate of convergence and convergence of the algorithm.

Therefore, some scholars have proposed the strategies of introducing memory (Deb et al. 2007; Zhang 2008; Wang and Li 2009; Isaacs et al. 2008; Guan et al. 2005; Goh and Tan 2009  and prediction (Hatzakis and Wallace 2006a, b; Zhou et al. 2007, 2013; Ma et al. 2011  to solve this problem. Memory strategy reuses the optimal solutions which are previously searched by the memory to rapidly response to changes in the new environment. This strategy can achieve good results for periodic problems, but for non-periodic problems or in the first period of changing environment, population is still in the process of blind evolution, and algorithm is difficult to obtain good convergence. For example, Deb et al. (2007) proposed a dynamic NSGA-II algorithm, Zhang (2008) proposed a dynamic multiobjective immune algorithm, and Goh and Tan (2009) proposed a dynamic competition–cooperation coevolutionary algorithm. These algorithms are simply introducing memory strategy to solve DMOPs, but the effect is not very good in the first period and changes in a particular environment. After each environmental change, methods based on prediction provide guidance direction for the population evolution by the prediction mechanism, and help algorithm to respond quickly to new changes. And forecast accuracy is the main difficulty. So far prediction method has few in-depth studies in DMOPs. Hatzakis and Wallace (2006a) proposed a feed-forward prediction strategy, referred to as FPS, in 2013, Zhou and Jin et al. (2013) proposed a population prediction strategy, referred to as PPS. Prediction strategies are used in these algorithms to solve the DMOPs, and have achieved good results, but there are some defects. For example, prediction is not accurate enough in FPS by simply using autoregressive model. And PPS has poor convergence due to the lack of historical information accumulation in the beginning stage of environmental change.

The combination of memory and prediction is a new trend to solve DMOPs (Yang et al. 2012). Prediction strategy can guarantee the algorithm to quickly respond to environmental changes, and in a timely manner to search for a new optimal solution set. Meanwhile, the memory strategy can improve the ability of dealing periodic problems and has reduced the error rate because of every prediction that may generate. In recent years, a few researches have used this hybrid strategy to solve DMOPs (Wang and Li 2009; Helbig and Engelbrecht 2011; Koo et al. 2009), for example, in 2010, Wang and Li proposed a multi-strategy ensemble evolutionary algorithm, referred to as dMS-MOEA (Wang and Li 2009), in the same year, Koo et al. (2009) proposed a predictive gradient strategy, referred to as dMO-EGS. Both these two algorithms use hybrid strategies which combine memory and prediction to solve DMOPs, but they have their own defects. In the aspect of prediction, dMS-MOEA uses an adaptive genetic and difference operator. In the early stage of running the algorithm, genetic operator can guarantee rapid convergence. And in the later stages of running, the difference operator can ensure a better distribution for the solution set having a better convergence. However, this prediction does not build on the basis of available information, but only increases the diversity of population, so this is a blind prediction. dMO-EGS predicts the evolutionary gradient of current population by recording the final two previous positions of nondominated population center which have been searched and the last predictive evolutionary gradients. This gradient prediction applies only to the same or a similar degree of change in the environment; when there is a large degree of the environmental change, the accuracy of prediction is difficult to be guaranteed. In the aspect of memory, dMS-MOEA produces new individuals on the basis of the searched final nondominated populations by the difference operator and Gaussian mutation operation, and they are stored to the archive. The archive uses Gaussian local search operator to get a new population. dMO-EGS stores past population centers and variance vectors of nondominated solution set, and then generate new populations according to the centers and the variance vectors by normal distribution sampling. Since the sampling process prone to distortion, so the two strategies are incomplete memory strategies.

To solve these problems, the paper proposes novel prediction and memory strategies, referred to as PMS. The strategy consists of three parts: (1) exploration operator based on population evolutionary direction; (2) exploitation operator based on the direction of nondominated solutions linkage; (3) memory strategy based on the optimal solution set.

The rest of the paper is organized as follows. Section 2 presents background information. Section 3 describes the novel PMS in detail. Section 4 introduces the test problems and evaluation metric. Section 5 gives experimental results and analysis. Section 6 outlines the conclusions and future work.

2 Background

2.1 Dynamic multiobjective optimization

A minimization problem is considered here without loss of generality. The DMOP can be described as (Farina et al. 2004):

$$\begin{aligned} \left\{ {\begin{array}{l} \mathop {\min }\limits _{x\in \Omega } F(x,t)=(f_1 (x,t),f_2 (x,t),\dots ,f_m (x,t))^\mathrm{T} \\ s.t.\;\;g_i (x,t)\le 0\;i=1,2,\dots ,p;\;h_j (x,t)=0\\ \;j=1,2,\dots ,q \\ \end{array}} \right. \end{aligned}$$

where t is the time variable, \(x=(x_1 ,x_2,\dots ,x_n)\) is the n-dimensional decision variables bounded by the decision space \(\Omega \), \(F=(f_1 ,f_2,\dots ,f_m )\) presents the set of m objectives to be minimized the functions of \(g_i \le 0\;i=1,2,\dots ,p\) and \(h_j =0\;j=1,2,\dots ,q\) present the set of inequality and equality constraints.

Definition 1

Pareto Dominance: p and q are any two individuals in the population, p is said to dominate q, denoted by \(f(p)\prec f(q)\;\mathrm{iff}\;f_i (p)\le f_i (q)\forall i=\left\{ {1,2,\dots ,m} \right\} \) and \(f_j (p)<f_j (q)\exists j\in \left\{ {1,2,\dots ,m} \right\} \).

Definition 2

Pareto Optimal Set (PS). x is the decision variable, \(\Omega \) is the decision space, F is the objective function, thus the PS (Coello Coello 2007) is the set of all nondominated solutions and is defined mathematically as:

$$\begin{aligned} \mathrm{PS}: = \left\{ {x \in \Omega \not \exists |x^{*} \in \Omega ,F(x^{*} ) \prec F(x)} \right\} \end{aligned}$$

Definition 3

Pareto Optimal Front (PF). x is the decision variable, F is the objective function, thus the PF (Coello Coello 2007) is the set of nondominated solutions with respect to the objective space and is defined mathematically as:

$$\begin{aligned} \mathrm{PF}:=\left\{ {y=F(x)\vert x\in \mathrm{PS}} \right\} \end{aligned}$$

To distinguish between different types of DMOPs, Farina et al. (2004) identified four different types of DMOPs:

  • Type I PS changes with time, but PF remains fixed.

  • Type II Both PF and PS change with time.

  • Type III PF changes with time, but PS remains fixed.

  • Type IV Both PF and PS remain fixed.

In real life, the above four types of changes may occur at the same time when the problem is changed. However, we mainly research the first three types of changes.

2.2 General steps of dynamic multiobjective evolutionary algorithm

Dynamic multiobjective evolutionary algorithm is generally divided into the following steps:

Step 1. Initialize a population, set the related parameters;

Step 2. Optimization algorithm: this paper chooses RM-MEDA which will be introduced in Sect. 2.2.1 as the MOEA optimizer.

Step 3. Change detection: if the environmental change has been detected, the mechanism of change reaction will be called to respond to the changes; otherwise, continue to perform optimization algorithm. A good change detection operator can exclude external factors related to disturbance, and determine environmental changes correctly.

Step 4. The mechanism of change reaction: the mechanism can respond quickly and correctly to environmental change, so that the population can quickly converge to the new optimal solution set. General mechanism has the following kinds: re-initialize population, parameter adjustment, dynamic migration, memory and prediction.

Step 5. Judgment of terminal condition.

Figure 1 illustrates the flowchart of dynamic multiobjective evolutionary algorithm:

Fig. 1
figure 1

Flowchart of DMOEA

2.2.1 RM-MEDA

Estimation of distribution algorithm, referred to as EDA Larrañaga and Lozano (2001), is a class of random optimization algorithm which newly emerged on the field of evolutionary computation. Different from genetic algorithm, EDA is built on the basis of the estimation of distribution of solutions. EDA describes the distribution of solutions in the solution space by establishing a probability model, and then generates new population by random sampling for the probability model.

Under certain smoothness assumptions, it can be induced from the Karush–Kuhn–Tucker condition that the PS of a continuous MOP defines a piecewise continuous (\(m-1\))-dimensional manifold in the decision space. Therefore, on the basis of EDA, combined with the distribution features of solution set of continuous multiobjective problem, Zhang et al. (2008) proposed a regularity model-based multiobjective estimation of distribution algorithm in 2008, referred to as RM-MEDA.

RM-MEDA can make full use of the global information provided by current population to generate the new individual, without the need for local information directly used from population, which has a strong ability of heuristic optimization. Since RM-MEDA utilized the regularity property of continuous MOPs, it has been verified in Zhou et al. (2013) to be an excellent optimization algorithm for DMOPs, and performs better than NSGA-II-DE in the DMOEA framework.

2.3 Performance analysis for three kinds of DMOEA

The method based on prediction is one of the effective ways to solve DMOPs, which is to generate the next new population under environmental change by prediction. It can accelerate the convergence of algorithm to the new optimal solution. Meanwhile, the way combined with memory which reuses the past information under the same or similar environment is another new way to solve DMOPs. This paper mainly discusses the comparison of two representative prediction strategies and a hybrid strategy combined prediction and memory and, on this basis, we propose a new PMS to solve DMOPs.

  1. (1)

    Feed-forward prediction strategy (FPS) FPS (Hatzakis and Wallace 2006a) records information of boundary point on the adjacent PF of objective space, and predicts the location of PS after environmental change by autoregressive model (AR). The initial population in FPS is composed of three parts: the nondominated solution set, the dominated solution set and the solution set by prediction. Nondominated solutions make the algorithm more conducive to solve the DMOPs whose PS does not change with time, and dominated solutions maintain the diversity of population, using the algorithm to search for the new optimal solution.

  2. (2)

    Population prediction strategy (PPS) In PPS Zhou et al. (2013), the optimal solution set is divided into two parts: the population center and shape. PPS uses autoregressive model (AR) to predict a new population center at the next time by storing population center on a continuous time series. Meanwhile, it predicts a new population distribution of the next time by recording the shape of population in the last two moments. These two parts make up the initial population after change. Comparing these two strategies, FPS produces only a small part of the new population by prediction, the majority of its population is inherited from the final population before environmental change; PPS generates a whole new population by prediction, rather than just a few initial individuals. With the algorithm running, the accumulated history information increases in PPS, and the accuracy of prediction will be improved. Therefore, at the early stages, PPS is slightly worse than the FPS in performance due to insufficient historical information. At the latter stages, the accumulation of experience make PPS performance becomes stable, which is better than FPS (Zhou et al. 2013).

  3. (3)

    Predictive gradient strategy (dMO-EGS) dMO-EGS (Koo et al. 2009) consists of three parts: mutation, prediction and memory. To increase the diversity of population, firstly, dMO-EGS performs mutation operation for part of the nondominated solutions in current population. Mutation uses the sampling of normal distribution with expectation of zero and standard deviation of 0.01. In the aspect of prediction, dMO-EGS predicts the evolutionary gradient of current population by recording the final two previous positions of nondominated population center which have been searched and the last predictive evolutionary gradients, and updates some individuals of the population with the predicted gradient and the product of a scaling factor. Meanwhile, we select an individual randomly from the original population and compare with a new individual. If the new individual is dominated, the scaling factor will be negated. However, the predictive gradient is reversed permanently at the end of the updating only if the scaling factor is negative for at least half of the time to allow the possibility of random fluctuations. In the aspect of memory, dMO-EGS designs a memory method based on center and variance. Firstly, calculate the center vector \(C\tau \) and variance vector \(\widetilde{C}\tau \) of the nondominated solution set after environmental change, then \(C\tau \) and \(\widetilde{C}\tau \) will be combined into a memory item to be stored. For the each item of the memory, generate new solutions according to the normal distribution with the expectation of \(C\tau \) and variance of \(\widetilde{C}\tau \).

The three operators in dMO-EGS are effective in increasing the diversity of population, which can better adapt to the DMOPs whose degree of change is smaller. However, this gradient prediction of simple correlation based on the solution set applies only to the same or similar degree of change in the environment; when there is a large degree of the environmental change, the accuracy of prediction is difficult to be guaranteed. Moreover, the last searched information is not intact in its memory. This will be prone to distortion in the course of normal sampling, and there is greater influence on the convergence of algorithm.

3 Novel PMS for dynamic multiobjective optimization

3.1 Prediction strategy

To solve the problem which cannot respond quickly to changes in DMOPs, in a timely manner to search for the optimal solution set of new problem, this paper proposes a prediction strategy of both exploration and exploitation. Exploration uses exploration operator based on population evolutionary direction, and generates a bunch of guide individuals in population evolutionary direction to ensure the algorithm’s ability to search for the optimal solution space. Exploitation uses exploitation operator based on the direction of nondominated solutions linkage to improve the searching accuracy of algorithm under similar environmental change.

3.1.1 Exploration operator

Whenever the environmental changes, the population will spontaneously evolve towards the direction of the new Pareto optimal solutions. So taking advantage of this characteristic, let the population evolve independently for a short time in the initial stage of environmental change, then we judge the evolutionary direction of the population in the current environment by recording the different center positions of current population before and after the change.

Figure 2 illustrates the principle of exploration operator, for the final searched population (Population\(_{t-1} )\) in the last time, and the new population (Population\(_{t-1} +\Delta t)\) after evolving spontaneously for time \(\Delta t\), it is the connecting direction between these two different population center positions at different time that represents the predicted population evolutionary direction. On this basis, firstly, evenly select several initial individuals from current population, and then a certain number of individuals are evenly distributed from the initial individuals to the boundary of the decision space according to the predicted direction, which corresponds to the guide individual in Fig. 2. After nondominated sorting these individuals, the relatively better individuals are selected, namely, the individuals which are closest to \(\mathrm{PS}_{t}\) and connected by dotted lines.

Fig. 2
figure 2

Principle of exploration operator

Determining operator for the predicted population evolutionary direction is defined as:

$$\begin{aligned} \partial&= \left\{ {\partial _1 ,\partial _2,\dots ,\partial _i,\dots ,\partial _n } \right\} \nonumber \\&= \left\{ C_1^{t\!+\!\Delta t} \!-\!C_1^t ,C_2^{t\!+\!\Delta t}\!-\!C_2^t,\dots ,C_i^{t\!+\!\Delta t} \right. \nonumber \\&\left. \!-\!C_i^t, \dots ,C_n^{t\!+\!\Delta t}\!-\!C_n^t \right\} , \end{aligned}$$
(1)

where \(\partial \) represents evolutionary direction vector of population in decision space, \(i = 1,2,{\ldots },n\). \(C^t\) is the population center at time t, \(C^{t+\Delta t}\) is the population center after population evolve spontaneously for time \(\Delta t\).

As the center of nondominated solutions represents the position of population center best, so this can avoid the situation illustrated in Fig. 3. When the population is not completely converged, the number of nondominated solutions is smaller than the size of population, so there may be a small number of poor individuals affecting the center of the whole population. Therefore, to define the position of population center in formula (1) more accurately, the center of nondominated solutions will be indicated by the position of population center. Let \(P_{\mathrm{non-do}}^t =\left\{ {x^t} \right\} \) represents nondominated solutions at time t, and then the position of population center is defined as:

$$\begin{aligned} C_i^t =\frac{1}{\left| {P_{\mathrm{non-do}}^t } \right| }\sum \limits _{x_i^t \in P_{\mathrm{non-do}}^t } {x_i^t } \end{aligned}$$
(2)

where \(\left| {P_{\mathrm{non-do}}^t } \right| \) is the size of nondominated solutions, \(x_i^t\) is the individual at time t, \(i = 1,2,{\ldots },n\). Determine the population evolutionary direction by calculating the position of population center each dimension in the decision space before and after the change. The results are denoted by a set of vectors that forward is positive and backward is negative.

Fig. 3
figure 3

The impact of judging the evolutionary direction when population is not fully convergent

The judgment about population evolutionary direction provides a basis for predicting the location of Pareto optimal solutions more accurately, for which the exploration operator is needed to obtain a bunch of guide individuals on the direction. Firstly, the method of calculating the crowding distance (Deb et al. 2002) of current population will be used in the decision space, so that it can get some initial individuals more evenly. From the initial individuals to the boundary of the decision space, a bunch of detecting individuals are distributed equal distance intervals according to the direction predicted in formula (1). Finally, we get ObSize guide individuals by a nondominated sorting. Exploration algorithm based on population evolutionary direction is described in Algorithm 1.

figure a

3.1.2 Exploitation operator

Changes in the environment present regularity in many cases; the current environment may have the same or similar degree of change with the previous environment. Therefore, taking advantage of this feature is conducive to improve the searching accuracy of algorithm under similar environmental changes. As shown in Fig. 4, the connection vectors between centers of the final searched populations under each environment change represent the degree of change in the environment, and the connection vector on the adjacent time series may have similar direction and distance. In the current environment, based on the connection vector of a recent time series, some predicted solutions can be produced; there will be some errors between these solutions and the true Pareto optimal solutions, but these errors are within the acceptable range.

Fig. 4
figure 4

Principle of exploitation operator

This paper adopts an exploitation operator based on the direction of nondominated solutions linkage. Exploitation operator is defined as:

$$\begin{aligned} x(t+1)=x(t)+C(t)-C(t-1), \end{aligned}$$
(3)

where \(C(t)\) is the center of nondominated solutions obtained according to the formula (2) at time t, \(C(t-1)\) is the center of nondominated solutions at time \(t-1\). We predict the new optimal solutions, according to the center of nondominated solutions before and after the change of two recent different environments. Exploitation algorithm based on the direction of nondominated solutions linkage is described in Algorithm 2.

figure b

3.2 Memory strategy

For periodic DMOPs, the optimal solution under the new environment may return to the past-searched position. Therefore, making full use of the past-searched optimal solutions can speed up the convergence rate which is equivalent to iterate on the original searching process. This paper proposes a memory strategy based on optimal solution set preservation mechanism. When the environment changes, we nondominating sort these stored individuals in memory pool, and select optimal individuals which adapt best to the new environment. After that, we store the nondominated individuals of current population to memory pool.

As shown in Fig. 5, the individuals stored in memory pool are optimal with best convergence and diversity in past environment. These individuals can cover a period of environmental changes and apply to different environmental changes which have been detected. When a new change is detected, the optimal individuals in the last same environment will be selected by nondominated selection, and continue to evolve in new environment after retrieval. So the effect of distortion can be avoided due to normal distribution sampling by reusing directly the optimal solutions which are previously searched. This method improves searching accuracy and convergence speed. When the size of memory pool exceeds the value of set, we use the principle of FIFO to update memory pool, which ensures that the algorithm does not consume too much extra storage space and evaluation. Memory algorithm based on optimal solution set is described in Algorithm 3.

Fig. 5
figure 5

Principle of memory strategy

figure c

3.3 Detailed description of PMS

PMS iterates under the general framework of DMOEA, the purpose is to obtain new initial population after each environmental change, so that the new population can quickly respond to changes and track to the new optimal solutions. The PMS is described in detail in Algorithm 4.

figure d

4 Test instances and performance metrics

4.1 Test instances

This section presents the DMOPs test problems used in the experiment. FDA test suite has been proposed in Farina et al. (2004) and is often used to examine the performance of DMOEA. DMOP test suite has been proposed in Goh and Tan (2009) and is an extension to the FDA. The two test suites are linear correlation between the decision variables. F5–F9 proposed in Zhou et al. (2013) are the new DMOPS test problems. The characteristic of F5–F9 is nonlinear correlation between the decision variables. Especially for F9, the environment changes smoothly in most cases and, occasionally, the Pareto set jumps from one area to another area, and the difficulty of convergence is larger.

(1) FDA and DMOP test suite

FDA test suite is proposed by Farina and Deb et al. for assessing the performance of DMOEA. Four test problems FDA1–FDA4 will be used in this paper. DMOP test suite has been proposed by Goh and Tan et al. and is an extension of the FDA. Table 1 lists all the seven test problems and their PF and PS in detail. FDA4 is a three-objective problem among them.

Table 1 Test problems of FDA and DMOP

(2) F5–F9 test suite

F5–F9 are newly proposed by Zhou and Jin et al., which have nonlinear correlation between decision variables. Among them, F5-F7 are two-objective problems, F8 is a three-objective problem, and F9 is a complicated problem which is more difficult to converge than other test problems. Table 2 lists all the five test problems and their PF and PS in detail.

Table 2 Test problems of F5–F9

4.2 Performance metrics

Some metrics have been designed for dynamic optimization (Cámara et al. 2009; Tantar et al. 2011; Helbig and Engelbrecht 2013). In this paper, we firstly introduce the dynamic generational distance (DGD) (Goh and Tan 2009) and inverted generational distance (DIGD) (Zhou et al. 2013) metric for DMOPs. The DGD and DIGD metrics are defined as follows:

$$\begin{aligned}&\mathrm{DGD}=\frac{1}{\left| T \right| }\sum \limits _{t\in T} {\mathrm{GD}(\mathrm{PF}^t,P^t)} ,\nonumber \\&\quad \mathrm{GD}(\mathrm{PF}^t,P^t)=\frac{\sum \nolimits _{v\in P^t} {d(\mathrm{PF}^t,v)} }{\left| {P^t} \right| } \nonumber \\&\mathrm{DIGD}=\frac{1}{\left| T \right| }\sum \limits _{t\in T} {\mathrm{IGD}(\mathrm{PF}^t,P^t)} ,\nonumber \\&\quad \mathrm{IGD}(\mathrm{PF}^t,P^t)=\frac{\sum \nolimits _{v\in \mathrm{PF}^t} {d(v,P^t)} }{\left| {\mathrm{PF}^t} \right| } \end{aligned}$$
(4)

where \(\mathrm{PF}^t\) is a set of uniformly distributed Pareto optimal points in the PF at time t, and \(P^t\) is the solution obtained at time t. \(d(\mathrm{PF}^t,v)=\min _{u\in \mathrm{PF}^t} \sqrt{\sum \nolimits _{j=1}^m {\left( {f_j^{( u)} -f_j^{( v)} }\right) ^2}}\) is the distance between v and \(\mathrm{PF}^t\), \(d(v,P^t)=\min _{u\in P^t} \sqrt{\sum \nolimits _{j=1}^m {\left( {f_j^{( v)} -f_j^{( u)}}\right) ^2}}\) is the distance between v and \(P^t\), T is a set of discrete time points in a run and \(\vert T\vert \) is the cardinality of T. DGD evaluates convergence of the algorithm, DGD value is lower, and the solution obtained set has better convergence. DIGD is a comprehensive metric to evaluate the convergence and distribution; lower DIGD value means that solution set obtained has a better convergence and distribution.

5 Experiments

5.1 Compared strategies and parameter settings

The novel strategy which proposed in this paper is denoted as PMS, and in this section, PMS will be compared to other three strategies: forward-looking prediction strategy (FPS) (Hatzakis and Wallace 2006a), PPS (Zhou et al. 2013), and predictive gradient strategy(dMO-EGS) (Koo et al. 2009) (Since only we compared the impact between different strategies for solving DMOPs, so in this section, mutation, PMS in dMO-EGS are denoted as MPMS, and integrated into an unified optimization algorithm), we choose RM-MEDA mentioned in Sect. 2.2.1 as the MOEA optimizer. In PMS, the number of guide individual in exploration operator is Obsize = 10, the number of initial individual is OP = 10, the number of aliquots is OB = 9, and the time of evolving independently is \(\Delta t = 2\). The number of selected optimal individuals from memory pool is Msize = 5 (Three objectives: 10). Other parameter settings of three strategies use the given setting in Zhou et al. (2013) and Koo et al. (2009). The detailed settings of other parameters are shown in Table 3.

Table 3 Other parameters setting for experiment

Since the proposed PMS in this paper and MPMS in dMO-EGS need to consume a certain number of evaluations in the prediction and memory, to be fair, the algorithm iterations require removing the number of evaluations consumed at every environmental change and reduce the corresponding number of iterations. Therefore, the frequency of change is set to be \(\tau _T = 22\) in PMS and \(\tau _T = 23\) in MPMS. We run each algorithm 30 times for each test instance independently. Each simulation runs for 2,500 generations (PMS: 2,200 generations, MPMS: 2,300 generations) and each strategy tracks to 100 times of environmental changes. As the dynamic test problems introduced in Sect. 4.1 are all periods, according to the parameter setting of \(n_T\), the environment will change periodically with unequal frequency ranging from 2 to 40 in FDA and DMOP test suites, and change with the frequency of 40 in F5–F9 test suite. So to discuss the performances of different strategies in each period, the result of the experiment is divided into three stages except for the first environmental change. Each stage tracks to 33 times of environmental changes and its average is taken as the result.

5.2 The statistical results of evaluation

5.2.1 Results on FDA and DMOP

The statistical results of DGD and DIGD on FDA and DMOP over 30 runs can be found in Table 4. It is clear from Table 4 that:

  1. (1)

    Besides FDA2 and DMOP1 whose PS is fixed, the mean DGD and DIGD of PMS are less than the other three strategies, and the metric values are relatively average in three stages, and become more and more stable. Especially in the first stage, the metric values are greatly better than the other three strategies.

  2. (2)

    In the latter two stages, the DGD and DIGD of PPS are relatively average and slightly better than the PMS. But overall, the metric values are similar to PMS. On FDA2, FDA3 and DMOP1 these three test problems, PMS shows better performance than PPS.

  3. (3)

    On FDA4 which is a three-objective problem, the DGD and DIGD of PMS are better than the other three strategies in each stage.

It is not hard to explain the results, mainly because the exploration operator of PMS does not need the accumulation of experience, combined with the introduction of exploitation operator, which cannot only improve the depth of search, but also improve the accuracy of search. Therefore, it is able to respond more quickly to changes in the environment. Meanwhile, with the combination of two kinds of operators, the algorithm can adapt not only to a greater degree of environmental change, but also to the environment whose degree of change is smaller or similar. Compared with FPS which uses AR model and MPMS that uses predictive gradient, the prediction of PMS is more accurate, so PMS is greatly better than the other three strategies in the first stage. For problem whose PS is fixed, the diversity of the population will hinder the rapid convergence of population. So on FDA2 and DMOP1 in these two test problems whose PS is fixed, the stored portion of the previous population will improve the convergence of the algorithm.

Table 4 Statistical result of DGD and DIGD metrics for four strategies on FDA and DMOP

In addition, since the environmental periodic changes and the accumulation of experience, the convergence and diversity of PPS, MPMS and PMS will stabilize in the latter two stages due to the introduction of memory strategy. But the memory of MPMS is an incomplete memory strategy, so the performance is not as good as PPS and PMS. PPS predicts on the basis of memory, which is slightly better than the PMS on simple linear problem, but overall the difference is small. And on FDA3, the accumulated experience of PPS is not enough to make accurate predictions, which need more times iteration of algorithm, so the average performance metrics of PMS are better than PPS on FDA3.

On FDA4 which is a three-objective problem, performance of four strategies is not very different, and PMS performs better than the other three strategies at each stage, which indicates that PMS is more suitable for high-dimensional DMOPs.

5.2.2 Results on F5–F8

The statistical results of DGD and DIGD on F5–F8 over 30 runs can be found in Table 5. F5–F8 are harder than FDA and DMOP because of the nonlinear correlation between decision variables. Since the same overhead is assigned to all the algorithms on all the test problems, the statistical results on F5–F8 are worse than those on FDA and DMOP. However, according to Table 5, it is clear that the advantage of PMS is more obvious in solving such problems. Except for the latter two stages of F7, PPS is slightly better than PMS, the other performance metrics of PMS at each stage are better than the other three strategies.

Table 5 Statistical result of DGD and DIGD metrics for four strategies on F5–F8

5.2.3 Results on F9

The statistical results of DGD and DIGD on F9 over 30 runs can be found in Table 6. The experimental results are similar to those in the previous section. PMS also performs relatively stable convergence and diversity on complicated problems.

Table 6 Statistical result of DGD and DIGD metrics for four strategies on F9

5.3 Comparison of distribution of final populations obtained

To visually analyze the performance of each strategy, we choose these five typical test problems, FDA1, DMOP2, F5, F6, and F9, and draw the distribution of final populations obtained for four strategies for solving different problems at different time, shown in Figs. 6, 7, 8, 9, 10.

Fig. 6
figure 6

Solution sets founded by four strategies at six different time steps on FDA1

Fig. 7
figure 7

Solution sets founded by four strategies at eight different time steps on DMOP2

Fig. 8
figure 8

Solution sets founded by four strategies at eight different time steps on F5

Fig. 9
figure 9

Solution sets founded by four strategies at eight different time steps on F6

Fig. 10
figure 10

Solution sets founded by four strategies at eight different time steps on F9

By comparison, it is not hard to come with the same conclusion as in Sect. 5.2, the convergence and diversity of PMS are far better than FPS and PPS at the beginning stages of environmental change, which indicates that PMS is able to respond to environmental changes more quickly and accurately. There will be several times when the convergence and diversity of MPMS are poor, indicating that when a greater degree of environmental changes occurs, MPMS cannot make accurate prediction. To the later stage of running, PMS is the same as PPS, which has a better convergence and distribution, and slightly better than PPS on the nonlinear problems. As to the ability to solve complicated problem F9, the advantage of PMS is more obvious; the other three strategies cannot achieve better convergence and distribution, while the PMS can more accurately track to a new optimal solutions and obtain Pareto optimal solution set with better convergence and distribution. It indicates that PMS is more suitable for solving complicated nonlinear problems than the other three strategies.

5.4 Influence of different strategies

To consider the influence and necessity of each strategy, we choose FDA1, DMOP2, F6 and F7 to test. The results of average DIGD values versus the time with different strategies are shown in Fig. 11.

Fig. 11
figure 11

The average DIGD over 30 runs versus time on a FDA1, b DMOP2, c F6 and d F7

From Fig. 11, we can see that the algorithm which only employs prediction strategy performs well in terms of convergence and diversity at the early stages. Especially for solving the FDA1 and DMOP2 with linear correlation between the decision variables, the average DIGD of obtained solutions is well. But for F6 and F7 with nonlinear correlation between the decision variables, some deviations may be generated in the prediction, leading to the poor performance of the obtained solutions.

The algorithm which only employs memory strategy has the worst performance. The reason might be that the environment does not form a periodic change at the early stages. The stored history optimal solutions cannot help the population to evolve in current environment. But with the environmental periodic change, the memory strategy makes the performance of algorithm gradually improved by reusing the previously searched elite solutions. To further confirm the conclusion, we choose DMOP2 and draw the distribution of final obtained populations of the algorithm which only employs memory strategy in the same environment of different periods, shown in Fig. 12. We can see the change of performance of obtained solutions in the same environment of different time, which shows a gradual improved trend.

Fig. 12
figure 12

Solution sets founded by memory strategy at five different time steps on DMOP2

The strategy of combining prediction and memory has the best performance. The prediction can guarantee the algorithm to quickly respond to environmental changes, and in a timely manner to search for a new optimal solution set. Meanwhile, the memory strategy can improve the ability of dealing periodic problems and has reduced the error rate because of every prediction that may generate. The results indicate that PMS is very promising for dealing with dynamic environments.

6 Conclusions and future work

In this paper, we have proposed novel PMS to enhance the performance of multiobjective optimization evolutionary algorithms in dealing with dynamic environments. In the proposed prediction strategy, the combination of exploration and exploitation provides the basis for the algorithm to accurately predict the new optimal solution set, which accelerates the speed of response to environmental change; thus the optimal solution set can be obtained with good convergence and diversity at the initial stages. Meanwhile, the memory strategy allows the algorithm to better adapt to periodic problems, so that the algorithm can obtain a stable optimal solution set in the late stages of running. Compared with other three strategies both on seven traditional benchmark problems and on five newly appeared ones, PMS has shown faster response to the environmental changes than peer strategies in solving whether linear or nonlinear problems, with its solution set having better convergence and diversity. It has highlighted superiority for solving complicated nonlinear problems. Finally, please note that, when the population is not completely converged, the judgment errors may occur about evolutionary direction of the population in the current environment by recording the different center positions of nondominated solutions before and after the change. The prediction is not accurate and convergence is not good under some environmental changes. Therefore, our future work will be designing a more accurate prediction model. Furthermore, our focus in the future will also be the influence of different optimization algorithms for solving different problems.