Abstract
Decomposition-based multiobjective evolutionary algorithms (MOEA/D) have achieved great success in the field of evolutionary multiobjective optimization, and their outstanding performance in solving for the Pareto-optimal solution set has attracted attention. This type of algorithm uses reference vectors to decompose the multiobjective problem into multiple single-objective problems and searches them collaboratively, hence the choice of reference vectors is particularly important. However, predefined reference vectors may not be suitable for dealing with many-objective optimization problems with complex Pareto fronts (PFs), which can affect the performance of MOEA/D. To solve this problem, we introduce a reference vector initialization strategy, namely, scaling of the reference vectors (SRV), and also propose a new reference vector adaptation strategy, that is, transformation of the solution positions (TSP) based on the ideal point solution, to deal with irregular PFs. The TSP strategy can adaptively redistribute the reference vectors through periodic adjustment to endow that the solution set with better convergence and a better distribution. Both strategies are introduced into a representative MOEA/D, called 𝜃-DEA-TSP, which is compared with five state-of-the-art algorithms to verify the effectiveness of the proposed TSP strategy.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Multiobjective evolutionary algorithms (MOEAs) have received extensive attention in recent decades. Due to their superior performance in both research and practical problem optimization, they have been widely used in the fields of engineering [1,2,3,4] and scientific research [5,6,7,8]. In these fields, an increasing number of multiobjective optimization problems (MOPs) are emerging in which it is necessary to optimize three or more conflicting objectives simultaneously. Therefore, studying such many-objective optimization problems (MaOPs) has important practical significance.
Without loss of generality, the minimum MOP can be defined as:
where x=(x1,x2,⋯ ,xm)T is the decision vector, fi(x) represents the i-th objective function in the objective space, m denotes the number of objective functions, and Ω is the feasible region in the decision space. An MOP can also be called an MaOP if m> 3 [9,10,11]. So-called PFs are generated by these MOEAs to help decision-makers choose a solution that strikes a desirable balance among the various conflicting objectives.
In recent years, the emergence of many excellent algo rithms, such as the adaptive-reference-point-based nondominated sorting genetic algorithm (ANSGA-III) [12], MOEA/D [13], and the reference-vector-guided evolutionary algorithm (RVEA) [14], has enabled many MaOPs to be solved. In paticular MOEA/D, which was first proposed by Zhang, has been a trending research topic in recent decades. Without prior knowledge, this algorithm generates a series of uniformly distributed reference vectors in the first quadrant. The algorithm then uses this set of predefined reference vectors to decompose the problem into several subproblems. Reference vectors in different directions are used to search different areas of the objective space. As described by Liu [15], the predefined reference vectors are constructed from a set of reference points on the hyperplane \( {\sum }_{i=1}^{m} f_{i}=1 \) in the normalized m-dimensional objective space, called the normal boundary intersection (NBI) [16]. The basic assumption of MOEA/D is that a uniform distribution of the reference vectors can ensure the diversity of the solution set, as shown in Fig. 1(a).
However, such predefined reference vectors do not allow satisfactory handling of MaOPs with irregular PFs. Because the generated reference vectors are not evenly distributed along the true PF, the performance of MOEA/D can be greatly affected. Under the assumptions that the PF is symmetric and the m objectives have been normalized to 0<fi< 1, the curvature of the PF can be expressed as \( {f_{1}^{p}}+{f_{2}^{p}}+\cdots +{f_{n}^{p}}=1 \), where p> 0. When p> 1, the PF is concave, and most of the solutions obtained are sparse in the central area and dense at the corners. When p< 1, the PF is convex, and most of the solutions obtained are dense in the center area and sparse at the corners. For the concave and convex cases, the solutions specified by ten uniformly distributed reference vectors on the PF of a simple two-objective problem are shown in Fig. 1(b) and (c) respectively.
According to Pan [17], irregular PFs also include discontinuous and degenerate Pareto surfaces. A discontinuous PF is defined as a PF whose geometry is incomplete and the PF of a degenerate MOP has fewer than m-1 dimensionals.
Therefore, to solve such problems, many scholars have set out to study how to ensure that the reference vectors in MOEA/D are evenly distributed on the real PF.
The improved MOEA/D (MOEA/D-TPN) proposed by Jiang [18] adopts a two-phase strategy. In the first phase, the unevenness of the PF of the MOP is judged, and then the reference vectors are adjusted in accordance with the results of the first stage and a corresponding decision on how to allocate computational resources, however, this algorithm introduces too many parameters, making it heavily dependent on the values of these parameters.
Dong [19] proposed an adaptive reference vector adjustment strategy based on chain segmentation, in which a chain topology relationship is established for the subproblems, the step length between adjacent individuals is calculated to divide the chain equally, and then, the positions of the reference vectors are determined through a distance strategy and the calculation of the ratios between two adjacent individuals. Finally, vector addition is used to obtain a more uniformly distributed reference vector set. However, the structure assumed for the topological relationship is suitable only for two-dimensional and three-dimensional MOPs.
Qi [20] proposed an improved MOEA/D with adpative weight adjustment (MOEA/D-AWA), in which the reference points are dynamically adjusted during the optimization process. The entire optimization process is also divided into two stages to balance convergence and diversity. Specifically, a sparsity level calculation formula is introduced to calculate the sparsity of each subproblem; then, overcrowded subproblems are periodically deleted in accordance with a preset value, and new reference vectors are generated in relatively sparse positions. However, the external archive mechanism and the complicated sparsity level calculation seriously affect the efficiency of the algorithm.
On the other hand, some scholars have combined machine learning methods to generate new reference vectors. Gu [21] proposed a reference point design method based on a self-organizing map (SOM). This method uses the weights of the neurons as the reference vectors and uses the objective vector of the nearest population to periodically train an SOM network of N neurons. Wu [22] proposed a learning decomposition method to adaptively determine the reference vectors. This method is divided into a learning module and an optimization module. Specifically, in the learning module, the current nondominated solutions are used as training data, and Gaussian process (GP) regression is used to train an analysis model. In the optimization module, feature vectors are extracted to adjust the reference vectors. However, the learning process is time consuming and can easily lead to overfitting.
For MaOPs with convex, concave, degenerate, or discontinuous PFs, many algorithms based on reference vectors may not be able to solve them well or may require considerable computational resources to solve them [23]. To solve such problems more effectively, we propose our TSP strategy, which can be combined with other decomposition-based MOEAs.
The main contributions of this paper can be summarized as follows.
-
1.
The proposed TSP strategy can be easily integrated into other algorithms to improve the convergence of the population.
-
2.
𝜃-DEA-TSP is proposed, which incorporates the TSP strategy and considers the diversity and convergence of the population throughout the whole execution of the algorithm, instead of focusing on the diversity of the population at the end of the algorithm, to avoid the algorithm falling into a local optimum in the early stage. At the same time, the algorithm has lower computational complexity.
-
3.
This research is expected to provide better solutions for MaOPs facing similar problems.
The rest of this paper is organized as follows. In Section 2, we introduce the background of our work, including the SRV strategy and the previously proposed strategy for the transformation of the solution locations (TSL), as well as the motivation for our work. In Section 3, we describe in detail the framework of the proposed TSP strategy and 𝜃-DEA-TSP. In Section 4, our experimental results illustrate the effectiveness of the proposed strategy and the comprehensive performance of 𝜃-DEA-TSP on MaOPs with complex PFs. Finally, in Section 5, this paper is summarized, and future research topics are proposed.
2 Related work
To solve MaOPs with complex PFs, Liang [24] proposed the SRV and TSL strategies for modifying the reference vectors.
Specifically, the SRV strategy scales the reference vectors generated via the NBI method on the basis of a predesignated center vector, to ensure that the reference vectors can be evenly distributed on PFs that are concave or convex in nature.
In the NBI method, the number of reference vectors N is counted as follows:
where m is the number of objectives, and H represents the the number of divisions on each axis. The N reference points are evenly distributed on the entire normalized hyperplane with m objectives.
Therefore, the reference points are mostly generated near the boundary, and only a few points are generated on the inner part of the hyperplane, where H ≥ m. In the SRV stragety, divides the boundary and inner layers are divided using values of as H1 and H2, respectively, to generate two-layer reference points. The N reference vectors are then generated as follows:
However, when the intent is to solve an MaOP with a concave or convex PF, the reference vectors thus generated still will not be evenly distributed on the PF, especially in the case of MaOPs. Therefore, a scaling function (SF) is applied to the original reference vectors, as expressed below:
Here, i = 1,2,⋯ ,N, Wi represents the i-th reference vector. Vc represents the designated center vector. The ∘ operator denotes the Hadamard product and ||⋅|| denotes the calculation of the norm. ri is the scale of the i-th reference vector, and pi ∈[0,1] represents the similarity between the i-th reference vector and the center vector.
To solve MaOPs with degenerate and discontinuous PFs, Liang [24] also proposed the TSL strategy. Specifically, this strategy consists of comparing the Euclidean distances between pair of solutions, filtering out less diverse solutions based on their the Euclidean distances, removeing the overcrowded solutions and transforming the remaining highly diverse solutions into a new set of reference vectors to deal with the degenerate or discontinuous PF. This strategy solves the problem that the reference vectors cannot be evenly distributed on the PF in the case of complex MaOPs. However, we have found that the TSL strategy considers only the diversity of the solutions when deleting overcrowded solutions. It ignores the convergence of the solutions, leading to population degradation. In the proposed algorithm 𝜃-DEA*, the TSL strategy is applied at the end of the algorithm to increase the diversity of the solution set. However, this can cause the algorithm to fall into a local optimum early.
In this article, to solve the problem that the TSL strategy results in degeneration of the population in the face of complex MaOPs, we propose the TSP strategy as an alternative. In the TSP strategy, when a new set of reference vectors is being constructed two solutions to be deleted are compared with the current ideal point by calculating the Euclidean distance, and the better solution is retained to prevent population degradation. In addition, the TSP strategy is applied to adjust the reference vectors throughout the whole execution of 𝜃-DEA-TSP, thereby guaranteeing good diversity of the population while improving convergence.
3 Methodologies
In this section, we describe the TSP strategy. Specifically, to better deal with MaOPs with degraded or discontinuous PFs, the positions of solutions are converted into new reference vectors to prevent the degradation of the population while maintaining its diversity. This strategy can be integrated along with the SRV strategy into popular MOEAs based on reference vectors to enhance the evolutionary search process.
3.1 Transformation of solution positions based on the ideal point solution
Oriented toward MaOPs with degenerate or discontinuous PFs, we propose the TSP strategy to fit an irregular PF by transforming the position of a solution with good distribution and convergence characteristics to obtain a new reference vector. Specifically, in the TSP strategy, the distances between pairwise solutions are initially calculated. If such a distance is less than a preset threshold, then the corresponding solution with poorer diversity is selected to calculate the Euclidean distance between this solution and the ideal point. Solutions with larger Euclidean distances, that is, poorer convergence are deleted, and the remaining solutions are converted into reference vectors.
The TSP process is shown in Algorithm 1. First, the algorithm initializes a population archive PA as the entire population, in which each individual is normalized. Second, the Euclidean distances Dist of the pairwise solutions in PA are calculated. The Euclidean distance is expressed as follows:
where X=(x1,x2,⋯ ,xm)T and Y=(y1,y2,⋯ ,ym)T are the decision vectors, and ω is the feasible region in the decision space. Lines 3-10 of the algorithm filter out the two solutions corresponding to the minimum Euclidean distance based on Dist, calculate the Euclidean distances between these two solutions and the current ideal point, and delete the solution corresponding to the larger distance. This process is repeated until Distmin is greater than the threshold 𝜖. The aim of Lines 11-15 is to find the solution among the remaining solutions that is nearest to be better solution of each solution pair processed as described above and generate a new solution S at the midpoint of the two. The new solution S is stored in PA, Dist is updated, and the process is repeated until the number of solutions in PA is N. In Lines 17-19, the solutions in PA are converted into new reference vectors to guide the generation of new offspring.
To illustrate the process more clearly, we present an example in Fig. 2. In Fig. 2(a), the algorithm finds four points corresponding to two Euclidean distances smaller than our preset threshold 𝜖, namely, points B and C and points D and E. Second, the Euclidean distances between B, C, D and E, and the current ideal point Z∗ are calculated, as shown in Fig. 2(b). Next, the algorithm deletes the individuals farther from the ideal point between B and C and between D and E, that is, B and D, as shown in Fig. 2(c). Finally, the individuals closest to B and D other than C and E are found, namely, A and F. New individuals \(B^{\prime }\) and \(D^{\prime }\) are generated at the midpoints between A and C and between E and F, as shown in Fig. 2(d).
3.2 The framework of 𝜃-DEA-TSP
The SRV and TSP strategies are two independent strategies that can be easily integrated into MOEAs. The algorithm framework is shown in Algorithm 2. First, the population Pop is initialized and reference vectors W are generated using the NBI method. In Lines 2-3, the SRV strategy mentioned in Section 2 is used to scale the reference vectors, and then, the objective normalization flag is set. Lines 4-14 are the main part of the algorithm. The purpose of line 5 is to generate N offspring. Then, based on the Boolean value of the objective normalization flag, it is judged whether it is necessary to deal with the problem of different ranges for each objective. Subsequently, the SRV strategy is used to adjust the generated reference vectors. We introduce a frequency parameter fr to control the frequency of application of the TSP strategy so that it will be applied throughout the entire execution of the algorithm instead of taking effect only at the end of the algorithm. Finally, N individuals are obtained through environmental selection.
3.3 Computational complexity
Since it is still necessary to calculate the Euclidean distances between pairs of solutions, the computational complexity of the TSP strategy is the same as that of the original TSL strategy, which is O(N2). However, the overall algorithm frameworks differ between the TSL and TSP strategies. In the overall framework of 𝜃-DEA∗, the TSL strategy takes effect every constant generation at the end of the algorithm, and the algorithm complexity is O(\(N^{2}* \frac {T_{max}-0.75*T_{max}}{C}\)), where Tmax is the maximum number of function evaluations (FEs), and C is the preset constant of the algorithm. In contrast, this article introduces the frequency parameter fr, and the complexity of 𝜃-DEA-TSP is O(N2 ∗\(\frac {T_{max}}{f_{r}*T_{max}})\). In the case of many FEs, the TSL strategy will take effect frequently. Therefore, in this case, the complexity of 𝜃-DEA-TSP will be better than that of 𝜃-DEA∗.
4 Experiment study
In this section, we combine the proposed strategy with 𝜃-DEA to obtain the resulting 𝜃-DEA-TSP algorithm for comparison with several typical algorithms: 𝜃-DEA* [24], ANSGA-III [25], RVEA [14], and the guided nondominated sorting genetic algorithm (gNSGA-II) [26]. 𝜃-DEA* is the algorithm proposed by Liang, which integrates the SRV and TSL strategies into 𝜃-DEA. 𝜃-DEA* is an algorithm based on decomposition. It incorporates the TSL strategy and selects a set of excellent solutions to guide the reference vectors for self-adaptation, enabling some improvements in solving MaOPs. ANSGA-III is adaptive in the dynamic deletion and inclusion of new reference points. It is an algorithm proposed by Deb based on the nondominated sorting genetic algorithm version II (NSGA-II) framework for solving MaOPs, providing denser PFs under the same computational workload and demonstrating excellent performance in dealing with constrained and unconstrained MaOPs. RVEA is also an optimization algorithm based on decomposition. It applies an adaptive strategy for dynamic adjustment of the reference vector distribution in accordance with the objective function, and reference vector regeneration strategies are used to deal with irregular PFs. Many papers have proven its effectiveness in solving MaOPs. gNSGA-II is an evolutionary algorithm based on dominance that introduces the concept of g-dominance. Through comparisons with this algorithm, the performance of the g-dominance method for solving MaOPs has been proven. Since DTLZ1 is an MOP with a regular PF, we test the above mentioned 5 algorithms on DTLZ2-7 [27]. Each test problem is run independently 20 times, and the average and standard deviation are taken. The above tests are all carried out on the PlatEMO [28] platform. To ensure fairness, the parameters of the algorithms mentioned above are all set to their default values in PlatEMO.
In the following sections, we first give the parameter settings for the simulation experiments. Second, we briefly introduce the evaluation metrics used in the comparative study. Then, we analyze the influence of the value of the parameter fr on the performance of our algorithm. Finally, we compare the experimental results obtained by independently running each algorithm 20 times on each test problem. In these comparisons, the symbol “-” indicates that 𝜃-DEA-TSP is significantly better than the other algorithm considered for comparison, the symbol “+” indicates that 𝜃-DEA-TSP is significantly inferior to the other compared algorithm, and the symbol “≈” indicates that the result obtained is not significantly different from that of the other algorithm.
4.1 Parameter setting
The parameters of all algorithms considered for comparison are set as follows: In 𝜃-DEA-TSP, 𝜃-DEA∗, ANSGA-III, RVEA, and gNSGA-II, the simulated binary crossover (SBX) operator and polynomial mutation are used to solve the test problems, and the detailed settings are shown in Table 1. To ensure fairness, the default parameters values are used for binary crossover and polynomial mutation. The relevant parameter settings of the test function are summarized in Table 2, where M is the number of objectives; D is the number of decision variables, and k is a parameter used to artificially change the length of a decision variable.
4.2 Performance metrics
In the experiments, we measure the performance of various algorithms on the test problems. We use the following three metrics to evaluate the convergence and comprehensive performance of the solution set.
(1) Generation Distance (GD) [34]
The GD metric is used to calculate the distance between the obtained solution set and the true Pareto solution set, which is defined as follows:
where n denotes the number of obtained solutions, di is the Euclidean distance between the obtained solution i and its nearest true solution, and p denotes the number of objectives. The smaller the GD value is, the stronger the algorithm performance.
(2) Inverse Generation Distance (IGD) [35]
The IGD indicator is calculated as the average value of the distance from each reference point to the nearest solution, and can be used to simultaneously evaluate the convergence and distribution of the solution set calculated by an algorithm. The indicator is defined as follows:
where P represents the solution set obtained by the algorithm, P∗ denotes the solution set of the ideal PF, and d(x∗, P) denotes the minimum Euclidean distance from x∗ to P. The smaller the IGD value is, the stronger the algorithm performance.
(3) Hypervolume Distance (HV) [36]
HV is another way to measure the comprehensive performance. It calculates the volume of the objective space enclosed by the reference points and the i-th non-dominated solution. The larger the HV value, the better the comprehensive performance. The metric can be defined as follows:
where ai is the hypercube formed by the solution x and the origin of the coordinate axes as diagonal lines. The space covered by ai is volume(ai).
Notably, the widely used evaluation metric HV is not selected for the DTLZ problems because HV and IGD show certain contradictions on concave PFs [37]. For MaOPs whose PFs are nonconvex, HV has a certain error [36]. The PF of DTLZ2-6 is concave, as shown in Table 2. GD is an indicator of convergence and, can still measure the convergence of an algorithm even if the diversity of the solution set is not good. To prove the effectiveness of the convergence enhancement strategy, we choose the GD and IGD metrics to evaluate the results on the DTLZ problems.
4.3 Sensitive analysis of parameter f r
fr is an important parameter in the reference vector adaptation strategy. Its value will affect the frequency of reference vector adaptation, and thus, in turn, affect the convergence and distribution of the solution set found by the algorithm. To study the impact of this parameter on the algorithm performance, with fr= 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, and 1, we tested the algorithm on DTLZ2-7 with 6, 8, 10, and 15 objectives, and all parameters except fr remaining unchanged. Fig. 3 shows the comprehensive performance on the test problems when fr takes each value. With taking different values of fr, the IGD ranking of 𝜃-DEA-TSP fluctuates greatly on the DTLZ2-7 test problems. However, it can be seen from Fig. 3 that when fr= 0.1 and M= 8, 𝜃-DEA-TSP has the best IGD ranking on the DTLZ2-7 test problems. Therefore, for the comparative experiment, we choose fr= 0.1 as the value of the frequency parameter of 𝜃-DEA-TSP.
4.4 Comparison experience
4.4.1 Test problems
In this section, 𝜃-DEA-TSP is compared with 𝜃-DEA∗, ANSGA-III, RVEA, and gNSGA-II on DTLZ2-7 [38], IDTLZ1-2 [12] and DTLZ5IM [33]. Here, the numbers of objectives (m values) of DTLZ2-7 are 6, 8, 10, and 15. Tables 3 and 4 show the performance of the algorithms in terms of the GD metric and the IGD metric, respectively, on DTLZ2-7.
Experiments were run 20 times independently on the DTLZ2-7 test problems to compare the convergence performance of 𝜃-DEA-TSP, 𝜃-DEA∗, ANSGA-III, RVEA, and gNSGA-II as represented by the GD metric. The mean and variance of the GD values are listed in Table 3. In Table 3, the average GD value of 𝜃-DEA-TSP is better than those of the other algorithms on 50% of the 24 problems tested. In particular, 𝜃-DEA-TSP shows obviously superior performance on DTLZ4-7. Especially for the concave and multipeak DTLZ6 problem [39], the results for the 6-, 8-, 10-, and 15-objective versions of this problem are all better than those of the other four algorithms. On 8-objecitve DTLZ6, the GD value of 𝜃-DEA-TSP is 64% lower than that of the second-ranking algorithm 𝜃-DEA∗. On the 10-objective DTLZ6 problem, the GD value of 𝜃-DEA-TSP is 54% lower than that of the second-ranking algorithm 𝜃-DEA∗.
The experimental results in terms of the IGD values on DTLZ2-7 are shown in Table 4. We can see that when solving DTLZ4-7, 𝜃-DEA-TSP exhibits excellent performance. Even when the number of objectives is large, the IGD value of this algorithm is still better than those of the other algorithms. It is not difficult to see that 𝜃-DEA-TSP is highly very competitive on such problems. In addition, it can be seen from the experimental results that RVEA shows excellent performance on DTLZ2 and DTLZ3 and on other test problems with regular PFs. However, on DTLZ5-7, which have discontinuous or degraded PFs, the performance of our algorithm is significantly better than the other four, thus proving the effectiveness and pertinence of the TSP strategy proposed in this article. To more intuitively illustrate the performance of the five algorithms on DTLZ2-7 for different numbers of objectives, we plot their ranking histograms in Fig. 4. The x-axis represents the number of objectives, the y-axis represents the algorithm, and the z-axis represents the average ranking. The lower the column height in this figure, the higher the ranking of the corresponding algorithm, and the better its performance. We can see that the performance of the 𝜃-DEA-TSP and RVEA algorithms is relatively stable, while the ranking of 𝜃-DEA∗ fluctuates greatly for problem with 8 objectives and 15 objectives. The performance of the ANSGA-III worsens as the number of objectives increases. Compared with the other four algorithms, gNSGA-II performs poorly on these problems. The optimization results of the 5 compared algorithms for 15-objective DTLZ5 and 6-objective DTLZ7 are shown in Figs. 5 and 6.
𝜃-DEA* selects excellent solution sets only in the middle and late stages of the algorithm and generates reference vectors to guide the algorithm. A strategy for screening the solution sets that places too much emphasis on diversity may result in the degradation of the solution set. RVEA optimizes the reference vectors globally, which is not beneficial for subregion-specific solution sets. Although ANSGA-III modifies the traditional concept of dominance applied in NSGA-II, it also introduces extreme point calculation and target normalization strategies to enhance the convergence of the algorithm. However, the experimental results are still not very good. When 𝜃-DEA-TSP solves a problem whose PF has degenerate or discontinuous characteristics, it will dynamically screen the excellent solution set throughout the whole process, delete locally inferior solutions, and generate reference vectors from the results to better guide the evolution of the algorithm. Hence, the 𝜃-DEA-TSP algorithm performs better on DTLZ2-7 in terms of both the GD and IGD evaluation metrics.
In Fig. 5, on the one hand, we can see that the convergence of gNSGA-II, ANSGA-III, and 𝜃-DEA∗ is poor, which may be because they lack convergence improvement methods or place too much emphasis on the diversity of the solution set. On the other hand, RVEA has poor diversity on this problem, which may be because, in high-dimensional situations, its current reference vectors tend to be excessively concentrated in a certain location, and its adaptive reference vector strategy dose not improve this situation. In comparison, 𝜃-DEA-TSP does not fully converge, but because it considers both convergence and diversity, it performs significantly better than the algorithms. In Fig. 6, with an increase in the number of dimensions, the selection pressure of the dominance-based algorithms decreases sharply, preventing them from converging to the optimal PF, and gNSGA-II does not fully converge. In contrast, the convergence and diversity performance of 𝜃-DEA-TSP, 𝜃-DEA∗, RVEA and ANSGA-III on 6-objective DTLZ7 are all good, and 𝜃-DEA-TSP is slightly better than the other two.
To further analyze and compare the experimental results, we present boxplots of the experimental results for some test problems, namely, 8-objective DTLZ5, 10-objective DTLZ5, 8-objective DTLZ6, and 10-objective DTLZ7 in Fig. 7. The average IGD values of our algorithm on 8-objective DTLZ5 and 10-objective DTLZ5 are significantly better than those of the other four algorithms. It is not difficult to see in Fig. 6(a) and (b) that the stability of 𝜃-DEA- TSP is also significantly better than that of the other four algorithms. It can be seen from the boxplot for 8-objectives DTLZ6 that the IGD values of 𝜃-DEA-TSP, 𝜃-DEA*, and RVEA are relatively close and the fluctuations of the IGD values of these three algorithms are relatively small. The boxplot for 10-objective DTLZ7 shows that, except for the outliers, the IGD values of 𝜃-DEA-TSP are relatively close to those of ANSGA-III. Among the compared algorithms, the fluctuation in the performance of gNSGA-II on this test problem is particularly obvious. In addition, RVEA yields too many outliers, although the fluctuation in the IGD values is similar to that of the other three algorithms.
The experimental results for the IGD values on DTLZ5IM and IDTLZ1-2 for all dimensionalities are shown in Table 5. For these test problems in all numbers of dimensions, 58.33% of the experimental results of 𝜃-DEA-TSP are better than those of the other algorithms considered for comparison. Especially for DTLZ5IM and 6-objective IDTLZ2, 𝜃-DEA-TSP shows excellent performance. A feature of IDTLZ1-2 is that many reference points created on the standardized hyperplane do not have a Pareto optimal advantage. 𝜃-DEA-TSP, 𝜃-DEA* and ANSGA-III all benefit from their adaptive strategies for adaptively screening reference points and reallocating them to obtain Pareto-optimal solutions. Therefore, they perform well on these two test problems. However, gNSGA-II and RVEA lack strategies for local adaptation, leading to unsatisfactory results.
The experimental results for the HV values on DTLZ5IM and IDTLZ1-2 for each dimensionality are shown in Table 6. The experimental results of 𝜃-DEA-TSP are better than those of the other compared algorithms on half of these test instances. The PF of DTLZ5IM is nonconvex. By comparing the experimental HV and IGD results, it can be seen that the HV and IGD values may contradict each other in the testing of the PF. Nearly half of the HV values for gNSGA-II are 0, because all obtained solutions do not form a supervolume with the lowest point.
4.4.2 Real world problem
To verify the performance of the proposed algorithm on real problems, we also conduct a comparative analysis of this algorithm on the multipoint distance minimization problem (M-DMP) [40]. The problem of minimizing the distance between a point and a set of predetermined fixed target points has many advanced applications in navigation and layout optimization [30]. In this section, the D value of the problem is set to 2, and the upper and lower bounds are 100 and -100 [40], respectively. The other experimental parameters are the same as above. For real problems, we usually choose HV to measure the quality of the solution set.
The experimental results for versions of M-MDP with different numbers of dimensions are shown in Table 7. It can be seen that the proposed algorithm performs very well on this real problem, and the experimental results of gNSGA-II are also surprisingly good, perhaps because the changes to the traditional concept of dominance introduced by the g-dominance concept applied in this algorithm allow it to perform only slightly worse than 𝜃-DEA-TSP on this problem. In contrast, because 𝜃-DEA* and ANSGA-III sacrifice their convergence to preserve the diversity of the solution set, their solution sets do not form a hypervolume with the reference point of the HV metric for 15-objective M-DMP, resulting in a result of 0.
5 Conclusion
We propose the TSP strategy to balance the convergence and diversity of the population during evolution and correspondingly propose the 𝜃-DEA-TSP algorithm. The proposed algorithm has a lower computational cost on MaOPs, and can better solve problems with complex PFs and balance the convergence and diversity of the solution set throughout the whole execution of the algorithm. For classic test problems, recent test problems and a real world problem with different numbers of objectives, the GD, IGD and HV values are better than those of other algorithms on at least half the tested problem instances, although some of the algorithms considered for comparison already perform extremely well on MaOPs, thereby confirming the effectiveness of the proposed MOEA. However, as described by Aggarwal [41], in the case of high dimensions, the Euclidean distance is not a good metric. Therefore, we must admit that in MaOPs, the proposed TSP strategy will still have shortcomings. In future work, we need to find another method of measuring the diversity and convergence between individuals to replace the calculation of the Euclidean distance in the TSP strategy.
References
Alkebsi K, Du W (2021) Surrogate-assisted multi-objective particle swarm optimization for the operation of co2 capture using vpsa. Energy 224(3):120078
Tawhid M A, Savsani V (October 2018) A novel multi-objective optimization algorithm based on artificial algae for multi-objective engineering design problems. Appl Intell 48(10):3762?3781. https://doi.org/10.1007/s10489-018-1170-x
Mirjalili S, Jangir P, Saremi S (January 2017) Multi-objective ant lion optimizer: A multi-objective optimization algorithm for solving engineering problems. Appl Intell 46(1):79?-95. https://doi.org/10.1007/s10489-016-0825-8
Khan I, Maiti M K, Basuli K (2020) Multi-objective traveling salesman problem: an abc approach. Appl Intell, 2
Li H, Deb K, Zhang Q, Suganthan P N, Chen L (2019) Comparison between moea/d and nsga-iii on a set of many and multi-objective benchmark problems with challenging difficulties. Swarm and Evolutionary Computation
Bugingo E, Zhang D, Chen Z, Zheng W (2020) Towards decomposition based multi-objective workflow scheduling for big data processing in clouds. Clust Comput, pp 1–25
Panagant N, Pholdee N, Bureerat S, Yildiz A R, Mirjalili S (2021) A comparative study of recent multi-objective metaheuristics for solving constrained truss optimisation problems. Archives of Computational Methods in Engineering, pp 1–17
Lu C, Gao L, Pan Q, Li X, Zheng J (2019) A multi-objective cellular grey wolf optimizer for hybrid flowshop scheduling problem considering noise pollution. Appl Soft Comput 75:728–749
Liang Z, Hu K, Ma X, Zhu Z (2019) A many-objective evolutionary algorithm based on a two-round selection strategy. IEEE Transactions on Cybernetics, PP(99)
Zhou J, Yao X, Gao L, Hu C (2021) An indicator and adaptive region division based evolutionary algorithm for many-objective optimization. Appl Soft Comput 99:106872
Luo J, Huang X, Yang Y, Li X, Feng J (2019) A many-objective particle swarm optimizer based on indicator and direction vectors for many-objective optimization. Inf Sci, 514
Fellow, IEEE, Jain H, Deb K (2014) An evolutionary many-objective optimization algorithm using reference-point based nondominated sorting approach, part ii: Handling constraints and extending to an adaptive approach. IEEE Trans Evol Comput 18(4):602–622
Zhang Q, Li H (2007) Moea/d: A multiobjective evolutionary algorithm based on decomposition. IEEE Trans Evol Comput 11(6):712–731. https://doi.org/10.1109/TEVC.2007.892759
Cheng R, Jin Y, Olhofer M, Sendhoff B (2016) A reference vector guided evolutionary algorithm for many-objective optimization. IEEE Trans Evol Comput, 20(5)
Liu H-L, Gu F-, Cheung Y- (2010) T-moea/d: Moea/d with objective transform in multi-objective problems. In: 2010 international conference of information science and management engineering, vol 2, IEEE, pp 282–285
Das B I, Dennis J E (1998) Normal boundary intersection: A new method for generating pareto optimal points in multicriteria optimization problems
Pan L, He C, Tian Y, Su Y, Zhang X (2017) A region division based diversity maintaining approach for many-objective optimization. Integrated Computer-Aided Engineering 24(3):279–296
Jiang S, Yang S (2015) An improved multiobjective optimization evolutionary algorithm based on decomposition for complex pareto fronts. IEEE transactions on cybernetics 46(2):421–437
Dong Z, Wang X, Tang L (2020) Moea/d with a self-adaptive weight vector adjustment strategy based on chain segmentation. Inf Sci 521:209–230
Qi Y, Ma X, Liu F, Jiao L, Sun J, Wu J (2014) Moea/d with adaptive weight adjustment. Evolutionary computation 22(2):231–264
Gu F, Cheung Y-M (2017) Self-organizing map-based weight design for decomposition-based many-objective evolutionary algorithm. IEEE Trans Evol Comput 22(2):211–225
Wu M, Li K, Kwong S, Zhang Q, Zhang J (2018) Learning to decompose: A paradigm for decomposition-based multiobjective optimization. IEEE Trans Evol Comput 23(3):376–390
Zhang Q, Zhu W, Liao B, Chen X, Cai L (2018) A modified pbi approach for multi-objective optimization with complex pareto fronts. Swarm and Evolutionary Computation 40:216–237
Liang Z, Hou W, Huang X, Zhu Z (2019) Two new reference vector adaptation strategies for many-objective evolutionary algorithms. Inf Sci 483:332–349
Jain H, Deb K (2013) An evolutionary many-objective optimization algorithm using reference-point based nondominated sorting approach, part ii: Handling constraints and extending to an adaptive approach. IEEE Transactions on evolutionary computation 18(4):602–622
Molina J, Santana L V, Hernández-Díaz A G, Coello CA C, Caballero R (2009) g-dominance: Reference point based dominance for multiobjective metaheuristics. Eur J Oper Res 197(2):685–692
Deb K, Thiele L, Laumanns M, Zitzler E (2002) Scalable multi-objective optimization test problems. In: Proceedings of the 2002 Congress on Evolutionary Computation. CEC’02 (Cat. No. 02TH8600), vol 1, IEEE, pp 825–830
Tian Y, Cheng R, Zhang X, Jin Y (2017) Platemo: A matlab platform for evolutionary multi-objective optimization [educational forum]. IEEE Comput Intell Mag 12(4):73–87
Yuan Y, Xu H, Wang B, Yao X (2015) A new dominance relation-based evolutionary algorithm for many-objective optimization. IEEE Trans Evol Comput 20(1):16–37
Tian Y, Cheng R, Zhang X, Su Y, Jin Y (2018) A strengthened dominance relation considering convergence and diversity for evolutionary many-objective optimization. Evolutionary Computation, IEEE Transactions on
Guo X, Wang X, Wei Z (2015) Moea/d with adaptive weight vector design. In: 2015 11th international conference on computational intelligence and security (CIS), IEEE, pp 291–294
Tian Y, Cheng R, Zhang X, Cheng F, Jin Y (2017) An indicator-based multiobjective evolutionary algorithm with reference point adaptation for better versatility. IEEE Trans Evol Comput 22(4):609–622
Deb K, Saxena D K (2005) On finding pareto-optimal solutions through dimensionality reduction for certain large-dimensional multi-ob jective optimization problems
Van Veldhuizen DA, Lamont GB (2000) On measuring multiobjective evolutionary algorithm performance. In: Proceedings of the 2000 Congress on Evolutionary Computation. CEC00 (Cat. No. 00TH8512), vol 1, IEEE, pp 204–211
Zitzler E, Thiele L (1998) Multiobjective optimization using evolutionary algorithms?a comparative case study. In: International conference on parallel problem solving from nature, Springer, pp 292–301
Zitzler E, Thiele L (1999) Multiobjective evolutionary algorithms: a comparative case study and the strength pareto approach. IEEE Trans Evol Comput 3(4):257–271
Jiang S, Ong Y, Zhang J, Feng L (2014) Consistencies and contradictions of performance metrics in multiobjective optimization. IEEE Transactions on Cybernetics 44(12):2391–2404
Deb K, Thiele L, Laumanns M, Zitzler E (2006) Scalable test problems for evolutionary multi-objective optimization
Wang L, Pan X, Shen X, Zhao P, Qiu Q (2020) Balancing convergence and diversity in resource allocation strategy for decomposition-based multi-objective evolutionary algorithm. Appl Soft Comput 100:106968
Xu J, Deb K, Gaur A (2015) Identifying the pareto-optimal solutions for multi-point distance minimization problem in manhattan space. Comput. Optim. Innov.(COIN) Lab., East Lansing, MI, USA, COIN Tech. Rep. 2015018
Aggarwal C C, Hinneburg A, Keim D A (2001) On the surprising behavior of distance metrics in high dimensional space. In: International conference on database theory, Springer, pp 420– 434
Acknowledgements
This work was supported in part by Natural Science Foundation of Zhejiang Province (LQ20F020014), in part by the National Natural Science Foundation of China (61472366, 61379077), in part by the Natural Science Foundation of Zhejiang Province (LY17F020022), in part by Key Projects of Science and Technology Development Plan of Zhejiang Province (2018C01080).
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of Interests
The authors declare that they have no conflict of interest.
Additional information
Human participants
This study does not contain any studies with human participants or animals performed by any of the authors.
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhang, L., Wang, L., Pan, X. et al. A reference vector adaptive strategy for balancing diversity and convergence in many-objective evolutionary algorithms. Appl Intell 53, 7423–7438 (2023). https://doi.org/10.1007/s10489-022-03545-w
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-03545-w