1 Introduction

Meta-heuristic algorithms are the product of the improvement of the heuristic algorithm, which combines the random algorithm and the local search algorithm. They are random search methods characterized by imitating various operational mechanisms [1]. Since this kind of algorithm can reduce the amount of calculation, they are used in computer science, engineering optimization design and other fields [2,3,4,5].

According to the mechanism they are inspired by, meta-heuristic optimization algorithms fell generally into two groups: algorithms that imitate biological processes and algorithms based on physical principles. The first division includes evolutionary-based and swarm-based algorithms. Genetic Algorithm (GA) [6] is the earliest evolutionary-based algorithm. Particle Swarm Optimization (PSO) [7] opens up a precedent for swarm-based algorithms. Then many specialists have successively proposed lots of swarm intelligence optimization algorithms. The main inspiration for Moth-Flame Optimization (MFO) [8] comes from the navigation method of moths called ‘horizontal orientation’. In light of the fact that ants can always discover the best route when foraging, Ant Colony Optimization (ACO) [9] is established. In addition, there are also some well-known algorithms, such as Chimp Optimization Algorithm (ChOA) [10] and Artificial Bee Colony (ABC) [11].

As a novel kind of meta-heuristic algorithms, Whale Optimization Algorithm (WOA) [12] imitates the special bubble-net predation strategy of humpback whales. WOA has fewer parameters and operates more easily. It is extensively used to manage engineering optimization problems [13,14,15,16]. According to the experimental results of Ref. [12], WOA converges faster than DE and PSO, and has strong competitiveness and more significant application potential.

However, due to the characteristics of the algorithm itself, there are still some deficiencies in the optimization process. The main disadvantages of the WOA are that it is easy to fall into the local optimum and the convergence accuracy is not high. Because of the random initialization and parameter randomization, the exploration and exploitation are unbalanced, which converges to the local optimum easily [17]. Further, the convergent ability needs to be improved. The WOA algorithm gradually exposes some shortcomings in solving complex optimization problems. To overcome the shortcomings, some scholars have proposed WOA variants to enhance its performance.

According to the existing literature, the improvements are generally classified into three parts: the improvement achieved by adjusting the parameters, the improvement achieved by the adjustment of the search strategy and the improvement achieved by mixing WOA with other algorithms.

Many works have improved the algorithm by adjusting the parameters. Kaur and Arora [18] applied chaos theory in WOA and proposed CWOA, which used various chaotic maps to adjust the main parameters. Zhong and Long [19] have proposed a nonlinear strategy for different control parameters. Li et al. [20] used a nonlinear tuning parameter in an improved WOA (IWOA). Due to the few parameters of WOA, this kind of improvement is relatively simple, and the improvement effect is general.

The second is to enhance the performance by adjusting the search strategy, generally used in initializing the population and updating the location. Rashmi proposed [21] a modified WOA based on roulette wheel selection to balance exploration and exploitation. Elaziz and Oliva [22] employed an opposition-based learning strategy for a better-initialized population. Zhang and Wang [23] used the nonlinear adaptive weight to enable whales to search adaptively. Chen et al. [24] proposed the RDWOA, which used the random spare strategy to overcome the shortcoming of low convergence speed. To enhance the global exploration ability, Liu et al. [25] used the Lévy flights strategy to increase the search space. Chakraborty [26] proposed a novel improved WOA method (ImWOA) with increased diversity in the solution. The entire iteration is clearly divided into exploration and exploration. Mohammad [27] proposed an enhanced WOA with three new effective search strategies named migrating, preferential selecting, and enriched encircling prey.

Barring the above two types of improvements, some works are devoted to hybridizing the WOA algorithm with other intelligent algorithms to solve complex problems. Indrajit and Pradeep [28] raised a hybrid PSO-WOA algorithm, where WOA was used for exploration. Hardi and Tarik [29] proposed an improved algorithm hybridized with GWO to get better solutions. The hunting mechanism of GWO was applied in the exploitation stage of WOA. Kaveh and Rastegar [30] hybridized WOA with some concepts of Collider Bodies Optimization. Seyed and Samaneh [31] hybridized WOA with Differential Evolution (DE) to obtain improved WOA. This approach combined the exploration stage of DE to get a better solution. Bentouati et al. [32] combined the Pattern Search algorithm (PS) with WOA to solve power system problems. Tang [33] proposed a WOA mixed with the Artificial Bee Colony (ACWOA) to solve the problems of slow convergence and low precision. Dey [34] used a novel hybrid WOA-Sine cosine algorithm (SCA) to minimize the generation cost of a low voltage (LV) grid-connected microgrid system.

Although a series of works on the improvement of WOA has been carried out and the efficiency of WOA has been upgraded, there are some shortages. On the one hand, the strategies are relatively simple, using only nonlinear convergence parameters or adaptive weight. The single strategy cannot effectively achieve significant improvements in WOA performance. It is challenging to solve complex optimization problems. According to Ref. [23], the control parameter a was altered by different nonlinear adjustment strategies, and the improvement was not particularly effective. On the other hand, the computational complexity is higher than WOA, which will increase a lot of workloads. Chakraborty et al. [35] proposed an enhanced WOA, which introduced the mutualism phase of the Symbiotic Organisms Search. However, the complexity was higher than WOA. In addition, the improvement of WOA mainly based on the hybrid algorithm has resulted in losing evolutionary direction, which is challenging to realize and easy to sink into local optimum [36].

By consulting a lot of literature about the improvement of WOA, it is concluded that the effect of single strategy improvement is not very obvious. It is difficult to solve complex problems. In addition, most improvements only unilaterally improve the performance, such as increasing the convergence speed, improving the exploration capability and other single aspects. In contrast, the combination of multiple strategies can greatly boost the efficiency of WOA. Yuan et al. [37] proposed the Multi-Strategy Ensemble WOA, which employed the chaos strategy, an improved random search mechanism, Lévy flight, and an enhanced position correction mechanism. It balanced local and global searches to avoid premature algorithm convergence. Sun [38] devised a modified WOA based on multi-strategy to address somewhat deficiencies of the original WOA, which used a tent map function, inertia weight and an optimal feedback strategy. At present, many types of research are devoted to introducing different new strategies into WOA. The effective combination of multiple strategies may enhance the performance of WOA [39]. Elite reverse learning-related strategies, mutation strategies and Lévy flight-related strategies have been proved to be simple and effective in enhancing meta-heuristic optimization algorithms. Inspired by these references, we introduce these strategies into WOA.

Hence we propose a modified WOA (MWOA) with Multi-strategy mechanism. The above drawbacks can be solved by introducing new strategies to the WOA algorithm. For the problem of slow convergence and local optimum, MWOA improves the initialization, control parameters and search strategy of WOA, respectively. Improving at different stages of the algorithm and assisting each other can lead to better results. For the random initialization of WOA, the elite reverse learning strategy is introduced to get a better population. Then in terms of improvement of parameters, the nonlinear convergence factor is used for strengthening exploitation and exploration and maintains a balance of them. The search strategy is boosted by introducing the mutation strategy and a Lévy flight disturbance strategy. To enhance the exploration capacity and avoid local optimum, the DE/rand/1 mutation strategy is introduced into the exploration phase of WOA. Lastly, utilizing the Lévy fight disturbance strategy to boost the spatial search and accelerate convergence, which can strengthen the global search ability. Compared with WOA, PSO, MFO, SOA, SCA and other four WOA variants on the CEC 2017 benchmark suite, MWOA has strong competitiveness and can enhance the performance of WOA.

The remainder of this article is as below. WOA is briefly introduced in Sect. 2. MWOA is presented in Sect. 3, which introduces the mechanism of the four strategies and the pseudocode of MWOA. In Sect. 4, the experimental results are presented and analyzed. Lastly, we conclude in Sect. 5.

2 WOA

WOA simulates the unique hunting behavior of humpback whales. These whales usually live in groups in nature. When the prey is found, groups of humpback whales spiral upward from the bottom while exhaling air bubbles to surround their prey. These humpback whales usually hunt through three phases (1) Surround the prey; (2) Bubble-net attacking strategy; (3) Global search for prey. According to the three stages of whale foraging modeling, the WOA is obtained.

2.1 Surround the prey

Whales must first recognize the location of the prey. Because they cannot obtain specific location information of prey in advance, the location of the best whale is taken as the target location. Other individuals will move toward the best whale, and this behavior is expressed in Eqs. (12).

$$\vec{D} = \left| {\vec{C} \cdot \overrightarrow {{X^{*} }} \left( t \right) - \overrightarrow {X } \left( t \right)} \right|$$
(1)
$$\overrightarrow {X } \left( {t + 1} \right) = \overrightarrow {{X^{*} }} \left( t \right) - \overrightarrow {A } \cdot \vec{D}$$
(2)

where t is the index of iterations, \(\overrightarrow {{X{ }}} \left( t \right)\) is the position of the current population, \(\overrightarrow {{X^{*} }} \left( t \right)\) is the optimal position that has been searched, \(\overrightarrow {{X{ }}} \left( {t + 1} \right)\) deems the position that needs to be updated in the next iteration, \(\vec{A}\) indicates the convergence factor and \(\overrightarrow {{C{ }}} { }\) represents the disturbance factor, computed as follows:

$$\overrightarrow {A } = 2\overrightarrow {a } \cdot \overrightarrow {{r_{1} }} - \overrightarrow {a }$$
(3)
$$\overrightarrow {{C{ }}} = 2\overrightarrow {{r_{2} }}$$
(4)

where random vectors \(\overrightarrow {{r_{1} }}\) and \(\overrightarrow {{r_{2} }}\) are in [0,1], \(a{ }\) represents the convergence factor decreased from 2 to 0 during the iteration, whose changes can be expressed by:

$$a = 2 - \left( {2t/t_{{{\text{max}}}} } \right)$$
(5)

where \(t\) is the index of iterations.

2.2 Bubble-net attacking strategy

Whales will conduct spiral bubble net attacks on the target prey. There are two main ways: the shrinking encircling mechanism and the spiral updating position. The following equation can realize the spiral updating position.

$$\overrightarrow {{X{ }}} \left( {t + 1} \right) = \overrightarrow {{X^{*} }} \left( t \right) + \overrightarrow {{D_{p} }} \cdot e^{bl} \cdot \cos \left( {2\pi l} \right)$$
(6)

where \(\overrightarrow {{D_{p} }} = \left| {\overrightarrow {{X^{*} }} \left( t \right) - \overrightarrow {{X{ }}} \left( t \right)} \right|\) represents the distance between the individual and the target prey, \(b\) denotes a constant constrained to a logarithmic spiral shape, and the random number \(l\) is in [−1,1].

Whales randomly update their location. There is a 50 percent chance of choosing either contraction encirclement or spiral updating position, \(p\) is defined as a random number in [0, 1]. This behavior can be obtained by:

$$\overrightarrow {{X{ }}} \left( {t + 1} \right) = \left\{ {\begin{array}{*{20}l} {\overrightarrow {{X^{*} }} \left( t \right) - \overrightarrow {{A{ }}} \cdot \overrightarrow {{D{ }}} ,} \hfill & {p < 0.5} \hfill \\ {\overrightarrow {{X^{*} }} \left( t \right) + \overrightarrow {{D_{p} }} \cdot e^{bl} \cdot \cos \left( {2\pi l} \right) ,} \hfill & {p \ge 0.5} \hfill \\ \end{array} } \right.$$
(7)

2.3 Global search for prey

Assuming a random whale as the best whale. Other whales will approach this whale. Humpback whales search for prey by a random walk. When |A |> 1, it focuses on exploration and performs the global search. This behavior is obtained as follows:

$$\vec{D} = \left| {\overrightarrow {{C{ }}} \cdot \overrightarrow {{X_{rand} }} \left( t \right) - \overrightarrow {{X{ }}} \left( t \right)} \right|$$
(8)
$$\overrightarrow {X } \left( {t + 1} \right) = \overrightarrow {{X_{rand} }} \left( t \right) - \overrightarrow {A } \cdot \vec{D}$$
(9)

In the WOA algorithm, individuals update the positions based on the best whale or a random whale. It can realize the transition between exploitation and exploration by adjusting the parameter \(a\). Whales will search for prey randomly when \(|\overrightarrow {A|}\) > 1. When \(|\overrightarrow {A|}\) < 1, whales will attack their prey, and WOA is devoted to exploitation at this time.

The flowchart of WOA is presented in Fig. 1. Although WOA has some advantages compared with other algorithms, it also exposes some shortcomings. The main defect is the slow convergence rate [36]. The exploration and exploitation are unbalanced, which will lead WOA into local optimum in complex environments. Aiming at these problems, we propose MWOA to overcome deficiencies and enhance performance.

Fig. 1
figure 1

The schematic diagram of WOA

3 MWOA

MWOA adopts the elite reverse learning strategy, nonlinear convergence factor, DE/rand/1 mutation strategy and Lévy flight disturbance strategy. Elite reverse learning strategy is used at initialization to accelerate the convergence and obtain an excellent population. The nonlinear convergence factor can well balance exploitation and exploration. Then introducing the mutation strategy to enhance exploration capabilities. MWOA introduces a new hybrid operator \(m\) to control exploration and exploitation. Finally, applying the Lévy flight perturbation strategy to improve the global search capability and avoid local optimum. The details are as follows.

3.1 Elite reverse learning strategy

At initialization, the individuals generally update the location information randomly. The search time will be longer when it converges to the optimal solution. Therefore, initialization can significantly influence the search performance of WOA. Using the elite reverse learning strategy to initialize the population can obtain a better-initialized population and reduce the time for searching for the optimal solution [40].

Elite reverse learning strategy is used to generate reverse learning individual \(\overline{{X_{i} }}\), which is given by:

$$\overline{{X_{i} }} = k\left( {lb + ub} \right) - X_{i}$$
(10)

where \(lb{ }\) represents the upper boundary, \(ub\) represents the lower boundary, and \(X_{i}\) indicates the current individual location information of the initialized population. Reverse individuals are obtained through elite reverse learning. Calculating the fitness of \({ }\overline{{X_{i} }}\) and comparing it with the fitness of the \(X_{i}\). Better individuals are retained, which can be expressed by Eq. (11).

$$X_{i} = \left\{ {\begin{array}{*{20}c} {\overline{{X_{i} }} ,f\left( {\overline{{X_{i} }} } \right) < f\left( {X_{i} } \right) } \\ {X_{i} ,f\left( {\overline{{X_{i} }} } \right) > f\left( {X_{i} } \right)} \\ \end{array} } \right.$$
(11)

3.2 Nonlinear convergence factor

The coefficient \({ }a\) is the main factor affecting convergence performance. Because it decreases linearly, the exploration and exploitation of WOA cannot reach a balance [41]. According to Ref. [19], Zhong and Long proposed a nonlinear tuning strategy with different control parameters in the improved algorithm. The improvement using the cosine nonlinear convergence parameter was proved to be the best in several experiments. This paper uses this strategy and the nonlinear convergence factor \(a_{1}\) calculated by Eq. (12).

$$a_{1} = (a_{{\text{max }}} - a_{{\text{min }}} ) \cdot \cos \left( {7 \cdot \frac{t}{{t_{{{\text{max}}}} }}\pi } \right)$$
(12)

where the \(a_{{{\text{max}}}} ,\;a_{{\text{min }}}\) are the maximal and minimal of the parameter \(a\).

3.3 Mutation strategy

The DE algorithm [42] has only been developed for more than 20 years since it was proposed. However, it has shown structural advantages such as firm performance, simple structure and parallel operation. The DE algorithm has a good exploration ability for the function optimization problem [43]. By introducing the mutation strategy of DE to other meta-heuristic optimization algorithms, better solutions can be effectively obtained, and the exploration ability can be enhanced [44]. Hu et al. [45] used the mutation operator of differential evolution at mutation phase of BWO which helps the algorithm escape from the local optima. In this section, we give a brief introduction to the DE algorithm. This is mainly to reference the mutation policy in the DE algorithm. This strategy prevents stagnation in the search process, and it can deviate from the local optimum. WOA is prone to fall into local optimum and is particularly prone to premature convergence.

MWOA introduces the DE/rand/1/ mutation strategy into the exploration stage of WOA, which can boost the exploration capability and avoid local optimum. The mutation formula is as follows:

$$V_{i} \left( t \right) = X_{{w_{1} }} \left( t \right) + F \cdot \left( {X_{{w_{2} }} \left( t \right) - X_{{w_{3} }} \left( t \right)} \right)$$
(13)

The mutation vector \(V_{i} \left( t \right)\) is obtained through the mutation operation, \(w_{1}\), \(w_{2}\) and \(w_{3}\) are random integers generated in [1, N], they are not equal to each other. F denotes the mutation operator, which controls the magnification of the deviation vector in practice. N is the population size.

Then, to increase the vector's diversity, a crossover operation is introduced on the population. It can be presented by:

$$U_{i} \left( t \right) = \left\{ {\begin{array}{*{20}l} {V_{i} \left( t \right),}& {rand\left( {0,1} \right) \le\, CR \, or \, j = j_{rand} }\\ {X_{i} \left( t \right),}& {others }\\ \end{array} } \right.$$
(14)

where \({ }U_{i} \left( t \right)\) is the test vector obtained by the crossover operation, and CR is the crossover operator.

Finally, the selection operation is performed. The generated test vector \(U_{i} \left( t \right)\) is compared with the original individual \(X_{i} \left( t \right)\). Better individuals are retained. The selection process can be represented by:

$${ }X_{i} \left( {t + 1} \right) = \left\{ {\begin{array}{*{20}l} {U_{i} \left( t \right),} \hfill & {fU_{i} \left( t \right) < fX_{i} \left( t \right)} \hfill \\ {X_{i} \left( t \right),} \hfill & { {\text{otherwise}}} \hfill \\ \end{array} } \right.$$
(15)

In MWOA, we introduce a new hybrid operator \(m\) to adjust the exploration and exploitation. When \(rand \le m\), it performs the exploration part. When and \(> m\), it performs the exploitation part. The operator \(m\) is calculated by Eq. (16).

$$m = 1 - \frac{t}{{{ }t_{{{\text{max}}}} { }}}$$
(16)

3.4 Lévy flight disturbance strategy

Since Yang [46] proposed the cuckoo search algorithm which used two mechanisms of cuckoo parasitic reproduction and Lévy flight. The method has been widely used to improve the meta-heuristic algorithms [47,48,49]. In recent years, relevant scholars have conducted a series of research and proved that it helps to enhance the exploration of the optimization algorithm [50]. In WOA, individuals gradually approach to the optimal whale. The whale population is relatively concentrated, which can weaken global search capabilities and trap in the local optimum.

The Lévy flight disturbance strategy introduced in this paper perturbs the population after each position updates. It can improve the phenomenon of local optimization and premature convergence. The new equation for updating the position can be expressed by:

$$U_{i} \left( t \right) = U_{i} \left( t \right) + \alpha \cdot Le{\acute{v}}y \left( \lambda \right)$$
(17)

where \({ }\alpha\) denotes the step size control factor, \(U_{i} \left( t \right)\) is the position of the current population, and Lévy (λ) is the path of random search, calculated by Eq. (18)

$$Le{\acute{v}}y\left( \lambda \right) = \frac{u}{{\left| v \right|^{1/\beta } }}$$
(18)

where \(u\) and \(v{ }\) follow normal distribution, \(u \sim N\left( {0,\delta_{u}^{2} } \right)\), \(v \sim N\left( {0,\delta_{v}^{2} } \right),\) \(\delta_{v} = 1\) and \(\delta_{u}\) can be calculated by Eq. (19).

$$\delta_{u} = \left\{ {\frac{{\Gamma \left( {1 + \beta } \right)\sin \left( {\pi \beta /2} \right)}}{{\Gamma \left[ {\left( {1 + \beta } \right)/2} \right]\beta 2^{{\left( {\beta - 1} \right)/2}} }}} \right\}^{1/\beta }$$
(19)

where \(\Gamma\) is the standard gamma function. \(\beta\) usually takes the value of 3/2.

The pseudocode of MWOA is shown below.

figure a

4 Experimental results

4.1 CEC2017 benchmark suite

CEC 2017 benchmark suite [51] is used to test the performance of MWOA. It consists of 30 functions, which are separated into four groups: Unimodal (F1–F3); Simple Multimodal (F4–F10); Hybrid (F11–F20) and Composition (F21–F30). Different test functions are used to test various properties. Since F2 is unstable during the experiment using MATLAB software, this experiment used 29 test functions other than F2.

4.2 Experimental settings

MWOA is compared with five classical algorithms and four WOA variants. In this paper, the total number of evaluations is 300,000. The population size is 30 and D is taken as 30. \(F\) and \(CR\) are 0.5 and 0.9, respectively, and \(\alpha\) is 0.01. During the experiments, each algorithm runs 51 times to avoid the occasional error. The optimal value is preserved in each iteration. We use the mean error as the statistical index. We calculate the mean (Mean) and standard (Std) deviation of the data to reduce the statistical error. The same hardware and software platform is used for this experiment, and the running environment is Windows 10 operating system with Intel(R) Core(TM) i7-8565U CPU. MWOA and other algorithms are realized in MATLAB 2020(a).

4.3 Comparison of MWOA with other classic algorithms

We compare MWOA with WOA and four classic algorithms. Four algorithms include: PSO [7], MFO [8], SOA [52], SCA [53]. The 29 test functions are independently tested 51 times. The results from six algorithms are shown in Table 1. Bold data is the optimal value among six algorithms.

Table 1 Comparison of MWOA with WOA, PSO, MFO, SOA, and SCA on CEC 2017 at D = 30

In terms of convergence accuracy, MWOA obtains the best results on 27 functions. WOA obtains optimal solutions on 2 functions. PSO, MFO, SOA and SCA have not achieved the best results on any function. MWOA is significantly better than PSO, MFO, SOA and SCA on all test functions, and it is the second best on F25 and F26 among all algorithms. The standard deviation of MWOA is the smallest on 22 functions, so the stability of MWOA is superior to other algorithms.

We further study the convergence capability of MWOA. The mean error of 51 independent runs is calculated. Figure 2 is the convergence curve, showing the iteration trend on some selected functions. The ordinate represents the logarithm of the average function error value, and the abscissa represents the iteration number.

Fig. 2
figure 2

Convergence curves of WOA, PSO, MFO, SOA, SCA and MWOA on 29 test functions

As seen in Fig. 2, the searchability of MWOA is better. MWOA has a good convergence speed and better convergence accuracy than WOA, PSO, MFO, SOA and SCA, showing a strong convergence ability in early iterations. The convergence accuracy of MWOA is superior to other algorithms on most functions. On F1, F5-F9, F12, F19, F20 and F29, the MWOA has the fastest convergence speed. Overall, MWOA can always get better solutions when handling 30-dimensional problems.

According to the figure, the strategies applied by MWOA have different effects on the algorithm. MWOA shows a faster convergence in early iterations, and it can save time for convergence to the optimal solution. This is the result of using elite reverse learning strategy. MWOA remains stable in mid-iteration. Exploration and exploitation are balanced because of the improvement of nonlinear parameters. The application of the mutation strategy and Lévy flight disturbance strategy makes the algorithm continue to search globally in the later stage to avoid local optimum. Using multiple strategies enables MWOA to obtain better convergence performance than other algorithms.

4.4 Comparison of MWOA with other WOA variants

In this part, We compare MWOA with other WOA variants, which are MSWOA [54], IWOA [55], CWOA [18] and OBWOA [22]. These four algorithms are representative improvement algorithms, which can improve the original WOA from different perspectives and use different mechanisms, and the improvement is relatively good. So they have better comparison value. MSWOA is proposed by Yang, using chaotic initialization, adaptive weights and EPD strategy. IWOA is proposed by Li and Luo with a nonlinear adjustment strategy for cosine control parameters and Gaussian perturbation operator. CWOA introduces chaotic logical sequences during population initialization, which is proposed by Kaur and Arora. OBWOA employs an Opposition-based Learning strategy. For fairness, MWOA and other algorithms run independently 51 times on the CEC2017 benchmark suite. The mean error is calculated as an evaluation index. Experimental results of MWOA, MSWOA, IWOA, CWOA and OBWOA are demonstrated in Table 2. Bold data reveals the optimal value between MWOA and other algorithms.

Table 2 Comparison of MWOA with MSWOA, IWOA, CWOA and OBWOA on CEC2017 at D = 30

According to Table 2, MWOA obtains the best solutions on 24 functions. CWOA acquires the optimal solutions on four functions which are F21, F24, F25 and F26. MSWOA gets the optimal solutions on F27. IWOA and OBWOA do not obtain the optimal solutions on any functions. Moreover, the results of MWOA on F21, F24, F25, F26 and F27 are the second-best among all algorithms. MWOA can generally achieve better results in dealing with various global problems. In terms of stability, MWOA has the slightest standard deviation on 21 functions, and the performance is relatively stable and superior to other WOA variants. The embedding of the mutation strategy strengthens the exploration ability and greatly improves the convergence accuracy. In general, the strategies used by MWOA greatly improve the convergence accuracy and stability, which is better than other WOA variants to a certain extent.

We make the convergence curves to study the performance of MWOA. Figure 3 is the convergence curves of MWOA, MSWOA, IWOA, CWOA and OBWOA over 51 independent runs on the selected functions.

Fig. 3
figure 3

Convergence curves of MSWOA, IWOA, CWOA, OBWOA and MWOA on 29 test functions

We can conclude that MWOA performs best than other four WOA variants. MWOA has apparent advantages in the convergence speed on F1, F5–F12, F16–F20, F23 and F30. MWOA exhibits a good convergence speed at the beginning of the iteration. Then it keeps exploration and exploitation stable. There is a tendency to continue to accelerate in the later iterations, especially on F9, F14 and F18. It shows that the algorithm continues to perform a global search at the later iterative stage, which helps avoid local optimum. Lévy flight is used for updating the position, which increases the search step and accelerates the algorithm convergence. It can strengthen the global search ability. Moreover, it enables the algorithm to find the best solution rapidly and effectively. In general, MWOA is better than MSWOA, IWOA, CWOA and OBWOA in terms of the improvement effect. MWOA can consistently achieve better results when solving global optimization problems.

In addition to the above experiments, MATLAB and EXCEL software are used to perform Wilcoxon’s signed rank test on MWOA and other nine algorithms. The experimental results are listed in Table 3. \(R^{ + }\) represents positive rank sum, \(R^{ - }\) represents negative rank sum. The P-value is a parameter used to determine the results of the hypothesis test and indicates the significant difference between MWOA and other algorithms. The \(R^{ + }\) value of MWOA is larger than \(R^{ - }\) value in all comparisons. The value of the \(R^{ - }\) is 0 in contrast with IWOA, PSO, MFO, SOA, and SCA. At the confidence level of \(\alpha\) = 0.05, the P-value is much less than 0.05. So MWOA significantly differs from other algorithms. Through the experimental results, MWOA is superior to other algorithms when handling 30-dimensional problems.

Table 3 The statistical results based on Wilcoxon’s signed rank test (D = 30)

5 Conclusion

In WOA, the individual generally initializes the position randomly. Since the optimal solution cannot be obtained easily, random initialization will increase the search range of feasible solutions and search time. The linear convergence factor makes the WOA encounter the phenomenon of unbalanced exploration and exploitation in optimization. It tends to converge prematurely and fall into local optimum. We propose a modified WOA with multi-strategy mechanism to solve the existing problems of WOA and improve its performance. The innovations of MWOA are as follows:

  1. 1.

    MWOA introduces the elite reverse learning strategy to obtain a better population, which can enhance population diversity and speed up convergence.

  2. 2.

    The nonlinear convergence factor is introduced to maintain the balance between exploration and exploitation.

  3. 3.

    Using mutation strategy can strengthen the global exploration capability. Furthermore, the ability to mine the optimal solution is effectively improved, so it can prevent the algorithm from falling into local optimum.

  4. 4.

    The application of the Lévy flight disturbance strategy can increase the search range of the algorithm space. It is beneficial for finding the global optimal solution. After updating the location, the Lévy fight strategy perturbs the population position to disperse the population. This operation can avoid premature convergence.

The CEC2017 benchmark suite is used to test the performance of MWOA. MWOA is compared with WOA, PSO, MFO, SOA, SCA, and other four WOA variants. It has been proved by many experiments that MWOA has strong competitiveness. The proposed algorithm has noticeable improvements in convergence speed and accuracy. It can well balance exploration and exploitation to avoid local optimum. In general, MWOA demonstrates outstanding performance and has great competitiveness in dealing with optimization problems.

Although this paper overcomes some shortcomings of WOA and enhances the performance of WOA to some extent, the global theoretical optimum is not found for some functions. In addition, there are some parameters in MWOA, which need to be adjusted. The algorithm proposed in this paper needs to be applied to solve practical problems. In the next work, we will apply WOA to solveengineering optimization problems.