1 Introduction

Nature has been the main inspiration for the majority of the population-based stochastic optimization techniques. As the name of such techniques implies, they perform optimization randomly. The optimization process is usually started by creating a set of random solutions. These initial solutions are then combined, moved, or evolved over a pre-defined number of steps called iterations or generations. This is almost the main framework of all population-based algorithms. What makes an algorithm different from others in this field is the mechanism of combining, moving, or evolving the solutions during optimization.

For instance, Genetic Algorithms (GAs) [1] utilize the survival of the fitter individuals in nature in order to select the best solutions and then combine them based on the reproduction of chromosomes. Particle Swarm Optimization (PSO) [2] was inspired by social and individual thinking of birds when flying, so it obliges the candidate solutions to move around a search space with respect to their own personal best position obtained so far as well as the best position that the swarm found so far. Gravitational Search Algorithm (GSA) [3] uses the Newtonian laws of motion in order to move its search agents towards the promising regions of a search space.

Another common concepts among different population-based algorithms are exploration and exploitation. The former refers to the phase that an algorithm tries to discover different promising regions of a search space globally. Generally speaking, abrupt changes in candidate solutions are fruitful at this stage. In contrary, the latter concept is the convergence ability an algorithm around the obtained promising solutions in the exploration phase. A proper balance between exploration and exploitation can guarantee proceeding towards the global optimum.

Recently, there has been a growing interest in proposing new algorithms or improving the current ones in this field. A significant number of practical applications also accompanies the theoretical works. The reason of this remarkable popularity might be originated from the so-called No Free Lunch (NFL) theorem for optimization [4]. This theorem has been proved logically that there is no optimization technique for solving all optimization problems. The NFL theorem, obviously, makes this area of research open, in which researchers are allowed to improve/adapt the current algorithms for solving different problems or propose new algorithms for providing competitive results compared to the current algorithms.

In this work, a novel stochastic population-based algorithm is proposed called Multi-Verse Optimizer (MVO). As its name implies, MVO is inspired by the theory of multi-verse in physics. Three main concepts of the multi-verse theory (white hole, black hole, and wormhole) are mathematically modelled to construct the MVO. The rest of the paper is organized as follows.

Section 2 provides the literature review of the stochastic optimization techniques. Section 3 discusses the concepts of multi-verse theory and proposes the MVO algorithm. The test beds and results are demonstrated in Sect. 4. The real engineering problems are solved and discussed at the end of Sect. 4 as well. Eventually, Sect. 5 concludes the work and suggests some directions for future studies.

2 Related works

Generally speaking, stochastic optimization techniques can be divided into two main categories: single-solution-based versus population-based. The former class of algorithms starts the optimization process with a single random solution and improves it over a pre-defined number of generations. Simulate annealing (SA) [5], local searches [6, 7], and hill climbing [8] belong to this class of algorithms. The advantages of single-solution-based algorithms are: simplicity and low number of function evaluation. However, the main disadvantage is the high probability of entrapment in local optima. In addition, since at every run a single solution is involved, there is no information sharing, and the algorithm should deal with lots of issues such as local optima, isolation of optima, deceptiveness, bias of the search space, and premature convergence with only one candidate solution.

In contrast to single-solution-based algorithms, population-based algorithms initiate the optimization process with a set of random solutions and improve them over the course of iterations. This set of solutions is sometimes called candidate solutions’ set. PSO, GA, Ant Colony Optimization (ACO) [9, 10], Artificial Bee Colonies (ABC) [11], and GSA [11] are some of the most popular algorithms in this class. The main advantage of the population-based algorithms is that there can be information exchange between the candidate solutions. Therefore, they can handle the issues such as local optima, isolation of optima, deceptiveness, bias of the search space, and premature convergence easier and faster. Another advantage is the less probability of entrapment in local solutions compared to the single-solution-based algorithms. The disadvantages of these algorithms are: less simplicity and the need for high number of function evaluation at each iteration.

The literature shows that the population-based algorithms have become a reliable alternative to single-solution-based algorithms due to the above-mentioned advantages. The application of these methods can also be found in a wide range of fields, emphasizing the merits of these techniques. Generally speaking, the design process of an algorithm starts with an inspiration. The inspiration could be from behaviour of creatures, natural phenomena, or social events. After the inspiration, different potential mathematical models are generated to design the algorithm. The best combination of mathematical models is then found by conducting experiments on various test beds.

The operators of algorithms in this field are usually designed to accomplish two phases: exploration versus exploitation. In the former phase, an algorithm should be equipped with mechanisms to search the search space as extensively as possible. In fact, promising regions of the search space are identified in this phase. In the exploitation phase, however, there should be emphasizes on local search and convergence towards promising areas obtained in the exploration phase. Exploration and exploitation are two conflicting stages with no specific mathematical definition. The exploration phase usually comes before exploitation, but there is a need to re-explore the search space in case of local optima stagnation, which is quite common in real problems with unknown search spaces.

Another challenge when designing an algorithm is the transition between exploration and exploitation. There is no clear rule for an algorithm to realize the most suitable time for transiting from exploration to exploration due to both unknown shape of search spaces and stochastic nature of population-based algorithms. The majority of population-based algorithms have been tuned adaptively to smoothly transit between exploration and exploitation. For instance, the inertia weight in PSO is mostly decreased linearly from 0.9 to 0.4 in order to reduce the impacts of velocity vectors on particle movements and emphasize exploitation as iterations increase.

The above paragraphs show the challenges that a designer encounters when developing a new meta-heuristic. The following section proposes a novel meta-heuristic based on the concepts of multi-verse theory.

3 Multi-Verse Optimizer

3.1 Inspiration

The big bang theory [12] discusses that our universe starts with a massive explosion. According to this theory, the big bang is the origin of everything in this world, and there was nothing before that. Multi-verse theory is another recent and well-known theory between physicists [13]. It is believed in this theory that there are more than one big bang and each big bang causes the birth of a universe. The term multi-verse stands opposite of universe, which refers to the existence of other universes in addition to the universe that we all are living in [13]. Multiple universes interact and might even collide with each other in the multi-verse theory. The multi-verse theory also suggests that there might be different physical laws in each of the universes.

We chose three main concepts of the multi-verse theory as the inspiration for the MVO algorithm: white holes, black holes, and wormholes. A white hole has never seen in our universe, but physicists think that the big bang can be considered as a white hole and may be the main component for the birth of a universe [14]. It is also argued in the cyclic model of multi-verse theory [15] that big bangs/white holes are created where the collisions between parallel universes occur. Black holes, which have been observed frequently, behave completely in contrast to white wholes. They attract everything including light beams with their extremely high gravitational force [16]. Wormholes are those holes that connect different parts of a universe together. The wormholes in the multi-verse theory act as time/space travel tunnels where objects are able to travel instantly between any corners of a universe (or even from one universe to another) [17]. Conceptual models of these three key components of the multi-verse theory are illustrated in Fig. 1.

Fig. 1
figure 1

White hole, black hole, and wormhole

Every universe has an inflation rate (eternal inflation) that causes its expansion through space [18]. Inflation speed of a universe is very important in terms of forming stars, planets, asteroids, black holes, white holes, wormholes, physical laws, and suitability for life. It is argued in one of the cyclic multi-verse models [19] that multiple universes interact via white, black, and wormholes to reach a stable situation. This is the exact inspiration of the MVO algorithm, which is conceptually and mathematically modelled in the following subsection.

3.2 MVO algorithm

As discussed in the preceding section, a population-based algorithm divides the search process into two phases: exploration versus exploitation. We utilize the concepts of white hole and black hole in order to explore search spaces by MVO. In contrast, the wormholes assist MVO in exploiting the search spaces. We assume that each solution is analogous to a universe and each variable in the solution is an object in that universe. In addition, we assign each solution an inflation rate, which is proportional to the corresponding fitness function value of the solution. We also use the term time instead of the iteration in this paper since it is a common term in multi-verse theory and cosmology.

During optimization, the following rules are applied to the universes of MVO:

  1. 1.

    The higher inflation rate, the higher probability of having white hole.

  2. 2.

    The higher inflation rate, the lower probability of having black holes.

  3. 3.

    Universes with higher inflation rate tend to send objects through white holes.

  4. 4.

    Universes with lower inflation rate tend to receive more objects through black holes.

  5. 5.

    The objects in all universes may face random movement towards the best universe via wormholes regardless of the inflation rate.

The conceptual model of the proposed algorithm is illustrated in Fig. 2.

Fig. 2
figure 2

Conceptual model of the proposed MVO algorithm (I(U 1) > I(U 2) > ··· > I(U n−1) > I(U n ))

This figure shows that the objects are allowed to move between different universes through white/black hole tunnels. When a white/black tunnel is established between two universes, the universe with higher inflation rate is considered to have white hole, whereas the universe with less inflation rate is assumed to own black holes. The objects are then transferred from the white holes of the source universe to black holes of the destination universe. This mechanism allows the universes to easily exchange objects. In order to improve the whole inflation rate of the universes, we assume that the universes with high inflation rates are highly probable to have white holes. In contrary, the universes with low inflation rates have a high probability of having black holes. Therefore, there is always high possibility to move objects from a universe with high inflation rate to a universe with low inflation rate. This can guarantee the improvement of the average inflation rates of the whole universes over the iterations.

In order to mathematically model the white/black hole tunnels and exchange the objects of universes, we utilized a roulette wheel mechanism. At every iteration, we sort the universes based of their inflation rates and chose one of them by the roulette wheel to have a white hole. The following steps are done in order to do this.

Assume that

$$U = \left[ {\begin{array}{*{20}c} {x_{1}^{1} } & {x_{1}^{2} } & \ldots & {x_{1}^{d} } \\ {x_{2}^{1} } & {x_{2}^{2} } & \ldots & {x_{2}^{d} } \\ \vdots & \vdots & \vdots & \vdots \\ {x_{n}^{1} } & {x_{n}^{2} } & \ldots & {x_{n}^{d} } \\ \end{array} } \right]$$

where d is the number of parameters (variables) and n is the number of universes (candidate solutions):

$$x_{i}^{j} = \left\{ {\begin{array}{*{20}c} {\begin{array}{*{20}c} {x_{k}^{j} } & {r1 < NI(Ui)} \\ \end{array} } \\ {\begin{array}{*{20}c} {x_{i}^{j} } & {r1 \ge NI(Ui)} \\ \end{array} } \\ \end{array} } \right.$$
(3.1)

where \(x_{i}^{j}\) indicates the jth parameter of ith universe, Ui shows the ith universe, NI(Ui) is normalized inflation rate of the ith universe, r1 is a random number in [0, 1], and \(x_{k}^{j}\) indicates the jth parameter of kth universe selected by a roulette wheel selection mechanism.

The pseudocodes for this part are as follows:

figure a

As may be seen in these pseudocodes and Eq. (3.1), the selection and determination of white holes are done by the roulette wheel, which is based on the normalized inflation rate. The less inflation rate, the higher probability of sending objects though white/black hole tunnels. Please note that −NI should be changed to NI for the maximization problems. The exploration can be guaranteed using this mechanism since the universes are required to exchange objects and face abrupt changes in order to explore the search space.

With the above mechanism, the universes keep exchanging objects without perturbations. In order to maintain the diversity of universes and perform exploitation, we consider that each universe has wormholes to transport its objects through space randomly. In Fig. 2, white points represent transferred objects through the wormholes. It may be observed that the wormholes randomly change the objects of the universes without consideration of their inflation rates. In order to provide local changes for each universe and have high probability of improving the inflation rate using wormholes, we assume that wormhole tunnels are always established between a universe and the best universe formed so far. The formulation of this mechanism is as follows:

$$x_{i}^{j} = \left\{ {\begin{array}{*{20}l} {\left\{ {\begin{array}{*{20}l} {X_{j} + {\text{TDR}} \times \left( {\left( {ub_{j} - lb_{j} } \right) \times r4 + lb_{j} } \right)} \hfill & {r3 < 0.5} \hfill \\ {X_{j} - {\text{TDR}} \times \left( {\left( {ub_{j} - lb_{j} } \right) \times r4 + lb_{j} } \right)} \hfill & {r3 \ge 0.5} \hfill \\ \end{array} } \right.} \hfill & { r2 < {\text{WEP}}} \hfill \\ {x_{i}^{j} } \hfill & {r2 \ge {\text{WEP}}} \hfill \\ \end{array} } \right.$$
(3.2)

where \(X_{j}\) indicates the jth parameter of best universe formed so far, TDR is a coefficient, WEP is another coefficient, \(lb_{j}\) shows the lower bound of jth variable, \(ub_{j}\) is the upper bound of jth variable, \(x_{i}^{j}\) indicates the jth parameter of ith universe, and r2, r3, r4 are random numbers in [0, 1].

The pseudocodes are as follows (assuming that ub and lb indicate upper bound and lower bound of the variables):

figure b

It may be inferred from the pseudocodes and mathematical formulation that there are two main coefficients herein: wormhole existence probability (WEP) and travelling distance rate (TDR). The former coefficient is for defining the probability of wormhole’s existence in universes. It is required to increase linearly over the iterations in order to emphasize exploitation as the progress of optimization process. Travelling distance rate is also a factor to define the distance rate (variation) that an object can be teleported by a wormhole around the best universe obtained so far. In contrast to WEP, TDR is increased over the iterations to have more precise exploitation/local search around the best obtained universe. Wormhole existence and travelling distance rates are illustrated in Fig. 3. The adaptive formula for both coefficients are as follows:

$${\text{WEP}} = { \hbox{min} } + l \times \left( {\frac{{{ \hbox{max} } - { \hbox{min} }}}{L}} \right)$$
(3.3)

where min is the minimum (0.2 in this paper), max is the maximum (1 in this paper), l indicates the current iteration, and L shows the maximum iterations.

$${\text{TDR}} = 1 - \frac{{l^{1/p} }}{{L^{1/p} }}$$
(3.4)

where p (in this paper equals 6) defines the exploitation accuracy over the iterations. The higher p, the sooner and more accurate exploitation/local search.

Fig. 3
figure 3

Wormhole existence probability (WEP) versus travelling distance rate (TDR)

Note that WEP and TDR can be considered as constants as well, but we recommend adaptive values according to the results of this paper.

After all, the general steps and pseudocodes of the MVO algorithm are provided in the “Appendix 1”.

In the MVO algorithm, the optimization process starts with creating a set of random universes. At each iteration, objects in the universes with high inflation rates tend to move to the universes with low inflation rates via white/black holes. Meanwhile, every single universe faces random teleportations in its objects through wormholes towards the best universe. This process is iterated until the satisfaction of an end criterion (a pre-defined maximum number of iterations, for instance).

The computational complexity of the proposed algorithms depends on number of iterations, number of universes, roulette wheel mechanism, and universe sorting mechanism. Sorting universe is done in every iteration, and we employ the Quicksort algorithm, which has the complexity of O(n log n) and O(n 2) in the best and worst case, respectively. The roulette wheel selection is run for every variable in every universe over the iterations and is of O(n) or O(log n) based on the implementation. Therefore, the overall computational complexity is as follows:

$$O\left( {\text{MVO}} \right) = O\left( { l\left( {O\left( {\text{Quick sort}} \right) + n \times d \times \left( {O\left( {\text{roulette wheel}} \right)} \right)} \right)} \right)$$
(3.5)
$$O\left( {\text{MVO}} \right) = O\left( {l\left( {n^{2} + n \times d \times { \log }\,n} \right)} \right)$$
(3.6)

where n is the number of universes, l is the maximum number of iterations, and d is the number of objects.

To see how the proposed algorithm theoretically has the potential to solve optimization problems, some observations are as follows:

  • White holes are more possible to be created in the universes with high inflation rates, so they can send objects to other universes and assist them to improve their inflation rates.

  • Black holes are more likely to be appeared in the universes with low inflation rates so they have higher probability to receive objects from other universes. This again increases the chance of improving inflation rates for the universes with low inflation rates.

  • White/black hole tunnels tend to transport objects from universes with high inflation rates to those with low inflation rates, so the overall/average inflation rate of all universes is improved over the course of iterations.

  • Wormholes tend to appear randomly in any universe regardless of inflation rate, so the diversity of universes can be maintained over the course of iterations.

  • While/black hole tunnels require universes to abruptly change, so this can guarantee exploration of the search space.

  • Abrupt changes also assist resolving local optima stagnations.

  • Wormholes randomly re-span some of the variables of universes around the best solution obtained so far over the course of iterations, so this can guarantee exploitation around the most promising region of the search space.

  • Adaptive WEP values smoothly increase the existence probability of wormholes in universes. Therefore, exploitation is emphasized during optimization process.

  • Adaptive TDR values decrease the travelling distance of variables around the best universe, a mechanism for increasing the accuracy of local search over the iterations.

  • The convergence of the proposed algorithm is guaranteed by emphasizing exploitation/local search proportional to the number of iterations.

The following section investigates these theoretical claims in practice.

4 Results and discussion

We selected 19 benchmark functions in the literature as test beds for comparison [2023]. To improve the difficultly of the test functions F1–F13, we randomly shift the optima of test functions at every run. The rest of the test functions, F14–F19, are six composite benchmark functions provided by CEC 2005 special session [24, 25]. It is worth mentioning here that the dimension of the test functions is also considered 50. The test functions are listed in Tables 1, 2, and 3. The shape of the composite functions is also illustrated in Fig. 4.

Table 1 Unimodal benchmark functions
Table 2 Multi-modal benchmark functions
Table 3 Composite benchmark functions
Fig. 4
figure 4

Search space of composite benchmark functions

The employed test functions can be divided into three groups: unimodal, multi-modal, and composite. Unimodal test functions have one global optimum, so they are suitable for benchmarking the exploitation of algorithms. The multi-modal test functions, however, have a global optimum as well as multiple local optima with number exponentially increasing with the dimension. Eventually, composite test functions are considered as the combination of the first two groups with random rotation, shift, and biasing. The composite test functions are more similar to the real search spaces and good for benchmarking the exploration and exploitation of algorithms simultaneously.

In order to have a fair comparison, the algorithm is run 30 times. We also conduct Wilcoxon’s nonparametric statistical test over this 30 runs in order to draw a statistically meaningful conclusion. Such statistical test must be done due to the stochastic nature of meta-heuristics [26, 27]. Any p values <0.05 evidence the statistical significant superiority of the results. After all, the statistical results are provided in Tables 4, 5, and 6. Note that the number of universes is 30 and the maximum number of iterations is equal to 500. WEP is increased linearly from 0.2 to 1, and TDR is decreased from 0.6 to 0 using Eqs. (3.3) and (3.4), respectively.

Table 4 Results of unimodal benchmark functions
Table 5 Results of multi-modal benchmark functions
Table 6 Results of composite benchmark functions

For the verification of the results, the MVO algorithm is compared to PSO [2] as the best algorithm among SI-based techniques, GA [1] as the best evolutionary algorithms, GSA [3] as one of the best and recent physics-based algorithms, and Grey Wolf Optimizer (GWO) [28] as one of the most recent algorithms.

4.1 Unimodal test functions and exploitation

The results of Table 4 show that the proposed algorithm is able to provide very competitive results on the unimodal test functions. The p values in Table 7 also prove that the superiority is significant in the majority of the cases. This testifies that the proposed algorithm has a high exploitation ability, which is due to the integrated adaptive WEP/TDR constants and wormholes combined that assist MVO to provide high exploitation.

Table 7 p values of the Wilcoxon rank-sum test over 10 runs (p ≥ 0.05 have been underlined)

4.2 Multi-modal test functions and exploration analysis

After proving the exploitation of MVO, we are going to discuss the exploration of this algorithm. The results of multi-modal test function can be seen in Table 5. The results of this table evidence that the proposed algorithm is again able to provide very promising performance. It may be observed that this algorithm outperforms others on F8, F11, F12, and F13. The p values of Table 7 suggest that the superiority of the MVO algorithm is statistically significant in the majority of times. In addition, p values show that the MVO algorithm rejects the null hypothesis on F9 and F10, proving that the results of this algorithm are very competitive.

Superior exploration of the proposed algorithm is due to the white and black holes that allow universes to exchange different objects. Travelling an object from a universe to the other causes sudden changes and promotes exploration. This is similar to the crossover operator in GA that highly emphasizes exploration of the search spaces.

Exploration always comes with local optima avoidance. In fact, stagnation in local solutions can be resolved by promoting exploration. In addition to the discussion provided in the preceding paragraph, local optima avoidance of MVO also can be reasoned by the employed black/white hole concepts. Abrupt changes of universe are very useful in case of local optima stagnation.

4.3 Composite test functions and exploration/exploitation analysis

This subsection discusses the balance between exploration and exploitation, which is an essential feature of stochastic algorithms for solving real problems. The composite test functions are suitable for examining the balance of exploration and exploitation in this regard. As presented in Table 6, the results of the algorithms on composite test functions are very close. This is due to the difficulty of this set of test functions. The results prove that the MVO algorithm provides highly competitive results on the composite test functions as well. This evidences that the MVO algorithm properly balances exploration and exploration during optimization. This is firstly originated from the employed adaptive WEP that assists to smoothly transit between exploration and exploitation phases.

4.4 Convergence analysis

To confirm the convergence of the proposed algorithm, we provide the search histories, trajectories in the first dimension, average fitness of all universes, and convergence curves in Fig. 7 similarly to [28, 29]. From search history point of view, second columns of Fig. 5 show that the universes tend to scatter around the best solutions in the search space over the course of iterations. In order to see the changes in the search agents’ variables, the trajectories of the first universe in the first dimension are depicted in the fourth column of Fig. 5 as well. It is evident from the trajectories that universes faced abrupt changes in the initial steps of iterations. These sudden changes are decreased gradually over the course of iterations. According to Berg et al. [30], this behaviour can guarantee that a population-based algorithm eventually converges to a point and search locally in a search space.

Fig. 5
figure 5

Search history, trajectory in first dimension, average fitness of all universes and convergence

Another evidence of convergence and improvement in universes can be concluded from the average fitness of all universes in Fig. 5. The fifth column of this figure shows descending trends in the average fitness over the iterations. This evidences that the model proposed for MVO algorithm can successfully improve the fitness of all universes and guarantee finding better solutions as iteration increases. This can be discussed and reasoned according to the white and black hole concepts of the proposed algorithm. Since the variables of universes tend to move from a fit universe to a less fit universe, similarly to the evolution concepts in GA, the whole universes and their fitness average tend to be improved over the course of iterations. In addition, we save the best universe and move it to the next iteration, so the best universe formed so far is always available to improve the fitness of other universes.

This is similar to the elitism concepts in GA and other evolutionary algorithms. Eventually, the convergence curves in Fig. 5 (last column) show that MVO algorithm finds a better solution iteration by iteration.

4.5 Constrained optimization and classical engineering problems

In this subsection, a set of five constrained real engineering problems is solved in order to further investigate the performance of the proposed MVO algorithm. The following problems have several inequality constraints, so the MVO algorithm should be able to handle them during optimization. Since search agents of the MVO algorithm are fitness independent, the mechanism of this algorithm does not need to be altered for handling constraints. We utilize the simplest constraint handling method, death penalty functions, where search agents are assigned big objective function values if they violate any of the constraints. Assigning big fitness values to the violated universes requires them to have high number of black holes and less number of white holes. Therefore, they mostly tend to receive objects rather than sending them.

4.5.1 Welded beam design using MVO

As its name implies, this problem deals with designing a welded beam with the lowest fabrication cost. The overall system and structural parameters are illustrated in Fig. 6 [31].

Fig. 6
figure 6

Design parameters of the welded beam design problem

This figure shows that there are four design parameters for this problem to be optimized: thickness of weld (h), length of attached part of bar (l), the height of the bar (t), and thickness of the bar (b). The obtained design should not violate seven constraints when a load applies on the top of the bar. Some of the constraints are as follows: side constraints, shear stress (τ), end deflection of the beam (δ), buckling load on the bar (P c ), and bending stress in the beam (θ). Full details for this problem can be found in the “Appendix 2”.

This problem has been very popular in the literature. Some of the works that solved this problem are as follows: GA [3133], CPSO [34], HS [35], Richardson’s random method, Simplex method, Davidon–Fletcher–Powell, and Griffith and Stewart’s successive linear approximation [36]. We compare our algorithm with these works and provide the results in Table 8.

Table 8 Comparison results of the welded beam design problem

The results of the algorithms on welded beam design problem show that the MVO algorithm is able to find a design with the minimum cost. The superiority of the results compared to mathematical approaches is due to the local optima avoidance of this algorithm. The proposed algorithm also outperforms PSO, GA, and GSA, which shows that this algorithm is able to effectively search for optimal solutions in real search spaces as well.

4.5.2 Gear train design using MVO

As may be seen in Fig. 7 and “Appendix 2”, this problem has four parameters, and the objective is the minimization of gear ratio [37]. The gear ratio is calculated as follows [38]:

$${\text{Gear}}\;{\text{ratio}} = \frac{{{\text{angular}}\;{\text{velocity}}\;{\text{of}}\;{\text{output}}\;{\text{shaft}}}}{{{\text{angular}}\;{\text{velocity}}\;{\text{of }}\;{\text{input}}\;{\text{shaft}}}}$$

The parameters of this problem are discrete with the increment size of 1 since they define the teeth of the gears (n A , n B , n C , n D ). There constraints are only limited the variable ranges.

Fig. 7
figure 7

Gear train design problem

We solve this problem with MVO and provide the results in Table 9. This table shows that the MVO algorithm provides similar results compared to other algorithms. The obtained number of gears of MVO and MBA [39] is similar. These results prove that the proposed algorithm is also able to provide very competitive results on the problem with discrete parameters.

Table 9 Comparison results of the gear train design problem

4.5.3 Three-bar truss design using MVO

Generally speaking, truss design problems are very popular in the literature of meta-heuristics [39, 42]. The objective is to design a truss with the minimum weight that does not violate the constraints. The most important issue in designing truss is constraints that include stress, deflection, and buckling constraints. Figure 8 shows the structural parameters of this problem.

Fig. 8
figure 8

Three-bar truss design problem

As can be seen in the full details of this problem in the “Appendix 2”, the objective function is very simple, but it is subject to several challenging constraints. We employ this test function as a specific test bed with emphasize on constraints. The results of the MVO algorithm are compared with other algorithms in Table 10.

Table 10 Comparison results of the three-bar truss design problem

The results of this table show that MVO provides very close results to DEDS and PSO-DE. This evidences that the proposed algorithm is able to effectively optimize challenging constrained problems as well.

4.5.4 Pressure vessel design using MVO

The pressure vessel design problem is another popular engineering test problem in the literature of meta-heuristics [42, 47]. This objective is the minimization of the fabrication cost of a vessel. As shown in Fig. 9, one side of the vessel is flat, whereas the other side has hemispherical shape. The structural parameters to be optimizers are (see Fig. 9 and “Appendix 2”) thickness of the shell (T s), thickness of the head (T h), inner radius (R), and length of the cylindrical section without considering the head (L).

Fig. 9
figure 9

Pressure vessel design problem

We found an optimal design for this problem using MVO and report the relevant results in Table 11. We compare the results to PSO [48], GA [4951], ES [52], DE [53], and ACO [54]. Mathematical methods used are augmented Lagrangian multiplier [40] and branch-and-bound [37].

Table 11 Comparison results for pressure vessel design problem

The results of this table show that the proposed MVO algorithm outperforms all other algorithms.

4.5.5 Cantilever beam design using MVO

The structural optimization problem is to minimize the weight of a cantilever beam with hollow square blocks. There are five squares of which the first block is fixed and the fifth one burdens a vertical load. As may be seen in Fig. 10 and “Appendix 2”, five parameters define the shape of cross section of the cubes.

Fig. 10
figure 10

Cantilever beam design problem

We also solve this problem by the MVO algorithm and present the results in the Table 12. The results of this algorithm for this problem are consistent to those of other real problems, in which the MVO algorithm outperforms all of other algorithms.

Table 12 Comparison results for cantilever design problem

To sum up, the results of the employed five real engineering design problems prove that the MVO algorithm can be very effective in optimizing real problems. In addition, the performance of the proposed algorithm is verified in solving constrained problems. The reason of the superior results of MVO on constrained problems is that this algorithm efficiently balances exploration and exploitation. For one, the concepts of white/black holes promote exploration, which can cover both feasible and infeasible regions of the search space. For another, increasing the existence of wormholes emphasizes exploitation and local search around the best solution obtained so far.

5 Conclusion

This work proposed a novel optimization technique inspired from multi-verse theory in physics. Three conceptual models of white, black, and wormholes were proposed to design the MVO algorithm. A comprehensive comparative study was conducted on 19 test functions and five real engineering problems to investigate the efficiency of the proposed algorithm in solving optimization problems. The exploration, exploitation, and convergence of MVO were benchmarked and discussed on the test functions. The findings demonstrated the value of the proposed algorithm:

  • White holes are more possible to be created in the universes with high inflation rates, so they send objects to other universes and assist them to improve their inflation rates.

  • Black holes are more likely to be appeared in the universes with low inflation rates, so they have higher probability to receive objects from other universes. This again increases the chance of improving inflation rate for the universes with low inflation rates.

  • White/black hole tunnels tend to transport objects from universes with high inflation rates to those with low inflation rates, so the overall/average inflation rate of all universes is improved over the course of iterations.

  • Wormholes tend to appear in any universe randomly regardless of the inflation rate, so the diversity of universe is maintained over the iterations.

  • While/black hole tunnels require universes to abruptly change, causing exploration of the search space.

  • Abrupt changes also assist resolving local optima stagnations.

  • Wormholes randomly re-span some of the variables of universes around the best solution obtained so far over the course of iterations, so this guarantees exploitation around the most promising region of the search space.

  • Adaptive WEP values smoothly increase the existence probability of wormholes in universes. Therefore, exploitation is emphasized during optimization process.

  • Adaptive TDR values decrease the travelling distance of variables around the best universe, a mechanism that increases the accuracy of local search over the iterations.

  • The convergence of the proposed algorithm is guaranteed by emphasizing exploitation/local search proportional to the number of iterations.

For future work, we are planning to design the multi-objective and binary version of the MVO algorithm.