1 Introduction

Solving real-world optimization problems is one of the most challenging tasks. Most of the real-world optimization problems are complex; therefore, many researchers paid attention to propose new optimization methods. In recent years, different meta-heuristic algorithms are developed inspiring from nature, mathematics, physics or even animal behaviour to solve complex real-world problems, efficiently. Mirjalili et al. [1] proposed a novel meta-heuristic algorithm named multi-verse optimizer (MVO) which mimics the physic laws in universe. It uses the definition of black holes and white holes to perform optimization. Mirjalili et al. [1] investigated the performance of the MVO against different state-of-the-art algorithms and showed that the MVO is able to provide very competitive results even in challenging composite benchmark functions. In a different research, Saremi et al. [2] presented a chaos version of the biogeography-based optimization (BBO). They showed that their modification significantly improved the performance of the algorithm. Salimi [3] presented a new meta-heuristic algorithm named stochastic fractal search (SFS). The SFS is based on strong mathematical concepts such as Levy flight and Gaussian walks. He showed that the SFS is very competitive with other algorithms and outperforms them in majority of the benchmark functions. Since most of the optimization problems deal with different types of decision variables such as binary variables, many researchers paid attention to modify existing algorithms to solve various complex problems. For example, Mirjalili et al. [4] proposed a binary bat algorithm to solve complex binary optimization problems. For more information, see [5].

One of the most recently developed algorithms is Grey Wolf Optimizer (GWO). The GWO first proposed by Mirjalili et al. [6]. GWO mimics the hunting behaviour of the grey wolves in nature. GWO is designed very intelligently, which enables it to make an appropriate trade-off between exploration and exploitation abilities of the algorithm. Mirjalili et al. [6] compared the performance of the GWO against state-of-the-art meta-heuristic algorithms and showed that the GWO is able to outperform other algorithms in most of the benchmark functions. Since the GWO performed very efficient in solving complex benchmark functions, many researches started to implement it to solve complex optimization problems in different fields of study. Jayakumar et al. [7] used GWO for optimization of combined heat and power dispatch with cogeneration systems. They showed that the GWO is able to solve the problem, efficiently. Song et al. [8] implemented GWO for parameter estimation in surface waves and demonstrated the superiority of the GWO. Pradhan et al. [9] showed that GWO performs very effective in solving economic load dispatch problem. By solving different problems, Medjahed er al. [10] showed that the GWO is able to solve the hyperspectral band selection problem efficiently. In addition to above researches, which aimed to develop new techniques to solve complex problems or to show the efficiency of the novel algorithms, some researchers aimed to optimize the performance of the existing algorithms. One of the most commonly used approaches to optimizing the efficiency of the meta-heuristic algorithms is to calibrate the values of input parameters of the algorithms using a robust design methodology. One of the most widely used robust design methods is Taguchi method [11, 12]. Najafi et al. [13] used Taguchi method to tune the parameters of the genetic algorithm (GA) to solve a complex resource investment problem. Sadeghi et al. [14] utilized Taguchi method to calibrate the parameters of the GA to solve a constrained nonlinear programming model of the vendor-managed inventory (VMI) problem. To show the efficiency of the Taguchi method, Sadeghi et al. [15] used Taguchi method to design a robust GA to solve a NP-hard hybrid problem.

In this research, the aim is to optimize the milling process which is one of the most widely used machining processes. In milling process, to enhance the final quality of machined products, many researches focused on the optimization of the total production cost, total production time, final surface roughness and material removal rate. In this category, the input parameters of the machining process must be selected attentively to achieve the favourable outcomes. Instead of single pass, multi-pass machining has been utilized in turning, boring and face milling processes to decrease the power consumption, machining forces, chatter phenomena and depreciation rate of the machine tools. Many experimental and analytical methods have been developed to optimize multi-pass milling parameters considering operational constraints such as permissible cutting force, power and tool life. The main parameters of the problem include the number of passes, depth of cut, cutting speed and feed rate. Therefore, there are different variables and operational constraints that must be considered in multi-pass milling process which results in a constrained nonlinear programing (NLP) model. Due to nonlinearity and complexity of the problem, traditional optimization algorithms usually obtain local optimal solutions. Conversely, meta-heuristic algorithms based on their exploration ability can obtain better solutions comparing to the traditional optimization techniques. In recent decades, scholars utilized meta-heuristic algorithms to optimize multi-pass milling process.

To specify the number of passes and optimum values of the input parameters in multi-pass milling process, Sonmez et al. [16] developed a dynamic programming (DP) and utilized geometric programming (GP) to determine the optimal cutting variables. The results demonstrated that different depths of cut are better than equal depths for multi-pass milling process. Wang et al. [17] proposed a combinational optimization method using GA and simulated annealing (SA). The archived values for the cutting speed, feed rate and machining time have been compared with the DP and conventional parallel GA (PGA). The results showed the high capability of the presented method to find proper values for milling parameters. Onwubolu [18] utilized various depths of cut to minimize production time with a new optimization method called Tribes.

To find the maximum total profit rate in the process, Baskar et al. [19] introduced a new algorithm using the genetic algorithm and hill climbing algorithm for optimization of the milling parameters. Yildiz [20] proposed a combinational optimization method to find maximum value of the total profit rate in milling processes and compared the results with genetic algorithm in the literature. Gao et al. [21] used cellular particle swarm optimization (CPSO) technique to optimize multi-pass milling process. The comparison showed that their approach was better than other optimization algorithms in obtaining more proper values for milling parameters. Other algorithms have been introduced by Rao and Pawar [22] and Pawar and Rao [23] named the artificial bee colony (ABC) and teaching–learning-based optimization algorithm (TLBO) to optimize machining processes, respectively. Yang et al. [24] used a plain method to predict the optimum values of the multi-pass milling parameters. Recently, Mellal and Williams [25] suggested a cuckoo search optimization algorithm approach to minimizing the total production time in the milling process. The results compared to the previous researches in the literature and showed that cuckoo algorithm is an appropriate optimization method to specify minimum total production time considering process constraints and variables. In different research, Mellal and Williams [26] used a cuckoo optimization algorithm to optimize machining process parameters. They showed that the CS is able to provide very competitive results. In more recent research, Mellal and Williams [27] utilized a cuckoo optimization algorithm to optimize production costs in multi-pass milling process. Beside the wide application of meta-heuristic algorithms, many researchers aimed to optimize machining, turning and grinding processes using hybrid methods which use neural networks and meta-heuristic algorithms, simultaneously. For more information, see [2832].

Most of the researches in the literature proposed a meta-heuristic algorithm to optimize the parameters of the milling process and find better solutions. Since the proposed mathematical formulation is a nonlinear programing formulation (NLP) and hard to be solved [33], a meta-heuristic algorithm is needed to solve the problem which can make a proper trade-off between exploration and exploitation phrases of the algorithm to avoid trapping in local optima and fast convergence, simultaneously.

In this paper, a robust nature inspired algorithm named Robust Grey Wolf Optimizer (RGWO) is developed to optimize multi-pass milling process parameters. Since the values of the main parameters of meta-heuristic algorithms significantly affect the performance, therefore Taguchi method is utilized to obtain the optimal values of the main parameters of the algorithm to increase its efficiency, while in previous researches in the literature, the values of the main parameters of the developed algorithms are determined by trial and error. In addition, an effective constraint handling technique is used to handle complex operational constraints of the problem. Efficiency and effectiveness of the proposed RGWO is investigated by solving a real-world problem proposed by Sonmez et al. [16]. The results are compared to the previous researches and demonstrated that the RGWO can find significantly better solutions comparing to other solution methodologies in the literature.

2 Mathematical model

In this research, the mathematical formulation presented by Sonmez et al. [16] is implemented to optimize the multi-pass milling process parameters, where the main parameters of the milling process are considered as decision variables of the problem include: feed per tooth and cutting speed. The depth of cut (a) as one of the main parameters of the process is considered to be known. Using different strategies, the value of depth of cut is given in the problem for each pass. The goal of this research is to minimize total production time in multi-pass milling process. The notations are presented in Table 1.

Table 1 Parameters and decision variables

2.1 Objective function

The total production time in the multi-pass milling process is the sum of the required times for machining the workpiece with several single passes as presented in Eq. (1).

$$\begin{aligned} &T_{\text{pr}} (f_{{z_{i} }} ,V_{i} ,a_{i} ) = T_{\omega } + T_{\varphi } = \left\{ {T_{\text{L}} + N_{\text{p}} T_{\text{a}} + \frac{{T_{\text{S}} }}{{N_{\text{b}} }}} \right\} \\&\quad + \left\{ {\sum\limits_{i = 1}^{{N_{\text{p}} }} {\frac{\pi DL}{{1000ZV_{i} f_{{z_{i} }} }}} + \frac{{\pi LV_{i}^{{\left( {\frac{1}{m} - 1} \right)}} a_{i}^{{\left( {\frac{{e_{\text{v}} }}{m}} \right)}} a_{r}^{{\left( {\frac{{r_{\text{v}} }}{m}} \right)}} \lambda_{\text{s}}^{{\left( {\frac{{q_{\text{v}} }}{m}} \right)}} Z^{{\left( {\frac{{n_{\text{v}} }}{m} - 1} \right)}} f_{{z_{i} }}^{{^{{\left( {\frac{{u_{v} }}{m} - 1} \right)}} }} T_{\text{d}} }}{{1000 \times D^{{\left( {\frac{{b_{\text{v}} }}{m} - 1} \right)}} \left( {B_{\text{m}} B_{\text{h}} B_{\text{p}} B_{\text{t}} } \right)^{{\left( {\frac{1}{m}} \right)}} C_{\text{v}}^{{\left( {\frac{1}{m}} \right)}} }}} \right\} \end{aligned}$$
(1)

2.2 Constraints

Multi-pass milling process has some experimental and operational constraints which depend on the material and geometry of the workpiece. These operational constraints significantly reduce the feasible solution space of the problem and make the problem hard to be solved.

2.2.1 Arbour stability

The optimal values of the decision variables should satisfy the arbour stability constraint as presented in Eq. (2).

$$ C_{zp} Za_{r} a_{i}^{{e_{z} }} D^{{b_{z} }} f_{{z_{i} }}^{{u_{z} }} - \frac{{0.1k_{\text{b}} d_{\text{a}}^{3} }}{{\left( {0.08L_{\text{a}} } \right) + \left\{ {0.65\sqrt {\left( {0.25L_{\text{a}} } \right)^{2} + \left( {\frac{{0.38461k_{\text{b}} D}}{{k_{\text{t}} }}} \right)^{2} } } \right\}}} \le 0 $$
(2)

where the first phrase of Eq. (2) shows the mean peripheral cutting force and the second phrase presents permissible force.

2.2.2 Arbour deflection

The arbour deflection constraint should be satisfied by the optimal values of the decision variables. Equation (3) presents the arbour deflection constraint.

$$ C_{zp} Za_{r} a_{i}^{{e_{z} }} D^{{b_{z} }} f_{{z_{i} }}^{{u_{z} }} - \left( {4Ed_{\text{a}}^{4} } \right)\frac{e}{{L_{\text{a}}^{3} }} \le 0 $$
(3)

where the second phrase of Eq. (3) shows the permissible force with regard to the arbour deflection.

2.2.3 Power constraint

The following constraint ensures that the power required for the cutting operation is less or equal to the effective transmitted power to the cutting point which is given by Eq. (4).

$$ \frac{{C_{zp} Za_{r} a_{i}^{{e_{z} }} D^{{b_{z} }} f_{{z_{i} }}^{{u_{z} }} V_{i} }}{6120} \le P_{\text{m}} \eta $$
(4)

The last constraints present the lower and upper bound constraints of the decision variables.

$$ f_{{z_{\hbox{min} } }} \le f_{{z_{i} }} \le f_{{z_{\hbox{max} } }} $$
(5)
$$ V_{\hbox{min} } \le V_{\text{i}} \le V_{\hbox{max} } $$
(6)

3 Solution methodology

The proposed mathematical model in the previous section is a constrained nonlinear programming (CNLP) model, which is hard to be solved using exact methods [33]. In last decade, many meta-heuristic algorithms have been proposed and commonly used to optimize complex NLP problems. Most of them are inspired from the nature such as water cycle algorithm [3436], or animal’s behaviour such as particle swarm optimization [37, 38], or other physical phenomena such as gravitational search algorithm [39]. In order to solve the above model, a new meta-heuristic algorithm named GWO is utilized [6]. As mentioned earlier, the GWO is widely used to solve engineering problems, while in this research GWO is used to optimize machining process parameters for the first time. GWO uses intelligent behaviour of the grey wolves in hunting and attacking prey in nature which helps the algorithm to perform well in solving complex problems. In this paper, a robust GWO is developed using Taguchi method.

3.1 Grey Wolf Optimizer

The GWO is inspired from grey wolves and their behaviour in nature. Grey wolves are one of the most famous predators in nature. They usually live in pack with five to twelve members. One of the most interesting behaviours of grey wolves is their social dominant hierarchy [6]. The hierarchy of the wolves and their responsibilities are presented in Fig. 1.

Fig. 1
figure 1

Grey wolves’ hierarchy and their responsibilities

The alpha wolf is known as the leader of the pack, and his/her responsibility is to make decisions about hunting and other activities in the pack such as find a place for sleep. The beta wolf helps the alpha wolf in making decisions. The omega wolves are the lowest in ranking in the pack. Omegas are responsible to submit information to alpha, beta and delta wolves. All other wolves are called delta. The delta wolves should respect to alpha and beta, and they dominate omega wolves. There are three steps in hunting process of grey wolves as described in following [6].

  • Tracking the prey.

  • Encircling the prey.

  • Attacking to the prey.

GWO is a mathematical representation of the hunting method of the grey wolves. GWO uses the hunting process of the grey wolves to solve complex optimization problems. For this purpose, the optimal solution of the problem is considered as prey. In each iteration of the GWO, first three best solutions are considered as alpha, beta and delta, respectively. The rest of the solutions are considered as omega. Hunting is leaded by alpha, beta and delta wolves, and the omegas follow these dominant wolves.

3.1.1 Encircling prey

The first step in hunting process is encircling prey. The mathematical formulation to mimic the encircling process is given in the following.

$$ Di\vec{s} = \left| {\vec{C} \cdot \vec{X}_{\text{p}} (t) - \vec{X}(t)} \right| $$
(7)
$$ \bar{X}(t + 1) = \vec{X}_{\text{p}} (t) - \vec{A} \cdot \vec{D} $$
(8)

where \( \vec{C} \) and \( \vec{A} \) are coefficient vectors, t is the current iteration, while \( \vec{X}_{\text{p}} \) and \( \vec{X} \) are the position of prey and a randomly chosen grey wolf, respectively [6]. The coefficients \( \vec{C} \) and \( \vec{A} \) are determined using following formulas.

$$ \vec{A} = 2\vec{a} \cdot \vec{r}_{1} - \vec{a} $$
(9)
$$ \vec{C} = 2 \cdot \vec{r}_{2} $$
(10)

where the \( \vec{a} \) vector decreases over the iterations of the GWO from two to zero and \( \vec{r}_{1} \), \( \vec{r}_{2} \) are randomly generated parameters between 0 and 1. Equations (9) and (10) allow the grey wolves to change their positions around the prey. This allows the GWO to search the n-dimensional solution space of the problem more efficiently.

3.1.2 Hunting

As mentioned above, the alpha, beta and delta wolves guide hunting process and rest of the wolves in the pack follow them. Mathematically speaking, the wolves update their position using the following equations.

$$ \vec{D}_{\alpha } = \left| {\vec{C}_{1} .\vec{X}_{\alpha } - \vec{X}} \right|, \;\; \vec{D}_{\beta } = \left| {\vec{C}_{2} \cdot \vec{X}_{\beta } - \vec{X}} \right|, \;\; \vec{D}_{\delta } = \left| {\vec{C}_{3} \cdot \vec{X}_{\delta } - \vec{X}} \right| $$
(11)
$$ \bar{X}_{1} = \vec{X}_{\alpha } - \vec{A}_{1} \cdot \vec{D}_{\alpha } ,\;\; \bar{X}_{2} = \vec{X}_{\beta } - \vec{A}_{2} \cdot \vec{D}_{\beta } ,\;\; \bar{X}_{3} = \vec{X}_{\delta } - \vec{A}_{3} \cdot \vec{D}_{\delta } $$
(12)
$$ \bar{X}(t + 1) = \frac{{\vec{X}_{1} + \vec{X}_{2} + \vec{X}_{3} }}{3} $$
(13)

3.1.3 Attacking prey

Decreasing the value of \( \vec{a} \) over the iterations from 2 to 0 decreases the value of \( \vec{A} \) over the iterations as well. As the value of \( \vec{a} \) decreases, the grey wolves are more likely to attack towards the prey and get closer and closer to the prey over the last iterations of the GWO. In other words, the search radius of the grey wolves decreases over the course of iterations. This makes a proper trade-off in exploration and exploitation by focusing on exploration at first iterations and exploitation in last iterations.

Actually, the proposed basic GWO is to solve unconstrained problems and no version is developed to solve constrained problems. Therefore, we modified the GWO to obtain the optimal values of the decision variables of the constrained NLP model. The pseudo-code of the developed GWO is given in Table 2.

Table 2 Pseudo-code of the GWO

3.1.4 Constraint handling

During constrained optimization, both inequality and equality constraints should be satisfied. In this process, solutions are divided into two solutions, feasible and infeasible solutions [40]. As classified by Coello [41], five methods have been used to handle constraints during constrained optimization process: penalty functions, repair algorithms, hybrid methods, special operators and separation of objective functions and constraints. The penalty functions are the simplest technique to handle constraints which penalize the infeasible solutions. Consider a NLP problem as follows:

$$ \begin{aligned} & {\text{Min}} \,\,z = f(\vec{x}) \\ & {\text{s}} . {\text{t}} .\\ & g_{i} (x) \le 0 \\ & h_{j} (x) = 0 \\ \end{aligned} $$
(14)

The penalty functions convert a constrained optimization model to an unconstrained one. The general formulation of the penalty functions can be presented as follows [41]:

$$ \beta (\vec{x}) = f(\vec{x}) \pm \left[ {\sum\limits_{i = 1}^{n} {r_{i} \times \hbox{max} \left[ {0,\;g_{i} (x)} \right]^{\beta } + \sum\limits_{j = 1}^{p} {c_{j} \times \left| {h_{j} (x)} \right|} } } \right] $$
(15)

where \( \beta (\vec{x}) \) shows the modified objective function and \( r_{i} \), \( c_{j} \) are positive penalty factors [41]. There are different penalty function techniques which aim to handle the constraints in different manners such as dynamic penalty, static penalty, adaptive co-evolutionary penalty, annealing penalty and death penalty functions [42].

Most of the researchers in the literature used death penalty method to handle constraints in solving parameter optimization problems. Death penalty technique assigns a big positive number in minimization problems to objective function value of any infeasible solution. In other words, this technique does not use any information of infeasible solutions. This point makes the death penalty a weak technique to handle complex constraints in an optimization problem, that is why the algorithm needs more time to find the feasible solution space. Besides, the information of the infeasible solutions can help the algorithm to find the feasible solution space much faster. For this purpose, in this research static penalty technique is used to handle complex constraints of the proposed mathematical model. This method uses the information of the infeasible solutions (violation of each constraint) to reach to the feasible solution space. In the constraint handling approach proposed by Homaifar et al. [43], the user determines different levels of violation, and penalty constants for all constraints in a way that the penalty coefficient increases as the algorithm reaches higher violations [4446].

$$ {\text{fitness}}\,(\vec{x}) = {\text{objective}}\, {\text{function}} \,(\vec{x}) + \sum\limits_{i = 1}^{n} {R_{k,i} \times \hbox{max} \left[ {0,g_{i} (\vec{x})} \right]^{2} } $$
(16)

where \( R_{k,i} \) are user defined positive penalty coefficients. This methodology is utilized in this paper to handle complex operational constraints of the problem.

4 Tuning the parameters of the GWO

In all of the meta-heuristic algorithms, the values of input parameters of the algorithms play a prominent role in the performance of the algorithms. As mentioned above, GWO, as a meta-heuristic algorithm, has three main input parameters which effect the performance of the algorithm. The first one is the number of grey wolves (number of search agents). Consider a milling process where the aim is to minimize the total production time considering four passes, three rough millings at first and one finish milling at the end. Figure 2 presents the convergence curve of the GWO with different numbers of wolves in solving the problem.

Fig. 2
figure 2

Effect of change in the number of search agents on the performance of the GWO

From Fig. 2, it is obvious that the efficiency of the GWO is significantly influenced by the number of grey wolves. The second parameter, which changes the performance of the GWO, is the convergence constant. The convergence constant is the most important parameter in the GWO, since it determines the search radius of the grey wolves and makes an appropriate trade-off between exploration and exploitation in GWO. To show the effect of change in this parameter, Fig. 3 is presented.

Fig. 3
figure 3

Effect of change in the convergence constant on the performance of the GWO

The third parameter (MaxIt) determines the maximum number of iterations. This parameter not only affects the performance of the GWO, but also effects the performance of all the meta-heuristic algorithms. Figure 4 depicts the effect of this parameter on the performance of the GWO.

Fig. 4
figure 4

Effect of change in the maximum number of iterations on the performance of the GWO

As mentioned above, the performance of the GWO is directly influenced by the main parameters of the GWO. Therefore, determining the best values of the main parameters of the GWO results in better solutions. Different methodologies are proposed by the researchers to obtain the optimal level of the parameters of a meta-heuristic algorithm such as Taguchi method and response surface methodology. In this research, the Taguchi method approach is utilized to develop a robust GWO.

Factorial designs are known as design of experiments (DOE) proposed by Fisher and aim to determine the effect of several factors on a response [11, 12]. In this research, the objective function value is considered as the response and the factors are the main parameters of the GWO. Therefore, the effect of change in main parameters of the GWO on objective function value is investigated to determine the optimal values of the main parameters of the algorithm. In the Taguchi method, the main factors are classified into two groups: controllable factors (S) and uncontrollable factors or noise factors (N). Taguchi method aims to determine control factors by minimizing the effect of noise factors to reduce the variation around response and to obtain a robust design [1315].

4.1 Taguchi method

The main parameters of the GWO are tuned using Taguchi method to obtain better solutions and as fast as possible without trapping in local minima.

4.2 Grey Wolf Optimizer parameter tuning

Four levels of possible values are considered for each parameter of the GWO. Table 3 presents the levels of the parameters of the GWO.

Table 3 Different levels of the parameters

The \( L^{9} \) orthogonal array of the Taguchi method is used to find the optimal level of the parameters of the GWO. GWO is utilized to solve a milling process where the aim is to minimize the total production time considering four passes including three rough millings and one finish milling at the end to obtain the optimal values of the GWO parameters. For this purpose, a real case provided by Sonmez et al. [16] is utilized to develop RGWO. The values of required parameters of the case are presented in Table 4.

Table 4 Input parameters of the numerical example

The problem is solved five times using GWO. Each repetition is shown by y i . Table 5 presents the objective function values in five runs.

Table 5 Experimental results of the GWO

Figure 5 depicts the performance of the GWO in fourth, eighth, twelfth, thirteenth and fifteenth orders.

Fig. 5
figure 5

Performance of the GWO in fourth, eighth, twelfth, thirteenth and fifteenth orders

The first column of Fig. 5 presents the best solution over the course of iterations. The second column of Fig. 5 shows the average objective function value of grey wolves which are in the feasible solution space of problem. The third column of Fig. 5 presents the average penalty values of the wolves that are out of the feasible solution space of the problem. The last column of Fig. 5 presents the convergence constant value over the iterations.

To perform the Taguchi method, first the obtained objective functions values are normalized using linear norm. The linear normalization assigns one to the best solution (minimum objective function value) and positive values between 0 and 1 to other solutions based on their deviation from the best solution. Therefore, the solution with larger normalized value is preferred. Since the aim is to find solutions with larger normalized value, highest rate in S/N ratio is suitable, see [11, 12]. The S/N ratio for this type objective function is calculated as follows:

$$ S/N = - 10\log \left( {\frac{1}{n}\sum\limits_{i = 1}^{n} {\frac{1}{{y_{i}^{2} }}} } \right) $$
(17)

where n is the replications which in our example is five and \( y_{i} \) is the normalized objective function value of the different combinations of the parameters levels. Table 6 presents the results.

Table 6 Normalized values of the experimental results of GWO

Figure 6 presents the results of the Taguchi method implementation. As in Fig. 6, the highest value of the S/N ratio is the best level of the parameter.

Fig. 6
figure 6

Results of the Taguchi method

From the results, the optimal values of the parameters of the GWO algorithm can be determined. The optimal value of the \( \alpha \) parameter (convergence constant) is 0.5. The optimal number of grey wolves in the initial population is 70, and the optimal number of iterations to solve the problem is 1150.

5 Results and discussion

The real case provided by Sonmez et al. [16] is utilized to evaluate the efficiency of the developed RGWO in optimizing the multi-pass milling process. Three main strategies are adopted for given values of \( a_{i} \) as reported in the literature. The first two strategies include three rough and one finish passes. The third strategy contains one rough and one finish passes. The value of input parameters of the case is as presented earlier in Table 4. Table 7 presents the three cutting strategies and the value of the parameter a in each pass.

Table 7 Three cutting strategies

To optimize the parameters of the multi-pass milling process using the RGWO, the initial pack of wolves is created as shown in Table 8 [4446], where rand is a random number between 0 and 1.

Table 8 Generating the initial population in GWO

Tables 9, 10 and 11 present the computational results of optimizing the multi-pass milling process using the developed RGWO within different strategies. In addition, the results of the RGWO are compared with other meta-heuristic algorithms in the literature such as particle swarm optimization, simulated annealing and artificial bee colony algorithm. To validate the results of the RGWO, the performance of the RGWO is compared to two novel meta-heuristic algorithms named dragonfly algorithm and multi-verse optimizer within the same number of function evaluations, as well [1, 47].

Table 9 Computational results of the first strategy
Table 10 Computational results of the second strategy
Table 11 Computational results of the third strategy

From Table 9, the proposed RGWO obtains better solutions which has significantly less production time comparing to other solution methodologies in the literature. In addition, the two novel meta-heuristic algorithms obtained two solutions which are significantly worst than the solution of the RGWO. A schematic view of the comparison of RGWO and previous researches is presented in Fig. 7.

Fig. 7
figure 7

Comparison of RGWO and other researches in the first strategy

The computational results of the second strategy are given in Table 10. The second strategy contains three rough and one finish passes.

From Table 10, it is clear that the RGWO obtained better solutions with less total process time comparing to other methods in the literature. In addition, the solution obtained by the RGWO satisfies all the constraints of the problem. Figure 8 depicts the comparison of the results of the RGWO and other solution methods in the literature in a graphical manner.

Fig. 8
figure 8

Comparison of RGWO and other researches in second strategy

Table 11 presents the computation results of the third strategy. In contradiction with the first two strategies, the third strategy includes one rough and one finish passes.

One of the important points in the results of the third strategy is that in most of the previous researches, the developed meta-heuristic algorithms could not obtain a solution which satisfy all constraints; in other words, their solutions are not feasible. Conversely, the developed RGWO in this research determines better feasible solutions which satisfy all constraint. This shows the efficiency and applicability of the proposed solution methodology in optimizing multi-pass milling process. In the GWO, the omega wolves follow the three leading wolves (alpha, beta and delta wolves) and update their position in each iteration with respect to the position of these three leading wolves. In other words, they share the best solution in the algorithm. This approach helps the search agents (omega wolves) to find the feasible solution space faster and the best solution as well [4446]. Figure 9 shows the comparison of the results of the RGWO in third strategy and other solution methodologies in the literature, where points in the red circle are infeasible solutions obtained in the previous researches.

Fig. 9
figure 9

Comparison of RGWO and other researches in third strategy

As is shown in the above comparison, the performance of the RGWO is significantly better than other algorithms in the literature as well as other novel meta-heuristic algorithms, but why? To perform well, each meta-heuristic algorithm has to explore and exploit the solution space of the problem, efficiently. GWO performs very well in making an appropriate trade-off between exploitation and exploration using the convergence constant. As mentioned earlier, the value of the convergence constant decreases over the course of iterations of the GWO. The convergence constant is similar to search radius of the grey wolves. Therefore, in the first iterations, when the value of the convergence constant is high, the GWO focuses on exploring of the solution space and finding better solutions. Then, the GWO starts to decrease the convergence constant, and this process decreases the search radius of the grey wolves. Thus, GWO aims to exploit the solution space more efficiently as the iterations go on. By making this appropriate trade-off, GWO is able to explore and exploit the solution space effectively and this helps the GWO to obtain better solutions. The second reason is that the GWO can reach the feasible solution space of the problem very fast. In the first iterations, most of the randomly generated solutions are infeasible, and the static penalty approach calculates the violation level of each infeasible solution and chooses the best infeasible solution as alpha. Therefore, the omega wolves move towards a solution which is the best infeasible solution with least violation level. Thus, all the wolves in the pack move very fast towards the feasible solution space, and when alpha becomes feasible, all other wolves start to follow him instantly towards the feasible solution space.

As is known, the values of the main parameters of the meta-heuristic algorithms significantly affect their performance. In this research, the main parameters of the GWO were tuned using a robust design approach within five repetitions to obtain the best possible solution. One of the most critical parameters of the GWO is convergence constant. The optimal value of this parameter is also obtained using Taguchi method. As mentioned earlier, this parameter makes the appropriate trade-off between exploration and exploitation abilities of the GWO. Thus, by obtaining the optimal value of this parameter, the GWO reached its maximum efficiency in solving the problem, since it makes the best possible trade-off between exploration and exploitation. All the above-mentioned three reasons make the GWO a powerful algorithm in solving the problem.

As is clear, the other algorithms in the literature lack from an efficient exploration and exploitation abilities. The best values of the main parameters of those algorithms are obtained by trial and error which does not guarantee the optimality of the values of the main parameters. In addition, they may lack from an efficient constraint handling procedure.

Note that in all the simulations, the number of function evaluations (NFE) of the RGWO was 80,500, while the computation time was <10 s. Although the NFE of the RGWO is high, it can solve the problem accurately and very fast.

6 Conclusion

This paper aims to minimize total production time in a multi-pass milling process. Since the mathematical model of the multi-pass milling process is nonlinear programming, traditional techniques cannot solve the problem efficiently. Therefore, in this research a novel RGWO developed to solve the optimization problem. A robust design methodology named Taguchi method was implemented to tune the parameters of the GWO. Thus, in contradiction to previous researches, costly experiments are not required to calibrate the values of input parameters of the algorithm. The optimal values of the main parameters of the GWO guarantee the best appropriate trade-off between exploration and exploitation abilities of the algorithm. Also, an efficient constraint handling technique was utilized to handle complex operational constraints of the problem. By solving an experimental example, the effectiveness and applicability of the proposed RGWO was investigated. The performance of the RGWO was compared to the other solution methods in the literature as well as two novel meta-heuristic algorithms named dragonfly algorithm and multi-verse optimizer. The results indicated that the RGWO outperforms the other solution methods in the literature as well as novel meta-heuristic algorithms in minimizing total production time. Using the robust design methodology and efficient constraint handling approach, the RGWO was able to obtain the best feasible solution for different cutting strategies.

For future research, it is very interesting to consider some of the main parameters of the mathematical model under uncertainty and develop more realistic models. Considering different other operational constraints of the multi-pass milling process as well as multiple objectives will absolutely help the decision-maker to choose the best possible solution. Developing other efficient solution methodologies to solve the problem would be worthwhile. Using other constraint handling techniques to solve the problem is another extension of this research.