Introduction

Sustainable and responsible in mining is the explicit goals of researchers in recent years since the negative impacts on environments of mining practices, especially surface mining (Bakhtavar et al., 2021; Bui and Ho 2020; Patra et al., 2016; Poudyal et al., 2019; Schueler et al., 2011; Tiwary, 2001; Zeng et al., 2020). Among the operations in open-pit mines, blasting is the most sensitive to the environment (Ainalis et al., 2017; Katsabanis, 2020). It contributes a large amount of dust and toxic gas to the air environment. Not only that, but blasting activities are even more dangerous when generating harmful waves to the surrounding environment, such as shock waves and air over-pressure (Faramarzi et al., 2014). At a gradual distance from the center of the explosion, the shock wave turned into blast-induced ground vibration (BIGV) (Nguyen et al., 2020a, b, c). Furthermore, the structures may be broken, vibration, benches may be instability and fail at a high intensity of BIGV (Bui et al., 2019). Therefore, accurate prediction of BIGV has a vital role in mitigating the damages of structures, buildings, and benches/slopes.

For predicting BIGV, many researchers proposed and applied different approaches, and they can be categorized into two groups: empirical equations and soft computing models. Whereas empirical equations are considered the “white box” mathematics, most soft computing models are taken into account as the “black box” mathematics (Le and Dang 2020; Nguyen 2020). However, the accuracy and reliability of the “black box” mathematics were recommended as better as the “white box” mathematics in predicting BIGV (Ghoraba et al., 2016; Saadat et al., 2014). As a matter of fact, many artificial intelligence (AI) models were proposed with promising results. The proposed AI models (in the previous papers) and their results are listed in Table 1.

Table 1 Overview of the AI models proposed in the past decade and their results

From a comprehensive review in Table 1, it is conspicuous that many state-of-the-art techniques based on AI have been introduced to predict BIGV with different results reported. Indeed, various meta-heuristic algorithms have been applied to enhance the accuracy of the classic models, such as PSO, GWO, GA, ICA, CSO, to name a few. They worked based on the swarm, evolutionary, and social behaviors to generate many solutions for predicting BIGV. Each solution corresponds to a fitness of the model, and finally, the best solution was selected based on the best performance of the benchmark models. Based on the recommendations of previous researchers, hybrid models based on meta-heuristic algorithms often providing better performance over the conventional models. Thus, current models' trend is often integrated between optimization algorithms and other powerful machine learning algorithms. This study introduces a novel and robust hybrid model for predicting BIGV based on the Hunger Games Search optimization algorithm (HGS) and ANN, named as HGS–ANN model. It is worth noting that the HGS algorithm was just introduced by Yang et al. (2021) on March 04, 2021. Yet, it has not been considered and combined with the ANN or other models for any problems. Therefore, the HGS–ANN model was proposed the first time in this study and applied for predicting BIGV. Note that it can be widely applied to other problems in real life. Besides, three other hybrid models, such as PSO–ANN, FFA–ANN, and GOA–ANN, were also developed and compared to the proposed HGS–ANN model to emphasize the robustness of the proposed HGS–ANN model.

Methodology

As introduced in the above section, this study considered and developed different ANN models based on various meta-heuristic algorithms (i.e., HGS, PSO, FFA, and GOA). Due to the FFA–ANN, PSO–ANN, and GOA–ANN models have been introduced in many published papers (Hasanipanah et al., 2016; Heidari et al., 2019; Le et al., 2019; Shang et al., 2019; Moayedi et al., 2020; C. Xie et al., 2021), they have not been mentioned in this section. This section only focuses on proposing the framework of the novel HGS–ANN model for predicting BIGV.

Artificial Neural Network (ANN)

ANN is incredibly famous for its applications in real life. It was developed based on the biological neuron’s functionality of the human brain (Basheer & Hajmeer, 2000; Falahian et al., 2015). There are many types of ANN, such as feedforward, regulatory feedback, physical, recurrent, convolutional, radial basis function, and dynamic neural networks, to name a few (Aggarwal & Song, 1998; T. Xie et al., 2011; Suzuki, 2013; Tsantekidis et al., 2017; Al Kheraif et al., 2019; Yazdinejad et al., 2020). In this work, the multiple layers perceptron (MLP), which is a class of feedforward neural networks, was used to predict BIGV.

Theoretically, the network topology of an MLP model includes an input layer, one or multiple hidden layers, and one output layer (Fig. 1). Accordingly, the information is moved from the input layer through the hidden layers and finally summaries the results to the output layer in one direction. In each layer of an MLP (except the input layer), neurons (nodes) are the computational units that can help the MLP model achieves the expected results. In order to implement the computation, ANN uses a training algorithm, such as backpropagation, feedforward, and Levenberg–Marquardt. It can compute the connections between nodes through their weights and biases. Besides, activation functions are also the critical components of an MLP neural network. They are nonlinear functions (e.g., relu, edlu, tanh, sigmoid) applied to the outputs of hidden layers in an MLP, and they are then used as the inputs for the next layer. In this way, MLP can detect and explain complex relationships of the dataset. In other words, without activation functions, the predictive power of the MLP will be limited and significantly reduced. The combination of activation functions between hidden layers is to help the model learn complex nonlinear relationships latent in the data. Due to the principle of MLP has been introduced in many published papers and books (Basheer & Hajmeer, 2000; Ertuğrul, 2018); therefore, this section only briefly introduces the principle of MLP neural network. Further details of MLP neural networks can be found in the following materials (Alanis et al., 2019; Bre et al., 2018; Hassoun, 1995; Livingstone, 2008; Suzuki, 2013). In this study, MLP neural network was simply called ANN for predicting BIGV, and its architecture is illustrated in Figure 1.

Figure 1
figure 1

General architecture of an ANN for predicting BIGV

Hunger Games Search (HGS) Optimization Algorithm

HGS is a new meta-heuristic algorithm introduced and proposed by Yang et al. (2021). It is designed inspired by the cooperative behavior of the animal and their hunger-driven activities. The mathematical model of HGS was proposed based on a simple structure, but the performance is very competitive. It includes two stages: (1) approach food; and (2) hunger role.

In the first stage, the cooperative communication and foraging behavior of individuals are described as:

$$\overrightarrow {{G\left( {t + 1} \right)}} = \left\{ {\begin{array}{*{20}l} {{\text{Game}}_{1} = \overrightarrow {{G\left( t \right)}} \cdot \left( {1 + {\text{rand}}\left( {} \right)} \right),} \hfill & {\sigma _{1} < l} \hfill \\ {{\text{Game}}_{2} = \overrightarrow {{W_{1} }} \cdot \overrightarrow {{G_{{\text{b}}} }} + \overrightarrow {R} \cdot \overrightarrow {{W_{2} }} \cdot \left| {\overrightarrow {{G_{{\text{b}}} }} - \overrightarrow {{G\left( t \right)}} } \right|,} \hfill & {\sigma _{1} > l,\,\sigma _{2} > S} \hfill \\ {{\text{Game}}_{3} = \overrightarrow {{W_{1} }} \cdot \overrightarrow {{G_{{\text{b}}} }} - \overrightarrow {R} \cdot \overrightarrow {{W_{2} }} \cdot \left| {\overrightarrow {{G_{{\text{b}}} }} - \overrightarrow {{G\left( t \right)}} } \right|,} \hfill & {\sigma _{1} > l,\,\sigma _{2} < S} \hfill \\ \end{array} } \right.$$
(1)

where \(\sigma _{1} \;{\text{and}}\;\sigma _{2}\) are random numbers \(\left( {0 \le \sigma _{1} ,\,\sigma _{2} \le 1} \right)\); \({\text{rand}}\left( {} \right)\) is also a random number but it must be satisfied a normal distribution; \(t\) is the current iteration; \(\overrightarrow {{W_{1} }} \;{\text{and}}\;\overrightarrow {{W_{2} }}\) stand for the weights of hunger behavior; \(\overrightarrow {{G_{{\text{b}}} }}\) stands for the best position of the best individual; \(l\) is a constant; S is a variation control, which is determined as:

$$S = {\text{hyp}}{\text{.}}\;{\text{func}}\left( {\left| {{\text{fitness}}_{i} - {\text{fitness}}_{{{\text{best}}}} } \right|} \right)$$
(2)

where \({\text{hyp}}.\;{\text{func}}\) is a hyperbolic function; \(\overrightarrow {R}\) is a number in the range of [− a, a] and it is used to control the range of activity:

$$\overrightarrow {R} = 2 \times \alpha \times \beta - \alpha$$
(3)

where \(\beta\) is a random number interval [0,1]; \(\alpha\) is the shrink of iterations, thus:

$$\alpha = 2 \times \left( {1 - \frac{t}{T}} \right)$$
(4)

where T is the maximum number of iterations.

In the second stage, the characteristics of hunger-driven activities are simulated. Accordingly, the weights of hunger behavior are represented in Eqs. 5 and 6:

$$\overrightarrow {{W_{1} \left( l \right)}} = \left\{ {\begin{array}{*{20}l} {\delta \left( i \right) \cdot \frac{{N_{{{\text{pop}}}} }}{{\sum \delta }} \times \sigma _{4} ,} \hfill & {\sigma _{3} < l} \hfill \\ 1 \hfill & {\sigma _{3} > l} \hfill \\ \end{array} } \right.$$
(5)

where \(\delta\) represents the hunger of each population; \(N_{{{\text{pop}}}}\) is the number of populations;

$$\overrightarrow {{W_{2} \left( l \right)}} = \left( {1 - \exp \left( { - \left| {\delta \left( i \right) - \sum \delta } \right|} \right)} \right) \times \sigma _{5} \times 2$$
(6)

where \(\sigma _{3} ,\,\sigma _{4} ,\;{\text{and}}\;\sigma _{5}\) denote the random numbers \(\left( {0 \le \sigma _{3} ,\,\sigma _{4} ,\sigma _{5} \le 1} \right)\). \(\delta\) can be provided by the following function:

$$\delta \left( i \right) = \left\{ {\begin{array}{*{20}l} {0,} \hfill & {{\text{fitness}}_{i} = = {\text{fitness}}_{{{\text{best}}}} } \hfill \\ {\delta \left( i \right) + \delta _{{{\text{new}}}} ,} \hfill & {{\text{fitness}}_{i} ! = {\text{fitness}}_{{{\text{best}}}} } \hfill \\ \end{array} } \right.$$
(7)

where \({\text{fitness}}_{i}\) is the fitness of each individual at the ith iteration; \({\text{fitness}}_{{{\text{best}}}}\) is the best fitness obtained at the current iteration; \(\delta _{{{\text{new}}}}\) is a new hunger, and it is calculated based on Eqs. 8 and 9:

$$\delta _{{{\text{temp}}}} = \frac{{{\text{fitness}}_{i} - {\text{fitness}}_{{{\text{best}}}} }}{{{\text{fitness}}_{{{\text{worst}}}} - {\text{fitness}}_{{{\text{best}}}} }} \times \sigma _{6} \times 2 \times \left( {{\text{bound}}_{{{\text{upper}}}} - {\text{bound}}_{{{\text{lower}}}} } \right)$$
(8)
$$\delta _{{new}} = \left\{ {\begin{array}{*{20}l} {\delta _{{{\text{largest}}}} \times \left( {1 + r} \right),} \hfill & {\delta _{{{\text{temp}}}} < \delta _{{{\text{largest}}}} } \hfill \\ {\delta _{{{\text{temp}}}} ,} \hfill & {\delta _{{{\text{temp}}}} \ge \delta _{{{\text{largest}}}} } \hfill \\ \end{array} } \right.$$
(9)

where \({\text{bound}}_{{{\text{upper}}}}\) and \({\text{bound}}_{{{\text{lower}}}}\) stand for the feature space bounds; \(\sigma _{6}\) is a random number \(\left( {0 \le \sigma _{6} \le 1} \right)\).

The pseudo-code of the HGS algorithm is shown in Figure 2.

Figure 2
figure 2

[modified from Yang et al. (2021)]

Pseudo-code of the HGS algorithm

Hybrid HGS–ANN Model

Like other meta-heuristic algorithms, the primary role of the HGS algorithm in this study is the optimization of weights of the ANN model for predicting BIGV. To do this, an initial ANN model was developed first. Next, the initial weights were transferred to the optimization procedure of the HGS algorithm to re-calculate and update the weights to the ANN model, called the HGS–ANN model. During optimizing the HGS–ANN model, the stopping conditions, such as RMSE and the maximum number of iterations, were used to evaluate the fitness of the proposed HGS–ANN model. If the accuracy of the HGS–ANN model is satisfied (i.e., with the lowest RMSE), the optimization process will stop, and the optimal HGS–ANN model will be selected for predicting BIGV. Otherwise, the optimization procedure will continuously be searching, calculating, and updating the weights of the ANN model in the subsequent iterations. Once the optimization procedure was repeated with the maximum of iterations, the lowest RMSE was selected, and the corresponding model will be assigned as the best model for predicting BIGV. The framework of the HGS–ANN model for predicting BIGV is proposed in Figure 3.

Figure 3
figure 3

Introduction of the HGS–ANN model for predicting BIGV

Performance Metrics for Model Evaluation

As is known, model evaluation is mandatory after the predictive models were developed, aiming to not only understanding the performance of the models (e.g., over-fitting or under-fitting) but also have a comprehensive assessment and comparison of individual models in terms of BIGV prediction. Herein, four performance metrics were used to evaluate the predictive models, including MAE, RMSE, R2, and MAPE. Whereas the MAE and RMSE metrics show the models' errors, the MAPE metric shows the percentage of the error, and R2 metric helps to understand the suitability of the dataset used for each model. They are calculated using the following equations:

$${\text{MAE}} = \frac{1}{{n_{{{\text{blast}}}} }}\sum\limits_{{i = 1}}^{{n_{{{\text{blast}}}} }} {\left| {y_{{i\_{\text{BIGV}}}} - \hat{y}_{{i\_{\text{BIGV}}}} } \right|}$$
(10)
$${\text{RMSE}} = \sqrt {\frac{1}{{n_{{{\text{blast}}}} }}\sum\limits_{{i = 1}}^{{n_{{{\text{blast}}}} }} {\left( {y_{{i\_{\text{BIGV}}}} - \hat{y}_{{i\_{\text{BIGV}}}} } \right)^{2} } }$$
(11)
$$R^{2} = 1 - \frac{{\sum\limits_{{i = 1}}^{{n_{{{\text{blast}}}} }} {\left( {y_{{i\_{\text{BIGV}}}} - \hat{y}_{{i\_{\text{BIGV}}}} } \right)^{2} } }}{{\sum\limits_{{i = 1}}^{{n_{{{\text{blast}}}} }} {\left( {y_{{i\_{\text{BIGV}}}} - \bar{y}_{{{\text{BIGV}}}} } \right)^{2} } }}$$
(12)
$${\text{VAF}} = \left( {1 - \frac{{\text{var} \,\left( {y_{{i\_{\text{BIGV}}}} - \hat{y}_{{i\_{\text{BIGV}}}} } \right)}}{{\text{var} \,\left( {y_{{i\_{\text{BIGV}}}} } \right)}}} \right) \times 100$$
(13)

where \(n_{{{\text{blast}}}}\) is the number of blasting events used in the evaluation; \(y_{{i\_{\text{BIGV}}}}\) is the measured BIGV value at ith blasting event; \(\hat{y}_{{i\_{\text{BIGV}}}}\) stands for the predicted BIGV value at ith blasting event; \(\bar{y}_{{{\text{BIGV}}}}\) denotes the mean of measured BIGV values.

Data and Statistical Analysis

To realize this study, 252 blasting operations were performed at an open-pit coal mine located in the Northern Vietnam (Fig. 4). Accordingly, BIGV values were monitored at the sensitivity locations, and the corresponding parameters of blasts were collected. It is a fact that this mine applied the non-electric millisecond delay blasting method to reduce BIGV (Guan et al., 2019; Nateghi et al., 2009). Nevertheless, several recorded results indicated that the BIGVs measured are high, and they may damage the surrounding environment, especially the benches and slopes of the surrounding open-pit mines, as shown in Figure 4. Thus, monitoring and predicting BIGV in this mine is necessary to control and mitigate the adverse effects of BIGV.

Figure 4
figure 4

Location and an aerial view of the study area

Before measuring BIGVs, blast patterns of 252 blasting events were collected first with the blasting parameters, such as hole length (HL), maximum explosive charge per delay (W), the number of blasting groups (NG), time delay for each group (TL), burden (B), powder factor (PF), spacing (S), and stemming (T). For each blasting event, the position(s) of the seismograph was/were determined, and a GPS receiver was then used to measure the distance from the blasting face to the position(s) of the seismograph (abbreviated as D). Finally, the blasts were detonated, and BIGVs were recoded by the Micromate devices (seismographs). The statistical analysis of the dataset collected is shown in Table 2. In addition, the visualization of this dataset is shown in Figure 5.

Table 2 Statistical analysis of the dataset used
Figure 5
figure 5

Visualizations of the dataset used: (a) histograms; (b) box plots; (c) variable correlations

In Figure 5, the details of the dataset used were illustrated. Accordingly, the distribution of the variables is illustrated by the histogram (Fig. 5a), the statistics and outliers of the dataset are illustrated by box plots (Fig. 5b), and the correlation between independent variables (inputs) and dependent variable (output, i.e., BIGV) is illustrated in Figure 5c. The correlation results show that all the input variables in this dataset should be used to evaluate their effects on BIGV since its correlation is not high (< 0.75). Therefore, we used the whole dataset with nine input variables to predict BIGV based on the introduced models, HGS–ANN, FFA–ANN, GOA–ANN, and PSO–ANN.

Results and Discussion

As introduced earlier, this work aims at proposing the novel HGS–ANN model for predicting BIGV in mine blasting. Three other hybrid models, namely PSO–ANN, FFA–ANN, and GOA–ANN, were also taken into consideration to predict BIGV and compared with the proposed HGS–ANN model.

Prior to developing these models, the blasting dataset was preprocessed using the MinMax scaling method with the interval [− 1,1] to normalize the dataset. Subsequently, the dataset was divided into two partitions that can be managed, developed and tested separately (70% for training and 30% for testing). Finally, the proposed HGS–ANN framework was applied, as shown in Figure 3. Similar approaches were applied for the PSO–ANN, FFA–ANN, and GOA–ANN models. Before training the ANN model by the optimization algorithms (i.e., HGS, PSO, FFA, and GOA), an initial ANN model and the algorithms’ parameters were established sufficiency. For the initial ANN model, a network topology with two hidden layers (e.g., ten neurons in the first layer, five neurons in the second layer) was selected to predict BIGV through the trial-and-error procedure, and the initial weights were computed. The parameters of the HGS, PSO, FFA, and GOA were set up as follows:

  • Numbers of hungers, swarms, fireflies, and grasshoppers (populations): 50, 100, 150, 200, 250, 300, 350, 400, 450, 500.

  • Maximum number of iterations: 1000.

  • PSO’s parameters: \(c_{1} = c_{2} = 1.2\); \(w_{{\min }} = 0.4\); \(w_{{\max }} = 0.9\).

  • GOA’s parameters: \(c_{{\min }} = 0.00004\); \(c_{{\max }} = 1\).

  • HGS’s parameters: \(L = 0.03\); \({\text{LH}} = 1000\).

  • FFA’s parameters: \(\gamma = 1\); \(\beta = 1\); \(\alpha = 0.2\); \(\alpha _{{{\text{damp}}}} = 0.99\); \(\delta = 0.05\); \(m_{{{\text{exponent}}}} = 2\).

Finally, the meta-heuristic algorithms were applied to optimize the initial weights of the initial ANN model. RMSE was used as the predictive models' objective function to evaluate the training error of the swarm-based ANN models during 1000 iterations. The results are shown in Figure 6.

Figure 6
figure 6figure 6

Performance of the optimal predictive models under various population sizes: (a) HGS–ANN; (b) FFA–ANN; (c) GOA–ANN; (d) PSO–ANN

As shown in Figure 6, the best hybrid models (optimal models) for predicting BIGV were defined with the following parameters:

  • HGS–ANN model: 400 populations, 684 iterations;

  • FFA–ANN model: 450 populations; 727 iterations;

  • GOA–ANN model: 300 populations; 992 iterations;

  • PSO–ANN model: 300 populations; 969 iterations.

Once the models were optimized, the testing dataset was imported to calculate their performance as unseen data in practice. The statistical analysis of the performances is given in Table 3. As indicated in the statistical indexes, it is possible to see that all four models are acceptable for predicting BIGV. However, the HGS–ANN and GOA–ANN models provided better performances than the other models (i.e., FFA–ANN and PSO–ANN). Nevertheless, it is difficult to evaluate which model is better among the HGS–ANN and GOA–ANN models. However, the training phase provides a better performance for the GOA–ANN model, and the testing phase provides a better performance for the HGS–ANN model. Therefore, we used the ranking method to sort the models based on their performances and rankings, as given in Table 4.

Table 3 Statistical criteria of the hybrid models developed
Table 4 Ranking of the hybrid models developed based on statistical criteria

From the results in Table 4, it is easy to recognize which model is the best based on the ranking of the models. Accordingly, the HGS–ANN model yielded the highest ranking, with a total ranking of 28. Meanwhile, the GOA–ANN model only provided a total ranking of 26. The other models, i.e., FFA–ANN and PSO–ANN, provided the same ranking with a total ranking of 13. Furthermore, based on the parameters of the meta-heuristic algorithms used and the training time of the models in Figure 6, it can be seen that the HGS is an outstanding algorithm and the HGS–ANN model is the most superior model in the training dataset. Whereas the HGS’s parameters are most of the number of populations and maximum iterations (except L and LH are coefficients), the other algorithms required additional parameters during training. Moreover, considering the training time of the hybrid models (Fig. 6), the PSO–ANN model provided the lowest training time with 1944.472 s. Following are the HGS–ANN model (i.e., 2121.498 s), GOA–ANN model (i.e., 6350.216 s), and FFA–ANN model (i.e., 7033.281 s). However, the training error of the HGS–ANN model is the lowest, and the training time is slightly higher than the PSO–ANN model. Overall, the HGS–ANN model is the most superior model in terms of development and modeling. For further assessment of the models, the relative error (RE), correlation of observed and predicted values, and a comparison of them are shown in Figure 7.

Figure 7
figure 7figure 7

Evaluation of the BIGV prediction models based on the relative error, performance prediction, and linear relationship: (a) HGS–ANN; (b) FFA–ANN; (c) GOA–ANN; (d) PSO–ANN

From Figure 7, it can be seen that the accuracy of the predictions obtained by individual models is high, and the convergence on the regression lines is very good. Nevertheless, taking a closer look at Figure 7, it is taken into consideration that the accuracy and convergence on the HGS–ANN model are superior to the other models. In particular, considering the models' RE graphics, we can see a peak point on the graphics at sample #32. Taking a closer look at this peak point in the developed models, it is no doubt that the RE in the HGS–ANN model is the lowest (~ − 550%), whereas it is ~  − 1300% for the GOA–ANN model, ~  − 800% for the PSO–ANN model, and ~  − 7100% for the FFA–ANN model. Hence, it can claim that the HGS–ANN model provided an excellent performance in this study for predicting BIGV. In other words, the HGS algorithm is more robust than the other algorithms (i.e., GOA, PSO, FFA) when combined with the ANN model for BIGV estimation in this work.

Sensitivity Analysis

In order to determine the robustness of the proposed HGS–ANN model, a sensitivity analysis was adopted to examine the extent to which the results are affected by changes in values of the input variables. In other words, this step aims to identify the important variables on which the results depend most. The sensitivity criteria (Dimopoulos et al., 1995) were calculated and evaluated, as shown in Figure 8. As illustrated in this figure, B, W, NG, S, D, and PF are the variables that have high effects on the BIGV. In contrast, the HL and T variables have low influences on BIGV, especially the T variable.

Figure 8
figure 8

Sensitivity analysis results of the variables used through the proposed HGS–ANN model

Conclusion

Blasting is an indispensable method in open-pit mines; however, its negative impacts on the environment are inevitable, especially BIGV. This study proposed a novel AI model with the accuracy improved for predicting BIGV, namely the HGS–ANN model. The optimization role of the HGS algorithm was thoroughly interpreted in this study when combined with the ANN model for predicting BIGV, and it is a robust optimization algorithm for engineering problems. The HGS–ANN model was proved as a robust soft computing model for predicting BIGV with the outstanding obtained results compared to the benchmark hybrid models (i.e., GOA–ANN, FFA–ANN, and PSO–ANN). As a recommendation, this model should be used in practical engineering to control and mitigate the adverse effects of BIGV in mine blasting.