Introduction

Risk is defined as the possibility of suffering damage which is considered to be a function of the probability of event occurrence and the resulting consequence in the engineering domain (Baecher and Christian 2005). Slope failures are such major events resulting in significant damage to human lives and the economy. For India, an annual landslide damage of $1.35 billion is reported for the total 89,000 km of roads in northern India's landslide-prone and fragile mountainous territories, along with significant casualties (Mathur 1982). The most important step for managing and preventing these risks is to conduct a probabilistic risk assessment of the rock slopes. Reliability technique is an important tool to estimate these risks using modern probabilistic and statistical methods. Different reliability methods such as Point Estimate Methods (PEMs), First Order Reliability Method (FORM), Second-Order Reliability Method (SORM) and Monte-Carlo Simulations (MCSs) have proven their accuracy and practical utility for a wide range of rock slope reliability problems (Low and Einstein 1992; Düzgün et al. 1995; Park and West 2001; Duzgun and Bhasin 2009; Ahmadabadi and Poisel 2016). Table 1 shows the studies and classification of the reliability studies on rock slopes based on the major types of failure mechanisms and analysis methods. It can be observed that the majority of initial studies were related to the reliability analysis of structurally controlled rock slopes (type 3 problems) and analysis were performed via the coupling of analytically formulated performance functions with methods like FORM or MCSs. However, due to the limitations of analytical formulations like inability to incorporate the advanced constitutive models, spatial variability, complex in-situ stresses and boundary conditions (Jiang et al. 2015; Qian et al. 2017; Kumar et al. 2018; Ji et al. 2020; Abdulai and Sharifzadeh 2021), the focus has currently shifted towards the coupling of advanced numerical tools like finite element, finite difference, etc., with reliability methods (Huang et al. 2017; Pandit et al. 2018; Tiwari and Latha 2019; Zhao et al. 2020; Zhang et al. 2021). However, due to the unavailability of explicit performance functions in many reliability problems, traditional reliability methods like FORM/SORM cannot be applied as they require the derivative of the performance function. Further, the simulation-based methods like MCSs are computationally inefficient for large rock slopes as they require repeated evaluation of the numerical models for the random sets of input properties.

Table 1 A summary of the studies on the reliability analysis of rock slopes

For these cases, Response Surface Methods (RSMs) can be employed which develop surrogate performance functions based on the limited numerical simulations. This provides a computationally efficient and accurate en route for the risk assessment and remedial measures design of rock slopes utilizing reliability methods (Pandit et al. 2018, 2019; Tiwari and Latha 2019). Studies on the consistent developments to enhance the accuracy and efficiency of the RSMs, and their practical applicability for different engineering problems are available in the literature (Hill and Hunter 1966; Mead and Pike 1975; Myers et al. 1989; Myers 1999; Bucher and Most 2008; Li et al. 2016; Zhu et al. 2019b). Table 2 summarizes the studies related to the application of RSMs for the reliability-based stability analysis of rock slopes. It can be observed that the studies on the development and application of RSMs for the reliability-based stability analyses of rock slopes are limited. Most of these studies employed the polynomial based RSM due to its mathematical simplicity despite many disadvantages related to its accuracy and efficiency (Shamekhi and Tannant et al. 2015; Pandit et al. 2016; Dadashzadeh et al. 2017; Basahel and Mitri 2019; Roy et al. 2019). From the extensive literature review, major gaps are identified as: (i) lack of any guideline for the selection of the most appropriate RSM among the available RSMs and their required sampling points for the reliability-based stability analysis of different types of rock slope problems, (ii) lack of studies demonstrating the applicability and advantages of advanced RSMs for rock slope problems as most studies employed polynomial RSMs due to their mathematical simplicity, and (iii) no studies on the applicability of RSMs for the reliability-based stability analysis of rock slopes considering the spatial variation of properties (type 2 problem).

Table 2 A summary of the studies on the application of response surfaces for the rock slope reliability problems

To address these issues, this study demonstrates and compares the applicability and advantages of traditional and advanced RSMs for the reliability-based stability analysis of different types of rock slope problems. These problem types are identified based on an extensive literature review: type (1) stress-controlled failures neglecting spatial variability, type (2) stress-controlled failures considering spatial variability, and type (3) structurally controlled failures neglecting spatial variability. Performances of sixteen traditional and advanced RSMs are evaluated for the reliability-based stability analysis of three real rock slopes corresponding to the above-mentioned problem types to identify the dependence of the accuracy and efficiency of RSMs on the problem types. Real and large natural rock slopes with complex geometries and geologies were used for the analyses to get a more realistic idea of the applicability of the RSMs. Based on the extensive numerical simulations (\(\sim 5\times {10}^{3}\)) and data analyses, guidelines are suggested for the selection of the most appropriate RSM and the number of the required sampling points for the stability analysis of different types of rock slope problems. This paper is organized into five sections. “Response surface methods (RSMs)” provides the theoretical details of the considered RSMs. “Evaluation of RSMs performance” provides the methodology adopted to compare the performances of the considered RSMs. “Application of RSMs to major slope reliability problems” describes the major rock slope reliability problems identified through literature review and analyses of the case studies to demonstrate the applicability of RSMs along with the suggestive guidelines. “Conclusions” provides the important conclusions and limitations of the study.

Response surface methods (RSMs)

RSMs basically consist of developing an explicit surrogate functional relationship between response parameters (e.g., Factor of Safety (FOS) for the rock slopes) with input random variables (e.g., rock properties) based on the limited numerical experiments. The basic mathematical formulation of the RSMs is given below (Bucher and Most 2008).

Let the required response parameter for the rock slopes, i.e., \(G({\varvec{Z}})\) (\(FOS({\varvec{Z}})\) for this study), depend upon a vector \({\varvec{Z}}\) of \(d\) input random variables (rock properties), i.e. \({\varvec{Z}} ={({z}_{1}, {z}_{2}, {z}_{3},\dots \dots {z}_{d})}^{T}\). In many rock slope problems, it is impossible to derive an explicit relation between the response parameter and input rock properties. RSMs provide an alternative relationship between response (\(FOS({\varvec{Z}})\)) and input variables (\({\varvec{Z}}\)) through a flexible function \(\widehat{G}({\varvec{Z}})\) and can be approximately represented as following.

$$FOS\left( {\varvec{Z}} \right) \approx \hat{G}\left( {\theta_{1} ,{ }\theta_{2} ,{ }\theta_{3} , \ldots \ldots ,\theta_{m} ;z_{1} ,{ }z_{2} ,{ }z_{3} , \ldots \ldots ,z_{d} } \right) \approx \hat{G}\left( {{\varvec{\theta}};{\varvec{Z}}} \right)$$
(1)

where, \({\varvec{\theta}}={\theta }_{1}, {\theta }_{2}, {\theta }_{3},\dots \dots ,{\theta }_{m}\) is the vector of the parameters of the approximating function \(\widehat{G}({\varvec{Z}})\) and is evaluated through numerical experiments carried out for limited input realizations. This section will provide the theoretical details of the considered RSMs and methodology to determine parameters for these RSMs. In the following sections, \(\widehat{G}\left({\varvec{\theta}};{\varvec{Z}}\right)\) is represented as \(\widehat{G}\left({\varvec{Z}}\right)\) for simplicity. The Latin Hypercube Sampling (LHS) based space-filling design is adopted for generating random input vector realizations (i.e., sampling points) from the distributions of input parameters (Montgomery 2001).

Polynomial model with cross terms (Q-I)

This is the most commonly used RSM in rock engineering due to its computational simplicity and its final form which is a closed-form algebraic expression (Krishnamurthy 2003; Li et al. 2016). The efficiency of this RSM depends on the order and cross-terms of the polynomial function, which are used to represent the whole input parameter domain. Figure 1a shows an example of the polynomial RSM. A general quadratic (second-order) polynomial model can be defined as following.

$$\hat{G}\left( {\varvec{Z}} \right) = a_{0} + \mathop \sum \limits_{i = 1}^{d} a_{i} z_{i} + \mathop \sum \limits_{i = 1}^{d} \mathop \sum \limits_{j = 1}^{d} a_{ij} z_{i} z_{j}$$
(2)

where, \({z}_{i}\) and \({z}_{j}\) are the design variables and \({a}_{0}\), \({a}_{i}\) and \({a}_{ij}\) are the unknown coefficients. The unknown coefficients (\(m\)= \(\left(d+1\right)\left(d+2\right)/2\)) can be determined uniquely with \(m\) numerical experiments corresponding to \(m\) sampling points. For a larger number of sampling points \(n\)(\(>m\)), the system of equations becomes overdetermined. In that case, the method of least squares that minimizes the sum of the squares of the residual error (i.e., the difference between the actual/true and the approximated response at a sampling point) can be used in the coefficient estimation process.

Fig. 1
figure 1

Illustration of performance function approximation concept for the Response Surface Methods (RSMs) a Polynomial b Radial Basis Function (RBF) c Moving least square (MLS) d Kriging e Support Vector Machine (SVM), and f Least Square SVM (LS-SVM)

Polynomial model without cross terms (Q-II)

This is a type of the previous RSM in which a quadratic polynomial without cross (interaction) terms is used. The expression for this RSM is defined as following.

$$\hat{G}\left( {\varvec{Z}} \right) = a_{0} + \mathop \sum \limits_{i = 1}^{d} a_{i} z_{i} + \mathop \sum \limits_{i = 1}^{d} b_{i} z_{i}^{2}$$
(3)

where, \({a}_{0}\), \({a}_{i}\) and \({b}_{i}\) are the \(m=\left(2d+1\right)\) unknown coefficients which can be determined as explained for the previous RSM (Q-I).

Classical radial basis function (RBF)

The RBF-based RSMs are developed to efficiently approximate the multi-variable performance functions which are highly non-linear in their entire input domain. The classical RBF methods approximate the numerical model output \(G({\varvec{Z}})\) by a weighted linear combination of radially symmetric non-linear functions which depend on the Euclidean norm between observation point \({\varvec{Z}}\) and \({i}^{th}\) sampling point input vector \({{\varvec{Z}}}_{i}\), i.e., \({\| {\varvec{Z}}-{{\varvec{Z}}}_{i}\| }_{2}\). Figure 1b illustrates the basic concept of the RBF RSM. The functional form for the RBF RSM can be written as given below (Krishnamurthy 2003).

$$\hat{G}\left( Z \right) = \sum\limits_{i = 1}^{n} {\lambda_{i} } \phi \left( {\left\| {Z - Z_{i} } \right\|_{2} } \right)$$
(4)

where, \(\phi\) is the radial basis function, \(n\) is the number of sampling points obtained using Latin Hypercube Sampling (LHS), and \({\lambda }_{i}\) are the interpolation/weightage coefficients to be determined. The unknown coefficients \({\varvec{\lambda}}\) can be determined by solving the system of linear equations as given below.

$$\left[ {\varvec{\lambda}} \right]_{n \times 1} = \left[ {\varvec{A}} \right]_{n \times n}^{{{-}1}} \times \left[ {\varvec{Y}} \right]_{n \times 1}$$
(5)

where, \({\varvec{Y}}={[G\left({{\varvec{Z}}}_{1}\right) G\left({{\varvec{Z}}}_{2}\right) ...\boldsymbol{ }\boldsymbol{ }G\left({{\varvec{Z}}}_{n}\right)]}^{{\varvec{T}}}\) is the known vector of numerical model outputs estimated by solving numerical models at \(n\) sampling points and matrix \({\varvec{A}}\) can be obtained as given below.

$${\varvec{A}}_{ij} \, = \,\phi \left( {{\varvec{Z}}_{i} - {\varvec{Z}}_{j2} } \right); i,j = 1,2, \ldots n$$
(6)

Popular RBFs include Linear (L–RBF), Cubic (C–RBF), Gaussian (G–RBF) and Multiquadric (MQ-RBF) basis functions, i.e., \(\phi\) as shown in Appendix 1.

The approximation accuracy of these RBFs models depends on a parameter, i.e., \(c\), known as the shape or shift parameter whose optimal estimation can be made with the leave-one-out cross-validation (\(LOOCV\)) technique (Rippa 1999). The \(LOOCV\) uses \(n-1\) sampling points to const`ruct the response surface and the remaining sample point as cross-validation. This procedure of leaving out one sampling point is repeated for all \(n\) points and squares of error are added cumulatively with varying shape parameter (\(c\)) values. Thus, the value of \(c\) for which \(LOOCV\) is minimum is adopted for RBF response surface construction.

$$LOOCV = \mathop \sum \limits_{i = 1}^{n} \left\{ {\hat{G}\left( {{\varvec{Z}}_{i} } \right) - G\left( {{\varvec{Z}}_{i} } \right)} \right\}^{2}$$
(7)

where, \(\widehat{G}({{\varvec{Z}}}_{i})\) is approximated output of the left-out point from the response surface constructed from \(n-1\) sampling points and \(G\left({{\varvec{Z}}}_{i}\right)\) is the actual output obtained from the numerical model at the left-out point.

Compactly supported RBF

Compactly supported RBFs are developed to increase the computational efficiency by converting the globally supported system of classical RBFs into a local one. This is achieved by defining the basis functions in such a way that they only give weightage to sample points that are within a certain distance (i.e., the radius of compact support) from the observation point. This makes the resultant matrix \({\varvec{A}}\) (Eq. 6) sparse and banded. Thus, compactly supported RBF approximations are more feasible for solving large-scale problems.

Apart from the basis function, definitions of all the other parameters in Eq. 4 are valid for compactly supported RBFs. In this study, the following two compactly supported RBFs are considered (Krishnamurthy 2003).

Compact-I (C-I-RBF):

$$\phi \left( r \right) = \left\{ {\begin{array}{*{20}l} {\left( {1{-}t} \right)^{5} \left( {8 + 40t + 48t^{2} + 25t^{3} + 5t^{4} } \right);} & {0 \le r \le r_{0} } \\ {0;} & {r > r_{0} } \\ \end{array} } \right.$$
(8)

Compact-II (C-II-RBF):

$$\phi \left( r \right) = \left\{ {\begin{array}{*{20}l} {\left( {1{-}t} \right)^{6} \left( {6 + 36t + 82t^{2} + 72t^{3} + 30t^{4} + 5t^{5} } \right);~} & {0 \le r \le r_{0} } \\ {0;} & {r > r_{0} } \\ \end{array} } \right.$$
(9)

where, \(t = {\raise0.7ex\hbox{$r$} \!\mathord{\left/ {\vphantom {r {r_{0} }}}\right.\kern-\nulldelimiterspace} \!\lower0.7ex\hbox{${r_{0} }$}}\); \(r_{0}\) is the compact support radius of the domain which is a free parameter chosen by minimizing the \(LOOCV\) error.

Augmented RBF

The compactly supported RBFs require a large number of sampling points to reproduce the simple polynomials, i.e., constant, linear and quadratic (Krishnamurthy 2003). Augmented RBFs are developed to approximate the simple polynomials efficiently by augmenting polynomial terms to the basis functions of compactly supported RBFs as given below (Krishnamurthy 2003).

$$\hat{G}\left( {\varvec{Z}} \right) = \mathop \sum \limits_{i = 1}^{n} \lambda_{i} \phi \left( {{\varvec{Z}} - {\varvec{Z}}_{i2} } \right) + \mathop \sum \limits_{j = 1}^{m} c_{j} P_{j} \left( {\varvec{Z}} \right)$$
(10)

where, \({P}_{j}({\varvec{Z}})\) are the monomial terms of polynomial \(P({\varvec{Z}})\) and \({c}_{j}\) are the \(m\) constant coefficients introduced due to augmentation of the polynomial. For quadratic polynomial without cross term, value of \(m\) is \(2d+1\) (i.e.,\({P}_{j}({\varvec{Z}})={[1 {z}_{1} {z}_{2} ... {z}_{d} {z}_{1}^{2}{ z}_{2}^{2} ... {z}_{d}^{2}]}_{1\times m}\)). Additional \(m\) coefficients are estimated by employing orthogonality conditions as given below.

$$\mathop \sum \limits_{i = 1}^{n} \lambda_{i} P_{j} \left( {Z_{i} } \right) = 0 ; {\text{for}} j = 1,2,...,m$$
(11)

By combining Eq. 10 and Eq. 11, unknown coefficients \({\varvec{\lambda}}\) and \({\varvec{c}}\) can be estimated by solving the following system of equations.

$$\left[ {\begin{array}{*{20}l} {{\varvec{\lambda}}_{n \times 1} } \\ {{\varvec{c}}_{m \times 1} } \\ \end{array} } \right] = \left[ {\begin{array}{*{20}c} {{\varvec{A}}_{n \times n} } & {{\varvec{C}}_{n \times m} } \\ {{\varvec{C}}_{m \times n}^{T} } & {0_{m \times m} } \\ \end{array} } \right]^{{{-}1}} \left[ {\begin{array}{*{20}c} {{\varvec{Y}}_{n \times 1} } \\ {0_{m \times 1} } \\ \end{array} } \right]$$
(12)

where, \({{\varvec{C}}}_{ij}={P}_{j}\left({{\varvec{Z}}}_{i}\right) \mathrm{for} i=\mathrm{1,2},...n; j=\mathrm{1,2},...m\) and\({\varvec{c}}={[{c}_{1} {c}_{2} ... {c}_{m}]}^{T}\). Matrices\({\varvec{\lambda}}\),\({\varvec{A}}\), and \({\varvec{Y}}\) can be formed similarly as mentioned in  “Classical Radial Basis Function (RBF)”. In this study, the two compactly supported RBFs defined in “Compactly supported RBF” (Eq. 8 and Eq. 9) are considered.

Moving least square (MLS)

The MLS is a local RSM with high approximation accuracy compared to the conventional least square method for the polynomial RSMs (Krishnamurthy 2003). Figure 1c shows the basic concept of the MLS RSM approximation. The MLS RSM approximates the polynomial surface from a set of sampling points within the influence domain of/around the observation point (\({\varvec{Z}}\)) through the least square measures. The sampling points are weighted according to their distance from the observation point. In MLS-RSM, an explicit polynomial approximation of the implicit performance function in terms of the random variables is made as given below (Krishnamurthy 2003).

$$\hat{G}\left( {\varvec{Z}} \right) = {\varvec{p}}\left( {\varvec{Z}} \right){\varvec{a}}\left( {\varvec{Z}} \right)$$
(13)

where, \({\varvec{p}}({\varvec{Z}})={[1 {z}_{1} {z}_{2} ... {z}_{d} {z}_{1}^{2} {z}_{2}^{2} ... {z}_{d}^{2}]}_{1\times m}\) is a quadratic polynomial basis of function (\(m=2d+1\)). \({\varvec{a}}({\varvec{Z}})\) is a set of unknown coefficients which is expressed as a function of the observation point \({\varvec{Z}}\) to consider the variation of the coefficients for every new observation point.\({\varvec{a}}({\varvec{Z}})\) can be determined as given below.

$${\varvec{a}}\left( {\varvec{Z}} \right) = {\varvec{A}}_{m \times m}^{{{-}1}} {\varvec{B}}_{m \times n} {\varvec{Y}}_{n \times 1}$$
(14)

where, \({\varvec{Y}}={[G\left({{\varvec{Z}}}_{1}\right) G\left({{\varvec{Z}}}_{2}\right)...\boldsymbol{ }G\left({{\varvec{Z}}}_{n}\right)]}^{{\varvec{T}}}\) is the vector of known performance function values obtained from the numerical modeling at \(n\) sampling points and matrix \({\varvec{A}}\) and \({\varvec{B}}\) are defined as given below.

$${\varvec{A}}\left( {\varvec{Z}} \right) = {\varvec{P}}_{m \times n}^{T} {\varvec{W}}_{n \times n} {\varvec{P}}_{n \times m} ; {\varvec{B}}\left( {\varvec{Z}} \right) = {\varvec{P}}_{m \times n}^{T} {\varvec{W}}_{n \times n}$$
(15)

with

$${\varvec{P}} = \left[ {\begin{array}{*{20}c} {{\varvec{p}}\left( {{\varvec{Z}}_{1} } \right)} \\ {{\varvec{p}}\left( {{\varvec{Z}}_{2} } \right)} \\ {\begin{array}{*{20}c} {...} \\ {{\varvec{p}}\left( {{\varvec{Z}}_{n} } \right)} \\ \end{array} } \\ \end{array} } \right]_{n \times m};\, {\varvec{W}} = \left[ {\begin{array}{*{20}c} {w_{1} \left( {\varvec{Z}} \right)} & 0 & 0 \\ 0 & {...} & 0 \\ 0 & 0 & {w_{n} \left( {\varvec{Z}} \right)} \\ \end{array} } \right]_{n \times n}$$
(16)

where, \({w}_{i}({\varvec{Z}})\) is the weighting or smoothing function representing the weightage/contribution of different sampling points to the approximated \(\widehat{G}\left({\varvec{Z}}\right)\) for observation point \({\varvec{Z}}\). Among the available weighting functions, the spline weighting function (\({C}^{1}\) continuous) with compact support is considered in this study which can be expressed as following.

$$w_{i} \left( \varvec{Z} \right) = \left\{ {\begin{array}{*{20}l} {1{-}6r_{i} ^{2} + 8r_{i} ^{3} {-}3r_{i} ^{4} ;} & {r \le 1} \\ {0;~} & {~r > 1} \\ \end{array} } \right.$$
(17)

where, \(r = {\raise0.7ex\hbox{${\left\| {{\varvec{Z}}{-}{\varvec{Z}}_{i} } \right\|_{2} }$} \!\mathord{\left/ {\vphantom {{\left\| {{\varvec{Z}}{-}{\varvec{Z}}_{i} } \right\|_{2} } {l_{i} }}}\right.\kern-\nulldelimiterspace} \!\lower0.7ex\hbox{${l_{i} }$}}\); \({l}_{i}\) = size of the domain of influence which is chosen as twice the distance between the \({(1+2d)}^{th}\) sample point and the observation point \({\varvec{Z}}\); \(d\) is the number of random variables and \(n\) is the number of sampling points.

Kriging


Kriging is a statistical RSM that assumes the deterministic response as a stochastic process. Figure 1d shows the concept of the kriging approximation. This method approximates the response at an observation point by giving more weightage to nearby sampling points utilizing spatial correlation among the sample points and minimizing the variance of the residual errors. Kriging model postulates the approximation of performance function in the form as given below (Krishnamurthy 2003).

$$\hat{G}\left( {\varvec{Z}} \right) = \beta + \left[ {{\varvec{r}}\left( {\varvec{Z}} \right)} \right]_{1 \times n}^{T} {\varvec{R}}_{n \times n}^{{{-}1}} \left( {{\varvec{Y}}_{n \times 1} {-}\beta {\varvec{f}}_{n \times 1} } \right)$$
(18)

where, \(\beta\) is the unknown coefficient; \({\varvec{Y}}\) is the matrix of \(n\) sample points response values obtained by numerical analysis; \({\varvec{f}}\) is a \(n\times 1\) matrix with all values equal to unity; \({\varvec{r}}({\varvec{Z}})\) is the matrix of correlations between the observation point \({\varvec{Z}}\) and all \(n\) sample points (\({{\varvec{Z}}}_{i}\)) in parameter space, and \({\varvec{R}}\) is the correlation matrix of all \(n\) sample points. The terms in the correlation matrices are computed using a Gaussian correlation function as following.

$${\varvec{r}}\left( {\varvec{Z}} \right) = exp\left[ {{-}\theta \mathop \sum \limits_{t = 1}^{d} z_{t} {-}z_{t}^{i2} } \right]; i = 1,2,..., n$$
(19)
$${\varvec{R}}_{ij} = exp\left[ {{-}\theta \mathop \sum \limits_{t = 1}^{d} z_{t}^{i} {-}z_{t}^{j2} } \right]; i,j = 1,2,..., n$$
(20)

The value of the parameter \(\theta\) is estimated from \(n\) sample response values via maximum likelihood estimation (\(MLE\)) and this \(MLE\) problem can be solved by reducing it to a one-dimensional optimization problem as following.

$$MLE = \min \left[ {\frac{1}{2}\left\{ {n ln\left( {\sigma^{2} } \right) + ln\left| {\varvec{R}} \right|} \right\}} \right]; s.t. 0 \le \theta \le \infty$$
(21)

where, \({\sigma }^{2}\) variance and unknown coefficient \(\beta\) are estimated using generalized least squares as following.

$$\sigma^{2} = \frac{{\left( {{\varvec{Y}}{-}\beta {\varvec{f}}} \right)^{T} {\varvec{R}}^{{{-}1}} \left( {{\varvec{Y}}{-}\beta {\varvec{f}}} \right)}}{n}$$
(22)
$$\beta = \left( {{\varvec{f}}^{T} {\varvec{R}}^{{{-}1}} {\varvec{f}}} \right)^{{{-}1}} {\varvec{f}}^{T} {\varvec{R}}^{{{-}1}} \user2{Y }$$
(23)

Support vector machine (T-SVM)

The SVM formulates the performance function approximation problem as an inequality constraint based convex optimization problem (quadratic programming problem). Figure 1e shows the concept of the SVM for linear approximation. The SVM based RSM sets a tolerance approximation margin (\(\pm \varepsilon\)) and uses only values outside the margin to approximate the performance function. For mapping of non-linearly separable functions, SVM uses kernel functions to transfer input space to high dimensional feature space and then linear separation is carried out. The optimization problem is given below (Zhao 2008).

$$\min \frac{1}{2}w^{2} + C\mathop \sum \limits_{i = 1}^{n} \left( {\xi_{i} + \xi_{i}^{*} } \right)$$
(24)
$$s.t. \left\{ \begin{gathered} Y_{i} {-}w.Z_{i} {-}b \le \varepsilon + \xi_{i}^{*} \hfill \\ w.Z_{i} + b{-}Y_{i} \le \varepsilon + \xi_{i} \,\,\,\,\,\,\,\,\,\,i = 1,2,...,n \hfill \\ \xi_{i} ,\xi_{i}^{*} \ge 0 \hfill \\ \end{gathered} \right.$$
(25)

where, \(C\) is the penalty parameter (\(C>0\)); \(\varepsilon\) is the approximation accuracy; \({\xi }_{i}\) and \({\xi }_{i}^{*}\) are the slack variables; \({\varvec{Y}}\) is the matrix of response values at \(n\) sample points obtained by numerical analysis; \(w\) is the weightage vector and \(b\) is a bias coefficient.

Solving the above optimization problem, the resulting SVM regression model for the non-linear function approximation takes the form as given below.

$$\hat{G}\left( Z \right) = \mathop \sum \limits_{i = 1}^{n} \left( {\alpha_{i} {-}\alpha_{i}^{*} } \right)K\left( {Z,Z_{i} } \right) + b$$
(26)

where, \(b\) can be estimated as following.

$$b = \left\{ {\begin{array}{*{20}c} {Y_{i} {-}\mathop \sum \limits_{j = 1}^{n} \left( {\alpha_{j} {-}\alpha_{j}^{*} } \right)K\left( {Z_{j} ,Z_{i} } \right){-}\varepsilon ;\alpha_{j} \in \left( {0,C} \right)} \\ {Y_{i} {-}\mathop \sum \limits_{j = 1}^{n} \left( {\alpha_{j} {-}\alpha_{j}^{*} } \right)K\left( {Z_{j} ,Z_{i} } \right) + \varepsilon ;\alpha_{j}^{*} \in \left( {0,C} \right)} \\ \end{array} } \right.$$
(27)

where, \({\alpha }_{i}\) and \({\alpha }_{i}^{*}\) are Lagrange multipliers and \(K(.)\) represents the kernel function. In this study, the Gaussian radial kernel (or covariance) function with kernel width \(\sigma\) is considered which is defined as following.

$$K\left( {{\varvec{Z}},{\varvec{Z}}_{i} } \right) = exp\left\{ {{-}\frac{{{\varvec{Z}} - {\varvec{Z}}_{i}^{2} }}{{2\sigma^{2} }}} \right\}$$
(28)

Least square-support vector machine (LS-SVM)

The LS-SVM is an improvement over the SVM method to increase its computational efficiency which involves the equality instead of inequality constraints. The LS-SVM works with the least-squares errors as shown in Fig. 1f instead of nonnegative errors to the loss function as for the SVM. In this way, the solution follows a linear Karush–Kuhn–Tucker system instead of a quadratic programming problem which thus reduces the computation time significantly.

For LS-SVM function, the optimization problem for approximations is formulated as following (Samui et al. 2013).

$$min\frac{1}{2}w^{2} + \frac{1}{2}\gamma \mathop \sum \limits_{i = 1}^{n} e_{i}^{2}$$
(29)
$$s.t. {\varvec{Y}}_{i} = w.{\varvec{Z}}_{i} + b + e_{i} ; \,\,\,\,\,i = 1,2,...,n$$
(30)

where, \({e}_{i}\) is the approximation error term for \({i}^{th}\) sampling point, \(\gamma \ge 0\) is the regularization constant. The resulting LS-SVM response surface for function approximation can be written as following

$$\hat{G}\left( {\varvec{Z}} \right) = \mathop \sum \limits_{i = 1}^{n} \alpha_{i} K\left( {{\varvec{Z}},{\varvec{Z}}_{i} } \right) + b$$
(31)

where, \({\alpha }_{i}\) are Lagrange multipliers and bias coefficient \(b\) can be estimated as following.

$$\left[ {\begin{array}{*{20}c} 0 & {{\varvec{f}}^{T} } \\ {\varvec{f}} & {{{\varvec{\Omega}}} + \gamma^{ - 1} {\varvec{I}}} \\ \end{array} } \right]\left[ {\begin{array}{*{20}c} {\varvec{b}} \\ {\varvec{\alpha}} \\ \end{array} } \right] = \left[ {\begin{array}{*{20}c} 0 \\ {\varvec{Y}} \\ \end{array} } \right]$$
(32)

where, \(\boldsymbol{\alpha }={\left[{\alpha }_{1} {\alpha }_{2}\dots {\alpha }_{n}\right]}^{T}\); \({\varvec{f}}={\left[1 1\dots 1\right]}_{n\times 1}^{T}\) and \({{\varvec{\Omega}}}_{ij}=K({{\varvec{Z}}}_{i}, {{\varvec{Z}}}_{j})\) \(i,j=1, 2, \dots , n\). The Gaussian radial kernel, as in Eq. 28, is considered for the LS-SVM method.

Gaussian progress regression (GPR)

The GPR is a Bayesian framework-based non-parametric RSM. It provides the probability of responses with prediction errors compared to other regression methods. GPR uses the concept of kernel functions-based regression and hence, can approximate the non-linear performance functions with high accuracy and efficiency. In GPR RSM, response \(\widehat{G}({\varvec{Z}})\) for the observations \({\varvec{Z}}\) can be modeled using the posterior distribution based on the Baye's rule as given below (Kang et al. 2015).

$$p\left( {\hat{G}\left( {\varvec{Z}} \right){|}{\varvec{Z}}_{{\varvec{i}}} ,{\varvec{Y}},{\varvec{Z}}} \right)\sim N\left( {{\varvec{\mu}},\varvec{ \sigma }} \right)$$
(33)
$${\varvec{\mu}} = {\varvec{K}}\left( {{\varvec{Z}},\user2{ Z}_{{\varvec{i}}} } \right)\left[ {{\varvec{K}}\left( {{\varvec{Z}}_{{\varvec{i}}} ,\user2{ Z}_{{\varvec{i}}} } \right) + \sigma_{n}^{2} {\varvec{I}}} \right]^{ - 1} \user2{ Y}$$
(34)
$${\varvec{\sigma}} = {\varvec{K}}\left( {{\varvec{Z}},\user2{ Z}} \right) - {\varvec{K}}\left( {{\varvec{Z}},\user2{ Z}_{{\varvec{i}}} } \right)\left[ {{\varvec{K}}\left( {{\varvec{Z}}_{{\varvec{i}}} ,{\varvec{Z}}_{{\varvec{i}}} } \right) + \sigma_{n}^{2} {\varvec{I}}} \right]^{ - 1} \user2{ K}\left( {{\varvec{Z}}_{{\varvec{i}}} ,\user2{ Z}} \right)$$
(35)

where, \({\varvec{Z}}\) represents sampling points and \({\varvec{Y}}\) is the corresponding numerical model output vector, \({\varvec{K}}\) is the covariance matrix and its elements are \({K}_{ij}=K({{\varvec{Z}}}_{{\varvec{i}}},{{\varvec{Z}}}_{{\varvec{j}}})\), \({\varvec{I}}\) is the unit matrix and \({\sigma }_{n}\) is the standard deviation of Gaussian noise for the training data. The squared exponential covariance function with automatic relevance determination (Rasmussen and Williams 2006) is used in this study as following.

$$K\left( {{\varvec{Z}}_{{\varvec{i}}} ,{\varvec{Z}}_{{\varvec{j}}} } \right) = \sigma_{f}^{2} \exp \left( { - \frac{1}{2}\left( {{\varvec{Z}}_{{\varvec{i}}} - {\varvec{Z}}_{{\varvec{j}}} } \right)^{T}{\varvec{\varLambda}}^{ - 2} \left( {{\varvec{Z}}_{{\varvec{i}}} - {\varvec{Z}}_{{\varvec{j}}} } \right)} \right)$$
(36)

where, \({\sigma }_{f}\) is the standard deviation of the field, \(\boldsymbol{\Lambda }=diag(\lambda )\) and a set of hyperparameter \(\theta =(\mathrm{ln}{\lambda }_{1}, \mathrm{ln}{\lambda }_{2}, \dots ,\mathrm{ln}{\lambda }_{n},\mathrm{ln}{\sigma }_{f})\) can be obtained optimally by the gradient‐based marginal likelihood optimization with the log‐likelihood function of the hyperparameter as following.

$$\log p({\varvec{Y}}|{\varvec{Z}}_{{\varvec{i}}} ,{\varvec{\theta}}) = - \frac{1}{2}{\varvec{Y}}^{T} \left[ {{\varvec{K}}\left( {{\varvec{Z}}_{{\varvec{i}}} ,\user2{ Z}_{{\varvec{i}}} } \right) + \sigma_{n}^{2} {\varvec{I}}} \right]^{ - 1} \user2{ Y} - \frac{1}{2}\log \left| {{\varvec{K}}\left( {{\varvec{Z}}_{{\varvec{i}}} ,\user2{ Z}_{{\varvec{i}}} } \right) + \sigma_{n}^{2} {\varvec{I}}} \right| - \frac{n}{2}\log 2\pi$$
(37)

Evaluation of RSMs performance

This section provides the details of the methodology adopted for the estimation of the reliability index (\(\beta\)) and the evaluation of the performances of RSMs.

Estimation of the reliability index

Estimation of \(\beta\) using RSMs involved two major steps. Firstly, RSMs were used to obtain the performance function (\(\widehat{G}\)) and then the MCSs were performed on this function to estimate the \(\beta\). MCSs method was adopted to estimate the reliability index (\(\beta\)) due to its accuracy, robustness and concept simplicity. MCSs involves the generation of a large number of random sets '\(N\)' of input rock properties, i.e., \({z}_{1}, {z}_{2}, {z}_{3},\dots ,{z}_{d}\), by random sampling based on their probability distributions (PDs). These random sets are treated as an outcome of the laboratory or field investigations. The values of the performance function (constructed using RSMs for this case), i.e., \(G\left({\theta }_{1}, {\theta }_{2}, {\theta }_{3},\dots ,{\theta }_{m};{z}_{1}, {z}_{2}, {z}_{3},\dots ,{z}_{d}\right)\) are estimated for generated random sets as a deterministic analysis. Figure 2 shows the basic concept of MCSs. The reliability index using RSMs-MCSs, i.e., \({\beta }_{RSM}\) can be estimated by evaluating the statistics of resulting performance function values as given below.

$$\beta_{RSM} = \frac{{E\left[ {\hat{G}} \right] - 1}}{{\sigma \left[ {\hat{G}} \right]}}$$
(38)
Fig. 2
figure 2

Basic concept of the Monte-Carlo Simulations (MCSs) method

where, \(E\left[\widehat{G}\right]\) and \(\sigma \left[\widehat{G}\right]\) are the mean and standard deviation (SD) of the performance function.

Additionally, the \(\beta\) was also estimated directly using the MCSs. This was done by evaluating \(G\left({\theta }_{1}, {\theta }_{2}, {\theta }_{3},\dots \dots ,{\theta }_{m};{z}_{1}, {z}_{2}, {z}_{3},\dots \dots ,{z}_{d}\right)\) for '\(N\)' random sets of input rock properties, i.e., \({z}_{1}, {z}_{2}, {z}_{3},\dots \dots {z}_{d}\) (generated via MCSs) using the original numerical program adopted to assess the stability of different rock slopes. Finally, \(\beta\) can be estimated similarly using Eq. 38 as explained above. This direct methodology and estimated \(\beta\) were termed as the direct MCSs (DMCSs) and \({\beta }_{DMC}\) respectively. Different accuracy indicators based on the \({\beta }_{DMC}\) and \({\beta }_{RSM}\) were then used to measure the accuracy of RSMs as explained in the next section.

Evaluation of the accuracy of RSMs

The accuracy of the RSMs was evaluated using two approaches for the current study as described below.

a) Initially, the accuracy of the RSMs was estimated using an accuracy indicator known as Nash–Sutcliffe Efficiency (NSE) (Moriasi et al. 2007). NSE can be estimated by generating \(k\) off-sample points (20 in this study) of input rock properties. For these off-sample points, FOSs were estimated using both the original numerical program and RSMs represented as \({FOS}_{i}^{original}\) and \({FOS}_{i}^{RSM}\) respectively. NSE can be estimated as given below.

$$NSE = 1 - \left[ {\frac{{\mathop \sum \nolimits_{i = 1}^{k} \left( {FOS_{i}^{original} - FOS_{i}^{RSM} } \right)^{2} }}{{\mathop \sum \nolimits_{i = 1}^{k} \left( {FOS_{i}^{original} - FOS^{mean} } \right)^{2} }}} \right]$$
(39)

where, \({FOS}^{mean}\) represents the mean value of \({FOS}_{i}^{original}\).

Performance of RSMs for NSE > 0.75 is considered very good. A complete description of the performance rating of RSMs based on NSE is shown in Appendix 2.

b) In addition to the NSE, the accuracy of RSMs was also evaluated with respect to the accuracy in the estimation of \(\beta\). For this analysis, an indicator namely Relative Difference in Reliability Index (RDRI) was used for this purpose which can be estimated as given below.

$$RDRI = \frac{{\left| {\beta_{DMC} - \beta_{RSM} } \right|}}{{\beta_{DMC} }}$$
(40)

Higher values of RDRI mean higher inaccuracies in the prediction of \(\beta\) by the RSMs.

Evaluation of computational efficiency of RSMs

The computational efficiency of the RSMs was defined in terms of the total time required to compute the performance function values for minimum number of sampling points (\({N}_{sp}^{min}\)) required to construct the RSMs with acceptable accuracy. This time was represented as \({T}_{sp}\) and can be estimated by multiplying \({N}_{sp}^{min}\) with the time required to evaluate one sample point (\({T}_{sp}^{1}\)) using the original numerical program as following.

$$T_{sp} = T_{sp}^{1} \times N_{sp}^{min}$$
(41)

Acceptable accuracy for the RSMs is the condition when the NSE values corresponding to very good performance rating (0.75–1.0) are achieved. NSE was used in the present study to determine the computational efficiency due to its clear guidelines on the expected performance of RSMs as compared to RDRI (Appendix 2). Further, the number of sample points used for the construction of RSMs was represented as \({N}_{sp}\) and the maximum number of sample points used for RSMs construction for an application example was represented as \({N}_{sp}^{max}\).

Application of RSMs to major slope reliability problems

This section investigates the applicability of RSMs for three types of rock slope reliability problems– type (1) stress-controlled slope failure neglecting spatial variability, type (2) stress-controlled slope failure considering spatial variability, and type (3) structurally controlled failure neglecting spatial variability. Three case studies corresponding to these problem types were analyzed using different RSMs to rank them based on their accuracy and computational efficiency.

Type 1 problem: stress-controlled slope failure neglecting spatial variability

Stress-controlled failures along the rock slopes occur generally along highly weathered or closely fractured rocks with approximately circular failure surfaces. These types of failures occur along the rock slopes when–i) individual blocks in the rock mass are very small compared to the slope dimensions or, in other words, discontinuity spacing is very small, and ii) no single discontinuity is dominating the slope behavior. Reliability analyses for this type of problem are performed by considering the homogenous distribution of input properties, i.e., neglecting their spatial variability, as shown in Table 1. This assumption is more appropriate for the slopes with higher correlation lengths between rock properties, i.e., low spatial variability. Reliability analyses for this problem types were generally performed by coupling analytical limit equilibrium methods with traditional reliability methods in the past. However, there were several subjective assumptions made in these analyses especially for the failure surfaces and material models (Wyllie and Mah 2004; Kumar et al. 2018; Abdulai and Sharifzadeh 2021). This often leads to inaccuracies in the estimated results. Hence, the analyses for these slopes are performed with advanced numerical tools (FDM, FEM etc.) in the recent years as it helps to overcome the limitations related to the prior assumptions of the failure surface and provides flexibility to consider different material models. However, it is not possible to obtain explicit performance functions for these analyses which are required to apply reliability methods efficiently (Pandit et al. 2019; Kumar and Tiwari 2022a). RSMs can provide a better alternative under these conditions which will be investigated through a systematic analysis of a case study for this article.

The case study considered for this problem type is a large rock slope (dimensions of 293 m × 196 m) supporting piers of the world's highest railway bridge crossing the Chenab River in Jammu and Kashmir, India. Rock present at the site was dolomite intersected with closely spaced three major joint/discontinuity sets along with some random joint sets (Tiwari and Latha 2019). Physical and mechanical properties of intact rock were estimated by International Society for Rock Mechanics (ISRM) suggested methods (ISRM 1981) and are summarized in Table 3. Table 4 shows the properties of the major joint sets along the rock slope. Rock mass along the slope was assumed to follow elastic-perfectly plastic behavior with the Hoek–Brown criterion. Numerical modeling for this slope was carried out in Phase2 (Rocscience 2014). FOS was estimated using the shear strength reduction (SSR) method (Dawson et al. 1999). Rock mass properties (i.e., \({m}_{b}\), \({s}_{b}\), \({E}_{m}\)) required as input for the numerical model were calculated using their well-accepted relations with rock mass classification rating GSI (Hoek et al. 2013) and intact rock properties, i.e., \({m}_{i}\), \({E}_{i}\) (Appendix 3) (Hoek et al. 2002; Hoek and Diederichs 2006). Figure 3 shows the prepared numerical model of the slope in Phase2 along with the corresponding shear strain contour, FOS and velocity vector pattern obtained after the analysis for a typical model.

Table 3 Statistical parameters of the rock properties for the rock slope considered for type 1 problem
Table 4 Details of the major joint set along the rock slope considered for type 1 problem
Fig. 3
figure 3

a Numerical model prepared in the finite element method (FEM) based software Phase2 for type 1 problem, and b Shear strain contour, factor of safety and velocity vector pattern evaluated for a typical model analysis


Performances of different RSMs were evaluated for this slope as explained in “Polynomial model with cross terms (Q-I)”. Necessary parameters required for the comparative analysis between the considered RSMs were estimated initially. Table 3 shows the properties considered in the analysis. DMCSs was performed for 1000 random sets of input properties, i.e., \({m}_{i}\), \({E}_{i}\), UCS and GSI (total 4 random properties; Table 3) generated according to their PD function (“Estimation of the reliability index”). Corresponding values of rock mass properties, i.e., \({m}_{b}\), \({s}_{b}\), \({E}_{m}\) were estimated using GSI based relations (Appendix 3). FOS values were then estimated for these random sets from the numerical model of the slope in Phase2. The number of DMCSs random sets was decided based on the convergence of FOS statistics, i.e., mean and standard deviation (SD). Figure 4 shows the variation of mean and SD of FOS with the number of random sets and a good convergence can be observed for 1000 random sets. The estimated FOS for these random sets ranged between 1.02 and 5.17 and Fig. 5a also shows the histogram of the FOS. The reliability index from DMCSs (\({\beta }_{DMC}\)) was estimated to be 1.99 through Eq. 38. Further, \({\beta }_{RSM}\) was estimated using 10,000 MCSs on the constructed RSMs. RSMs were constructed with different numbers of sample points (\({N}_{sp}\)) generated using LHS sampling as explained in “Response Surface Methods (RSMs).” Figure 5b-c shows the results for this application example. It can be seen that with the increase in the number of sample points (\({N}_{sp}\)), NSEs of the considered RSMs were continuously increasing. Values were converging towards their maximum magnitude of 1, indicating the continuous improvement in their accuracy with the increasing \({N}_{sp}\). This trend was similar for the estimated β using RSMs as the RDRI was also converging towards 0% for all the considered RSMs with the increasing \({N}_{sp}\). It can be observed that the acceptable accuracies for all the considered RSMs were achieved for \({N}_{sp}\)<\({N}_{sp}^{max}\)(100 points). However, there were noticeable differences in the minimum required sampling points (\({N}_{sp}^{min}\)) for the RSMs. Table 5 shows the \({N}_{sp}^{min}\) and the associated \({T}_{sp}\) for the RSMs. It can be observed that the Kriging and LS-SVM were the most and C-II RBF was the least computationally efficient RSM with \({N}_{sp}^{min}\) values of 9 \(({T}_{sp}\)=1.35 × 104 s) and 75 \(({T}_{sp}\)=1.12 × 105 s), respectively. It can be concluded that the accuracies of the considered RSMs were highly comparable. However, differences exist in the computational efficiencies of the RSMs for type 1 problem.

Fig. 4
figure 4

Demonstration of convergence of factor of safety statistics with the increasing Monte-Carlo random sets a mean, and b standard deviation (SD)

Fig. 5
figure 5

Results of Response Surface Methods (RSMs) based reliability analysis for type 1 problem a Histogram of Factor of Safety (FOS) evaluated using Direct Monte-Carlo Simulations (DMCSs), b Nash-Stucliffe Efficiency (NSE) vs number of sample points (\({N}_{sp}\)) showing the improving accuracy of RSMs with increasing sampling points, and c Relative Difference in Reliability Index (RDRI) vs number of sample points (\({N}_{sp}\)) concluding the similar result as (b)

Table 5 Results regarding computational efficiency of the Response Surface Methods (RSMs) for the rock slope considered for type 1 problem

Type 2 problem: stress-controlled slope failure considering spatial variability

Failure mode and in-situ rock conditions required for these types of slope failure analysis remain the same as that of type 1 problem. However, if the correlation length for properties is low, i.e., spatial variability is high then the reliability analyses performed by assuming a homogenous distribution of input properties can lead to high inaccuracies. In these cases, reliability analysis performed by considering spatial variability of rock properties is more accurate and realistic (Huang et al. 2017; Pandit et al. 2018). The analysis is generally performed using direct MCSs on randomly generated models. Limited studies are available on the analysis of this type of problem (Huang et al. 2017; Pandit et al. 2018) possibly due to complex mathematical formulation and the requirement of high computational efforts (Table 1). RSMs can provide a better alternative to the traditional reliability analysis under these conditions which will be investigated through a systematic analysis of a case study.

The case study considered for the type 2 problem is a gold mine slope located in the Haveri district of Karnataka state in India (Pandit et al. 2018). The major rock type present at the site was greywacke. All the properties were estimated by detailed ISRM suggested methods (ISRM 1981) and relevant physical and mechanical properties are shown in Table 6. Table 7 shows the details of the major joint sets along the rock slope. Rock mass along the slope was assumed to follow elastic-perfectly plastic behavior with the Hoek–Brown criterion. Uniaxial compressive strength (UCS) was modeled as a Random Field (RF), GSI as a Random Variable (RV) and all other properties as deterministic variables based on their sensitivity towards FOS. More details on the estimation of properties and sensitivity analysis can be seen in Pandit et al. 2018. FOS was estimated using the SSR method (Dawson et al. 1999).

Table 6 Statistical parameters of the rock properties for the rock slope considered for type 1 problem
Table 7 Details of the major joint set along the rock slope considered for type 2 problem

Spatial variability of rock properties was considered in the analysis using the random field modeling. For random field modeling, a single exponential 2D autocorrelation model was adopted which can be expressed as following.

$${\rho }_{w}\left(\Delta x,\Delta z\right)=exp\left(-\frac{2\left|\Delta x\right|}{{\delta }_{x}}- \frac{2\left|\Delta z\right|}{{\delta }_{z}}\right)$$
(42)

where, \({\rho }_{w}(\Delta x,\Delta z)\) is correlation value at a distance of \(\Delta x\) in the horizontal direction and \(\Delta z\) in the vertical direction from the point \(\left(x,z\right)\); \({\delta }_{x}\) and \({\delta }_{z}\) represent the horizontal and vertical scale of fluctuations (SOFs) estimated from the rock properties along different boreholes using the maximum likelihood method in the logarithmic domain. For the present study, the values of both SOFs were considered as 64 m. For the discretization, the Expansion Optimal Linear Estimation (EOLE) method (Li and Kiureghian 1993) was adopted, which discretizes the Gaussian random field \(Z\) as shown below.

$$\tilde{Z}\left( {x,z} \right) \cong \mu_{\ln Z } + \sigma_{\ln Z } \mathop \sum \limits_{j = 1}^{N} \frac{{\xi_{j} }}{{\sqrt {\lambda_{j} } }}\phi_{j} {\Omega }$$
(43)

where, \({\mu }_{\mathrm{ln}Z}\) and \({\sigma }_{\mathrm{ln}Z}\) are the mean and standard deviation of the random field \(Z\) in logarithmic space. \({\lambda }_{j}\), \({\phi }_{j}\) are eigenvalues and eigenvectors of \(\Sigma\), and \(\Omega\) is the correlation vector between the random field value at a point \(\left(x,z\right)\) and all other points in the grid. \({\xi }_{j}\) is a vector of \(N\) independent standard normal random variables where \(N\) is the number of terms retained in the EOLE. \(N\) was obtained by arranging the eigenvalues in decreasing order and selecting the first \(N\) eigenvalues so that the mean error variance was smaller than the target value. Expression of the error variance is given as follows.

$$Var\left[ {Z\left( {x,z} \right) - \tilde{Z}\left( {x,z} \right)} \right] = \sigma^{2} - \mathop \sum \limits_{j = 1}^{N} \frac{1}{{\lambda_{j} }}\left[ {\left( {\phi_{j} } \right)^{T} \Omega } \right]^{2}$$
(44)

where, \(Var\) is variance, \({\left({\phi }_{j}\right)}^{T}\) is the transpose of the vector \({\phi }_{j}\) and \(Z\left(x,z\right)-\widetilde{Z}\left(x,z\right)\) is the difference between the true and approximate value of the random field at point \(\left(x,z\right)\). The target mean error variance was taken as 10% in this study giving the value of \(N\) = 43. This means that the underlying continuous random field was approximated by 43 discrete random variables. Numerical modeling was carried out using a random field approach implemented in FLAC-2D (Itasca 2011) coupled with MATLAB. Figure 6 shows a typical numerical model along with the corresponding shear strain contour, FOS and velocity vector pattern obtained after the analysis for a typical model.

Fig. 6
figure 6

a Numerical model prepared in the Finite Difference Method (FDM) based software FLAC-2D for type 2 problem, and b Shear strain contour, Factor of Safety (FOS) and velocity vector pattern evaluated for a typical model analysis

For comparison between the performances of RSMs, required parameters were estimated initially. DMCSs was performed by generating 1000 random finite difference models (RFDMs) with spatial variability in rock properties. For each simulation in DMCSs, the random input set \({\xi }_{j}\); \(j=1,\dots , N\) were obtained and utilized in Eq. 43 to generate spatially variable properties. For the next simulation in DMCSs, a different set of \({\xi }_{j}\); \(j=1,\dots , N\) was employed. Figure 6 shows a typical random finite difference model (RFDM) with spatial variation in UCS. FOS values for random finite difference models (RFDMs) were estimated using the shear strength reduction (SSR) method in FLAC. The estimated FOS for these random finite difference models were ranging between 2.17 and 7.10 and Fig. 7a shows the histogram of the estimated FOS. \({\beta }_{DMC}\) for the slope was estimated to be 3.07 using Eq. 38. Further, \({\beta }_{RSM}\) was estimated using 10,000 MCSs on the constructed RSMs. RSMs were constructed with different numbers of sample points generated using Latin Hypercube Sampling (LHS) as explained in “Response Surface Methods (RSMs).”. Figure 7b-c shows the results for this application example. Similar to the previous case, the accuracy of RSMs was continuously improving as indicated by the increasing Nash–Sutcliffe Efficiency (NSE) and decreasing Relative Difference in Reliability Index (RDRI) with the increasing number of sample points (\({N}_{sp}\)). However, some of the RSMs could not achieve the acceptable accuracy within the considered maximum number of sample points (\({N}_{sp}^{max}=150)\) for this application example. Further, a considerable difference exists in the estimated minimum number of sample points (\({N}_{sp}^{min}\)) and the associated total numerical models run time required (\({T}_{sp}\)) for the considered RSMs in this study. Table 8 shows the \({N}_{sp}^{min}\) and the associated \({T}_{sp}\) for the RSMs. It can be observed that almost all of the RBF based RSMs and LS-SVM are the most while Q-I, Q-II, MLS, Kriging, T-SVM and GPR are the least computationally efficient RSMs with \({N}_{sp}^{min}\) values of 75 \(({T}_{sp}\)=3.38 × 105 s) and undetermined (> \({N}_{sp}^{max}\)) respectively. It can be observed that the overall requirement of the \({N}_{sp}^{min}\) for this example was significantly higher than the previous example due to the involvement of the higher number of random variables (i.e., 44 RVs) (Zhou and Huang 2018). This should be emphasized that the number of the sampling points required for constructing RSM was problem type dependent. This could be due to the positive correlation between the number of sampling points and the number of input random variables. It can be seen that the number of random variables for type 2 problem was 90% higher as compared to type 1 problem (number of variables were 4). Hence, the required sampling points to achieve the acceptable accuracy for type 2 problem was 90% higher for most RSMs. Further, many RSMs could not achieve the acceptable accuracy for this problem type even for a higher number of sampling points possibly due to the higher number of input variables. This issue was not present in the type 1 problem where all RSMs achieved acceptable accuracy between 50 and 75 sampling points. For type 2 problem, RSMs like polynomial (Q-II) and C-RBF have significantly lower accuracy for sampling points ranging between 75 and 150 with an approximate percentage difference of 50% in NSE and RDRI as compared to other RSMs. Hence, it is suggested to choose the RSM for this type of problem with utmost attention to get more accurate results along with computational efficiency.

Fig. 7
figure 7

Results of Response Surface Methods (RSMs) based reliability analysis for type 2 problem a Histogram of Factor of Safety (FOS) evaluated using Direct Monte-Carlo Simulations (DMCSs), b Nash-Stucliffe Efficiency (NSE) vs number of sample points (\({N}_{sp}\)) showing the improving accuracy of most of the RSMs with increasing sampling points and failure of Q-I, Q-II, MLS and GPR to achieve acceptable accuracy within the maximum number of sample points (\({N}_{sp}^{max}\)), and c Relative Difference in Reliability Index (RDRI) vs number of sample points (\({N}_{sp}\)) concluding the similar result as (b)

Table 8 Results regarding computational efficiency of the Response Surface Methods (RSMs) for the rock slope considered for type 2 problem

Type 3 problem: structurally controlled failure neglecting spatial variability

Structurally controlled failures along the rock slopes are generally referred to the block failures along discontinuities. These types of failures occur along those slopes for which the discontinuity orientation and shear strength favor the block movement along them (Park and West 2001; Park et al. 2005; Tatone and Grasselli 2010; Ahmadabadi and Poisel 2016). The potential of structurally controlled failures and identification of their failure modes (planar, wedge, toppling) can be evaluated using kinematic analysis (Park and West 2001; Wyllie and Mah 2004). Reliability analyses for the type 3 problem are generally performed by considering the homogenous distribution of discontinuity properties along with other rock properties in the past (Table 1). Reliability analysis is generally performed by coupling analytical formulation of performance functions using limit equilibrium methods with traditional reliability methods. However, some studies have been recently performed to study these problems with advanced numerical tools (Tiwari and Latha 2016; Dadashzadeh et al. 2017; Sengani et al. 2021) like DEM, FEM etc., to study the displacements of slopes and the effect of post-peak strength behavior of discontinuities. Limit equilibrium methods cannot be applied under these conditions and hence, it becomes difficult to obtain an explicit performance function under these conditions (Dadashzadeh et al. 2017; Pandit et al. 2018, 2019). RSMs can provide an efficient alternative under these conditions. This section investigates the applicability of RSMs for these problems through a systematic analysis of a case study.

The case study considered for the type 3 problem is that of a potential massive rockslide along the Rishikesh-Badrinath highway located in the Uttarakhand state of India. The expected failure mechanism for this potential rockslide was planar failure as per kinematic stereographic analysis (Pain 2012). The major rock type present at the site was Quartzite. All the properties of the intact rock were estimated by ISRM suggested methods (ISRM 1981) as shown in Table 9. Table 10 also shows the properties of the major discontinuities.

Table 9 Statistical parameters of the rock properties for the rock slope considered for type 3 problem
Table 10 Details of the joint set for the rock slope considered for type 3 problem

Along the rock slope. Geometrical parameters of the slope and discontinuity were considered deterministic. As the slope showed the potential of failure along the discontinuity, strength properties of the discontinuities were considered to be random variables. Numerical modeling of the slope was carried out in Phase2 (Rocscience 2014) and FOS was evaluated using the SSR method (Dawson et al. 1999). Figure 8 shows the numerical model of the slope prepared in Phase2 along with the corresponding displacement contour, FOS and velocity vector pattern obtained after the analysis for a typical model.

Fig. 8
figure 8

a Numerical model created in the Finite Element Method (FEM) based software Phase2 for type 3 problem and b Displacement contour, Factor of Safety (FOS) and velocity vector pattern evaluated for a typical model analysis

For comparison between the performances of RSMs, required parameters were estimated initially. 1000 random sets of the random discontinuity properties (total 3 random properties; Table 9) were generated to perform the DMCSs. FOS values were estimated for these randomly generated random sets of rock properties using Phase2. Figure 9a shows the histogram of the estimated FOS ranging between 1.20 and 2.10. The \({\beta }_{DMC}\) was estimated to be 3.54 using Eq. 38. Further, \({\beta }_{RSM}\) was estimated using 10,000 MCSs on the constructed RSMs. RSMs were constructed with a different number of sampling points generated using LHS sampling as explained in “Response surface methods (RSMS)”. Figure 9b-c shows the results for the considered application example. Similar to the previous cases, the accuracy of the RSMs was improving as indicated by the increasing NSEs and decreasing RDRI with the increase in the \({N}_{sp}\). All of the RSMs achieved acceptable accuracy at \({N}_{sp}\) much lesser than the \({N}_{sp}^{max}\) (100 points). However, a difference was observed in \({N}_{sp}^{min}\) for the considered RSMs. Table 11 shows the \({N}_{sp}^{min}\) and \({T}_{sp}\) for the considered RSMs. It can be observed that the AC-I-RBF, AC-II-RBF, Kriging and LS-SVM are the most, while L-RBF, G-RBF, C-II-RBF and MLS are the least efficient RSMs with \({N}_{sp}^{min}\) values of 7 \(({T}_{sp}\)= 6.3 × 103 s) and 25 \(({T}_{sp}\)=2.25 × 104 s) respectively. It can be observed that accuracies of the RSMs were highly comparable similar to type 1 problem. However, differences exist in the computational efficiencies of the RSMs for this problem type.

Fig. 9
figure 9

Results of Response Surface Methods (RSMs) based reliability analysis for type 2 problem a Histogram of Factor of Safety (FOS) evaluated using Direct Monte-Carlo Simulations (DMCSs), b Nash-Stucliffe Efficiency (NSE) vs number of sample points (\({N}_{sp}\)) showing the improving accuracy of RSMs with increasing sampling points, and (c) Relative Difference in Reliability Index (RDRI) vs number of sample points (\({N}_{sp}\)) concluding the similar result as (b)

Table 11 Results regarding computational efficiency of the Response Surface Methods (RSMs) for the rock slope considered for type 3 problem

Suggested RSMs for reliability analysis of modelled rock slopes

It was discussed in the literature review that most of the previous studies employed polynomial RSM due to its mathematical simplicity. Although advanced RSMs provide a better alternative to traditional RSMs for most engineering problems, their applicability is questionable for the rock slope problems. This issue has been addressed in the previous sections as most of the advanced RSMs achieved acceptable accuracy with fewer sampling points for all problem types. This could be possibly due to their inherent feature of weighing the data samples near the observation points more heavily rather than giving them equal weights. However, one major issue for the users will be the selection of the most appropriate RSM among these advanced RSMs and the corresponding number of sampling points. To the author's best knowledge, no such guidelines for the rock slope problems exist in the current literature. This study has made an effort to provide the guidelines for the users for the initial selection of the most appropriate RSM and an appropriate number of sampling points based on the type of rock slope reliability problem under consideration.

Table 12 provides a comparison between the performances of the RSMs based on their accuracy and computational efficiency for different types of rock slope reliability problems. Symbol ‘’ means that the RSMs have achieved the acceptable accuracy for \({{N}_{sp}<N}_{sp}^{max}\), while ‘’ represents the RSM which could not. For computational efficiency, RSMs were ranked according to their estimated ‘\({T}_{sp}\)’ with ‘1’ referring to the most efficient RSM. It can be observed that for—a) type 1 problems, Kriging and LS-SVM, b) type 2 problems, L-RBF, G-RBF, MQ-RBF, C-II-RBF, AC-I-RBF, AC-II-RBF and LS-SVM, and c) type 3 problems, AC-I-RBF, AC-II-RBF, Kriging and LS-SVM were the accurate and most efficient RSMs. Another important recommendation is regarding the number of sampling points to obtain RSMs with acceptable accuracy. It is recommended to use \({N}_{sp}^{min} \approx 2\times d\) (number of involved random variables) as the initial trial and then increase the \({N}_{sp}\) if the acceptable accuracy is not achieved.

Table 12 Recommendations for selecting the Response Surface Methods (RSMs) for different types of rock slope reliability problems

Conclusions

This study investigates the performance of different RSMs for the risk assessment of rock slope stability problems on the criteria of accuracy and computational efficiency. Three types of slope reliability problems were identified from a detailed literature review – type (1) stress-controlled failures neglecting spatial variability, type (2) stress-controlled failures considering spatial variability, and type (3) structurally controlled failures neglecting spatial variability. Analyses were performed for large sized rock slopes corresponding to these problem types along different engineering sites in India. In general, RSMs based reliability analysis was observed to be highly accurate and computationally efficient for all problem types. However, the suitability of different RSMs was observed to be problem dependent. Traditional global RSM such as polynomial was observed to perform inferior to advanced RSMs which weigh the data samples near the observation points more heavily rather than giving equal weights. Some specific recommendations were provided based on the extensive simulations and data analyses for the selection of the appropriate RSM based on the problem types. Kriging and LS-SVM for type 1; L-RBF, G-RBF, MQ-RBF, C-II-RBF, AC-I-RBF, AC-II-RBF and LS-SVM for type 2; and AC-I-RBF, AC-II-RBF, Kriging and LS-SVM for type 3 problems were suggested to be the most suitable RSMs. The number of sampling points required for the construction of RSM was also observed to be dependent on the type of problem. A possible reason could be the direct dependence of the number of sampling points on the number of the input random variables which in turn depends upon the problem type. For this study, the number of random variables for type 2 problem was 44 which was 90% higher as compared to type 1 (number of variables were 4) and 3 (number of variables were 4) problems. Hence, the number of sampling points required to achieve the acceptable accuracy for type 2 problem was 90% higher for most RSMs. Further, the minimum number of sampling points required to construct the response surface of acceptable accuracy was recommended to be twice the number of input random variables and more points can be included quantitatively by assessing the accuracy of the constructed RSM.

Although the study provides important conclusions on the applicability of RSMs for the reliability analysis of rock slopes, some issues need to be investigated in future studies. For type 2 problem, suitability of RSMs for rock slopes with different auto-correlation lengths, especially smaller lengths, should be further studied. This is because of the increasing number of random variables for the problems with smaller correlation lengths. Further, the applicability of RSMs for the reliability analysis of structurally controlled rock slopes considering spatial variability of properties should be investigated.