Keywords

1 Introduction

During the last three decades, a number of laboratory model test results and few field test results have been published that are related to the ultimate bearing capacity of shallow foundation resting over homogeneous sand bed and clay extending to a great depth. The presence of a hard layer within a certain depth below the foundation base can significantly influence the ultimate bearing capacity supported by the soil. In many practical problems, the thin soil layer may be underlain by rigid rock base. In these situations, the ultimate bearing capacity of shallow foundation resting on soil layer of limited thickness is influenced by the lower rigid boundary. Most of the experimental studies were related to centric vertical loading. However, none of the published studies address the effect of load eccentricity, load inclination, and the effect of rigid base located at a limited depth on the ultimate bearing capacity of circular foundation on limited thickness of sand layer using ANN. The purpose of this study is to develop a neural network model from the results of laboratory model tests to estimate the reduction factor. Artificial neural network (ANN) is an artificial intelligence system inspired by the behavior of human brain and nervous system. In the present study a feed forward back propagation neural network model has been used to predict the reduction factor of eccentrically inclined loaded circular foundation. Backpropagation neural network is most suitable for prediction problems and Levenberg-Marquadrt algorithm is adopted as it is efficient in comparison to gradient descent backpropagation algorithm (Goh et al. 2005; Hornik et al. 1989). By drawing a neural interpretation diagram relationship between input and output are found out. A prediction model is developed based on the weights of the ANN model. The developed reduction factor is compared with the experimental reduction factor.

2 Analysis and Data

All the laboratory model tests were conducted using a poorly graded sand with effective grain size D10 = 0.325 mm, uniformity coefficient, Cu = 1.45, and coefficient of gradation, Cc = 1.15. Relative density Dr of sand is 69% and angle of internal friction, ϕ = 40.9°. Model foundations used for the tests had dimensions of 100 mm diameter. Mild steel plate 30-mm thick was used to make the model foundations. The bottom of the foundation was made rough by applying glue and rolling the steel plate over sand.

One hundred laboratory model tests were conducted. Three parameters H/B, α/ϕ and e/B are used as inputs in the ANN model, and the output is the reduction factor RF given by

$$ RF = \frac{{q_{u(H/B,\;\alpha /\phi ,\;e/B)} }}{{q_{u(\alpha /\phi ,\;e/B)} }} $$
(1)

where \( q_{u(H/B,\;\alpha /\phi ,\;e/B)} \) is the ultimate eccentrically inclined load per unit area of a circular foundation on the surface of a sand layer of limited thickness (Table 1) and \( q_{u(\alpha /\phi ,\;e/B)} \) is the ultimate eccentrically inclined load per unit area on a sand layer extending to a great depth (Table 2).

Table 1. Database used for ANN model and compared with experimental results
Table 2. Database used for ANN model

Out of 100 tests, 75 tests are considered for training and the remaining 25 are considered for testing. All the inputs and output are normalized in the range of [−1, 1] before training. A feed-forward back-propagation neural network is used with hyperbolic tangent sigmoid function and linear function as the transfer function. The network is trained with Levenberg-Marquardt (LM) algorithm as it is efficient in comparison to gradient descent back-propagation algorithm. The ANN has been implemented using MATLAB V 7.11.0(R2015b).

3 Results and Discussion

Three inputs and one output parameters were considered in the ANN model. The schematic diagram of the ANN architecture is shown in Fig. 1, which was computed from the database. The number of neurons in hidden layer is varied and the optimum number was taken based on mean square error (mse) value which was maintained at 0.001. In this ANN model there were two neurons evaluated in hidden layer as shown in Fig. 2. Therefore the final ANN architecture as 3-2-1[i.e. 3 (input) – 2 (hidden layer neuron) – 1 (output)].

Fig. 1.
figure 1

ANN architecture

Fig. 2.
figure 2

Variation of hidden layer neuron with mean square error (mse)

Mean square error (MSE) is defined as

$$ MSE = \frac{{\sum\limits_{i = 1}^{n} {\left( {RF_{i} - RF_{p} } \right)^{2} } }}{n} $$
(2)

Coefficient of efficiency, R2 is defined as

$$ R^{2} = \frac{{E_{1} - E_{2} }}{{E_{1} }} $$
(3)

where,

$$ E_{1} = \sum\limits_{i = 1}^{n} {\left( {RF_{i} - \overline{RF} } \right)}^{2} $$
(4)

and

$$ E_{2} = \sum\limits_{i = 1}^{n} {\left( {RF_{p} - RF_{i} } \right)}^{2} $$
(5)

where, \( RF_{i} ,\overline{RF} \) and \( RF_{p} \) are the experimental, average experimental, predicted RF values respectively; and n = number of training data.

The coefficient of efficiency (R2) is found to be 0.997 for training and 0.996 for testing as shown in Figs. 3, and 4. The weights and biases of the network are presented in Table 3. These weights and biases can be utilized for interpretation of relationship in between the inputs and output, sensitivity analysis and framing an ANN model in the form of an equation. The residual analysis was carried out by calculating the residuals in between experimental reduction factor and predicted reduction factor for training data. Residuals can be defined as the difference between the experimental and predicted RF value and is given by

Fig. 3.
figure 3

Correlation between predicted reduction factors with experimental reduction factor for training data

Fig. 4.
figure 4

Correlation between predicted reduction factors with experimental reduction factor for testing data

Table 3. Values of connection weights and biases
$$ e_{r} = RF_{i} - RF_{p} $$
(6)

The residuals are plotted with the experiment number as shown in Fig. 5. It is observed that the residuals are evenly distributed along the horizontal axis of the plot. Therefore it can be said that the network is well trained and can be used for prediction with reasonable accuracy.

Fig. 5.
figure 5

Residual distribution of training data

4 Sensitivity Analysis

Sensitivity analysis was carried out for selection of important input variables. Different approaches have been suggested to select the important input variables. The Pearson correlation coefficient is one of them in selecting proper inputs for the ANN model. It was approached by Guyon and Elisseeff (2003) and Wilby et al. (2003). Goh (1994), Shahin et al. (2002), Behera et al. (2013), Sahu et al. (2017, 2018), Sethy et al. (2017) have used Garson’s algorithm (Garson 1991) in which the input-hidden and hidden-output weights of trained ANN model are partitioned. In Garson’s algorithm the absolute values of weights are taken to select the important input variables. It does not provide information on the effect of input variables in terms of direct or inverse relation to the output. Olden et al. (2004) proposed a connection weights approach based on the neural interpretation diagram (NID), in which the actual values of input-hidden and hidden-output weights are taken. Table 4 shows the cross-correlation of the three input parameters with the reduction factor (RF) value. From the Table 4 it can be seen that RF is highly correlated to H/B with a values of −0.71 followed by α/ϕ = −0.11 and e/B = −0.01. The relative importance, quantified through the parameter Si of three input parameters as per Garson’s algorithm is presented in Table 5. The H/B is found to be the most important input parameters with relative importance value being 78.71% followed by 19.45% for α/ϕ and 1.84% for e/B. As per the connection weight approach (Olden et al. 2004) the relative importance of the present input variables is also presented in Table 5. H/B is also the most important input parameter (Si = −11.28) followed by α/ϕ (Si = −0.5) and e/B (Si = −0.05). The Si values being negative imply that H/B, α/ϕ, and e/B are indirectly related to RF. In other words increase in H/B, α/ϕ, and e/B leads to decrease in RF and leads to decrease in ultimate bearing capacity.

Table 4. Cross-correlation matrix of input and output for reduction factor
Table 5. Relative importance of different inputs as per Garson’s algorithm and connection weight approach

5 Neural Interpretation Diagram (NID)

Ozesmi and Ozesmi (1999) proposed neural interpretation diagram for visual interpretation of the connection weight among the neurons. For the present study with the weights as obtained and shown in Table 1, an NID is presented in Fig. 6. The lines joining the input-hidden and hidden output neurons represent the weights. The positive weights are represented by solid lines and negative weights by dashed lines and the thickness of the line is proportional to its magnitude.

Fig. 6.
figure 6

Neural interpretation diagram (NID) showing lines representing connection weights and effects of inputs on reduction factor (RF)

It is seen from Table 5 that Si values for parameters H/B, α/ϕ, and e/B are negative indicating that the parameters are indirectly related to RF values, whereas Si values for. This is shown in Fig. 6. Therefore, the developed ANN model is not a black box and could explain the physical effect of input parameters on the output.

6 ANN Model Equation for Reduction Factor Based on Trained Neural Network

A model equation is developed using the weights obtained from trained neural network model (Goh et al. 2005). The mathematical equation relating input parameters (H/B, α/ϕ, and e/B) to output given by

$$ RF_{n} = f_{n} \left\{ {b_{0} + \mathop \sum \limits_{k = \;1}^{h} \left[ {w_{k} f_{n} \left( {b_{hk} + \mathop \sum \limits_{i\; = \;1}^{m} w_{ik} X_{i} } \right)} \right]} \right\} $$
(7)

where RFn is the normalized value of RF in the range [−1, 1], \( f_{n} \) is the transfer function, h is the number of neurons in the hidden layer, Xi is the normalized value of inputs in the range [−1, 1], m is the number of input variables, \( w_{ik} \) is the connection weight between the ith layer of input and kth neuron of hidden layer, \( w_{k} \) is the connection weight between the kth neuron of hidden layer and single output neuron, \( b_{hk} \) is the bias at the kth neuron of hidden layer and \( b_{0} \) is the bias at the output layer.

The model equation of RF of shallow circular foundations on sand layer of limited thickness subjected to eccentrically inclined load was formulated using the values of the weights and biases shown in Table 3 as per the following steps.

Step 1

The input parameters were normalized in the range [−1, 1] by the following expressions

$$ X_{n} = 2\;\left( {\frac{{X_{n} - X_{\hbox{min} } }}{{X_{\hbox{max} } - X_{\hbox{min} } }}} \right) $$
(8)

Step 2

Calculate the normalized value of reduction factor (RFn) using the following expressions

$$ A_{1} = - 4.89\left( {\frac{H}{B}} \right)_{n} - 0.13\left( {\frac{\alpha }{\phi }} \right)_{n} - 0.01\left( {\frac{{D_{f} }}{B}} \right)_{n} - 5.19 $$
(9)
$$ A_{2} = - 6.01\left( {\frac{B}{L}} \right)_{n} - 3.63\left( {\frac{e}{B}} \right)_{n} - 0.34\left( {\frac{{D_{f} }}{B}} \right)_{n} - 2.59 $$
(10)
$$ B_{1} = 2.23\left( {\frac{{e^{{A\,_{1} }} - e^{{ - A\,_{1} }} }}{{e^{{A_{\,1} }} + e^{{ - A\,_{1} }} }}} \right) $$
(11)
$$ B_{2} = 0.06\left( {\frac{{e^{{A\,_{1} }} - e^{{ - A\,_{1} }} }}{{e^{{A_{\,1} }} + e^{{ - A\,_{1} }} }}} \right) $$
(12)
$$ C_{1} = B_{1} + B_{2} + 1.32 $$
(13)
$$ RF_{n} = C_{1} $$
(14)

Denormalize the RFn value obtained from Eq. 14 to actual RF as

$$ RF = 0.5(RF_{n} + 1)\left( {RF_{\hbox{max} } - RF_{\hbox{min} } } \right) + RF_{\hbox{min} } $$
(15)
$$ RF = 0.5(RF_{n} + 1)\left( {7.84 - 1.01} \right) + 1.01 $$
(16)

Figure 7 Shows the comparison of reduction factor (RF) obtained from Eqs. 16 and 1. It can be seen that the results predicted by using artificial neural network (ANN) are closer to the results obtained from model tests. The deviation between the RF obtained from model tests and those predicted RF is within ±10% except two values as shown in Table 1. The proposed ANN model can be used as an effective tool in predicting the reduction factor (RF) and hence, the ultimate bearing capacity shallow circular foundation on sand layer of limited thickness underlain by rigid rough base subjected to an eccentrically inclined load.

Fig. 7.
figure 7

Comparison of ANN results with experimental RF

7 Conclusion

Based on developed neural network model, the following conclusions may be drawn.

  1. 1.

    The errors are distributed evenly along the centerline as per residual analysis. It can be concluded that the network was well trained and can predict the reduction factor (RF).

  2. 2.

    Based on Pearson correlation coefficient and Garson’s algorithm, it was observed that H/B is the most important input parameter followed by α/ϕ and e/B.

  3. 3.

    The developed ANN model could explain the physical effect of inputs on the output, as described in NID. It has been observed that H/B, α/ϕ, and e/B is inversely related to RF.

  4. 4.

    A model equation is developed based on the trained weights of ANN to calculate the ultimate bearing capacity of circular foundation on sand layer of limited thickness.

  5. 5.

    The deviation between the RF obtained from model tests and those predicted from ANN model is within ±10%