Introduction

From the engineering point of view, soil is a very important and universal ancient construction and base material. Soil is very popular due to its versatile properties, such as moisture content, quartz content, porosity, mineral content, and organic content. It is very popular in construction industries and geotechnical field. Furthermore, with the help of some soil properties, we can compute thermal conductivity of soil by various methods like conventional method that comes from ancient and another by using machine learning which uses a variety of machine learning algorithms. Thermal conductivity is world-widely very important from many points of view such as large and heavy construction, agriculture field, geothermal resources development, land surface processing, and underground construction. Thermal conductivity is a characteristic of soil that manages the thermal properties on various types of soil depending upon various types of parameters related to that soil. Among these parameters, some parameter like quartz content, water content and cross-section, and thermal conductivity is highly reliable. Thermal conductivity is basically saying about thermal energy transformation via soil to either atmosphere or building.

Many researchers have used experimental and laboratory methods to determine heat conductivity, including Sepaskhah and Boersma (1979) who have used a laboratory method to find the soil’s thermal conductivity with heat probe method, and they have considered that soil’s thermal conductivity depends upon water content and temperature. The measurement of thermal conductivity was done for loam, loamy sand, and soils with silty clay loam with the help of a cylinder-shaped heat sensor method at different water contents and temperatures. Abu-Hamdeh and Reeder (2000) have done a laboratory experiment to estimate the thermal conductivity of soil and have taken moisture content, bulk density salinity level, and organic substance as input parameters. The results showed that for the same salt type, thermal conductivity measurements for sand were greater than those for clay loam. Abu-Hamdeh et al. (2001) have done a comparison of soil heating and cooling techniques to find the soil’s thermal conductivity for sand, loam, sandy loam, and clay. They have observed that based the thermal conductivity based on heating techniques is slightly more than cooling method. Abu-Hamdeh (2003) analyzed the thermal characteristics of soils that are influenced by water content and density. The result showed that the sandy soil had lesser volumetric heat capacity and specific heat in comparison to clay soil at equal soil density and moisture contents. Cosenza et al. (2003) have established the relationship between volumetric water content and soil’s thermal conductivity. They have demonstrated through the use of a mathematical modelling approach that the relationship between volumetric water content and thermal conductivity of soil is affected by the microscopic structure of water. Nusier and Abu-Hamdeh (2003)Hamdeh (2003 used laboratory methods to find the thermal conductivity of soil such as sand and loam, and finally it has been observed that sand has a higher heat conductivity than loam. Malek et al. (2021) have done a laboratory work to find the impact of different soil properties on thermal conductivity of soil. The output showed that the soil’s water content has the greatest impact on its thermal conductivity. Dong et al. (2015) have reviewed thermal conductivity models for unsaturated soils for critical condition. The models that they have used are classified into three groups: empirical models, mixing models, and mathematical models. The typical thermal conductivity (k) can be arranged in decreasing order as kmineral > ksaturated soil > kwater > kdry soil > kair. Tokoro et al. (2016) analyzed how soil thermal conductivity is affected by water and propose a new soil thermal conductivity model and empirical equation. Farouki (1981) used various methods to find the thermal conductivity of soil in cold region. They observed that Johansen’s method gives better result over others. Hu et al. (2016) used the conduction–convection equation to estimate a Fourier series–based analytical approach in order to determine the soil’s temperature and thermal characteristics. The conclusion obtained that the Fourier series model shows better observed field temperatures on comparing with the sine wave model. Tong et al. (2016) presented a simple empirical model for calculating soil thermal conductivity. In comparison to other complicated models, this one is more useful for massive computations of surface energy and water fluxes. Rubin and Ho (2021) give a semi-analytical way to evaluate the soil’s thermal conductivity. It is one of the few models that currently exists that captures the physical geometry of different soil structure stages. Adrinek et al. (2022) gave the idea regarding the analytical computation of thermal conductivity based on sediment’s physical characteristics.

Some of the researchers have also used machine learning model in the geotechnical field. TaeBang et al. (2020) used machine learning approach to anticipate a model to compute thermal conductivity for compacted bentonite soil. Rizvi et al. (2020) used soft computing approach such as thermal lattice element method (TLEM) and deep neural network (DNN) in which TELM shows best performance. Kardani et al. (2021) proposed new computational approaches along with hybridization of two metaheuristic optimization techniques to predict the thermal conductivity of unsaturated soil. The result shows that ELM-IFF gives the highest possible prediction accuracy. Da et al. (2022) used several machine learning algorithms that are broadly applicable to find thermal diffusivity and thermal conductivity of bentonite. Random forest, gradient boosting algorithms, and extra trees regressor have advantage in creation of thermal conductivity model. Li et al. (2022a, b) implemented six machine learning algorithms to predict soil’s thermal conductivity. Outcome shows that adaptive boosting methods (AdaBoost) gives better predicted values. Li et al. (2022a, b) proposed a unified model based on artificial neural network algorithms, and the accuracy of the model is judged by evaluating it with current prediction models to compute thermal conductivity of soil. Mustafa et al. (2022) created four hybrid ANFIS models to find the factor of safety against overturning. They observed that ANFIS-PSO shows superior prediction across all four models. Wang et al. (2022) analyzed that based on artificial intelligence (AI), support vector machine (SVM) is the best model to predict soil’s thermal conductivity. Liu et al. (2023) made an experimental judgment between machine learning (ML) models and traditional parametric models. He observed that extreme gradient boosting (XGBoost) has better recreation accuracy and simplification ability. Mustafa et al. (2023) examined the gravity retaining wall’s resistance to bearing failure with the help of machine learning and concluded that deep neural network (DNN) performed the best. In this study, we have used three machine learning (ML) models such as extreme learning machine (ELM), adaptive neuro fuzzy inference system (ANFIS), and extreme gradient boosting (XGBoost) to predict thermal conductivity of soil.

Details of present studies

Soil’s thermal conductivity is an essential property which establishes how easily heat can flow through the soil. It is a key factor in understanding how energy is exchanged between the ground and the atmosphere. The soil’s thermal conductivity plays a vital role in geothermal energy systems, where heat is extracted from or injected into the ground. Soils with higher mineral content tends to have higher thermal conductivity because minerals are better heat conductors than organic matter or water. In this study, we have used an empirical equation of thermal conductivity of soil (Ks) which was given by Johansen (1975) and can be expressed as follows:

$${TC}_{s}={7.7}^{{Q}_{C}}{.2}^{\left(1-Qc\right)}$$
(1)

where Qc is the quartz content and TCS is soil’s thermal conductivity.

Using Eqs. (2) and (3), we can compute the saturated thermal conductivity (TCsat) and dry thermal conductivity (TCdry). By using the values of TCsat, TCdry, and KN, we can compute unsaturated thermal conductivity of soil (TCunsat).

$${TC}_{sat}={0.57}^{\upeta }.{TC}_{s}^{\left(1-\upeta \right)}$$
(2)
$${TC}_{dry}=\frac{137\upgamma +64.7}{2700-947\upgamma }$$
(3)
$${K}_{N}=\left\{0.7log\left(\frac{S}{100}\right)+1.0\right\}$$
(4)

where η is the porosity, \(\gamma\) is dry unit weight, S is the degree of saturation, and KN is Kersten number. The unsaturated thermal conductivity of soil can be computed as:

$${TC}_{\text{unsat}}=\left({TC}_{sat}-{TC}_{dry}\right){K}_{N}+{TC}_{dry}$$
(5)

In this study, we have considered four input parameters, namely, porosity (η), degree of saturation (S), dry density (γ), and Quartz content (Qc) and one output that is soil’s thermal conductivity in watts per meter Kelvin (W m−1 K−1).

Proposed AI-based model

Adaptive neuro fuzzy inference system (ANFIS)

The ANFIS is based on Takagi–sugeno fuzzy inference system. The Takagi–sugeno fuzzy inference system serves as the foundation for ANFIS, which is the addition of artificial neural network (ANNs), and fuzzy logic includes membership function and fuzzy logic operators. The accuracy of any ANFIS model depends upon the suitability of the data provided.

The goal of ANFIS model is in order to reduce the error between the expected and actual output from the training datasets and provide integrated learning capacity using its five-layer system. Furthermore, a new study recently did ANFIS-PSO (adaptive neuro fuzzy inference system-particle swarm optimization) for accurate prediction by Mustafa et al. (2022).

The following fuzzy inference relies on a fuzzy rule that was established for two inputs (A and B) and one output (M).

Rule1: If A is X1 and B is Y1, then,

$${d}_{1}={e}_{1}A+{f}_{1}B+{K}_{1}$$
(6)

Rule 2: If A is X2 and B is Y2, then,

$${d}_{2}={e}_{2}A+{f}_{2}B+{K}_{2}$$
(7)

where A and B are the two inputs and X1, X2, Y1, and Y2 are membership function (MF) linked with inputs A and B connected to node function. The parameters e1, f1, K1, and e2, f2, and K2 are linked with output functions d1 and d2. A usual ANFIS illustration is shown in Fig. 1. It demonstrates that the ANFIS model has five layers. Every layers’ functions is listed below:

Fig. 1
figure 1

Adaptive neuro fuzzy inference system

Layer 1: The input layers have been turned into a fuzzy membership function (MF) through this layer. Premise parameters are referred to as layer 1 parameters. The node’s output may be calculated as:

$${Q}_{j}^{1}={\mu }_{Xj}\left(A\right) \mathrm{for j}=1, 2, {Q}_{j}^{1}={\mu }_{Yj-2}\left(B\right) \mathrm{for j}=3, 4$$
(8)

where both A and B represent the node j’s inputs and Xj and Yj are fuzzy sets associated with node j in parameterized form.

Layer 2: The nodes in this level are fixed, serve as multipliers, and are referred to as neural network layers.

$${Q}_{j}^{2}={p}_{j}={\upmu }_{Xj}\left(A\right).{\upmu }_{Yj}\left(B\right) \mathrm{for j}=1, 2$$
(9)

Layer 3: Each node in layer 3 has been fixed and classified by H. Fuzzy inference method firing intensities are standardized. Each node that may be accessed in this level computes its weight which is standardized. The term “outputs” refers to normalized firing intensities.

$${Q}_{j}^{3}=\overline{{p }_{j}}=\frac{{p}_{j}}{{p}_{1}+{p}_{2}}\mathrm{for j}=1, 2$$
(10)

Layer 4: The nodes’ parameters in this layer can be adjustable. A parameter collection is present in this fuzzy logic node {ej, fj, kj} and the output of previous layer \(\overline{{p }_{j}}\).The node’s output is indicated as follows:

$${Q}_{j}^{4}=\overline{{p }_{j}}{d}_{j}=\overline{{p }_{j}}\left({e}_{j}A+{f}_{j}+{k}_{j}\right)\mathrm{ for j }= 1, 2$$
(11)

Layer 5: One particular node is fixed in this layer, and the output has been calculated as the sum of all entering signals. This node’s output function is calculated as follows:

$${Q}_{j}^{5}=\sum \overline{{p }_{j}}{d}_{j}=\frac{{\sum }_{j}{p}_{j}{d}_{j}}{{\sum }_{j}{p}_{j}}{Q}_{j}^{4}=\overline{{p }_{j}}\left({e}_{j}A+{f}_{j}B+{K}_{j}\right) \mathrm{for j}=1, 2$$
(12)

Extreme learning machine (ELM)

Extreme learning machine is a type of computational algorithm which comes from feedforward neural networks. In many types of application, i.e., classification, regression, clustering, and feature extraction tec. It was introduced to improve the accuracy of single-hidden-layer feedforward network (SLFNs) (Huang et al. 2011). ELM is a new AI-based model developed in the recent decade (Huang et al. 2006). Application of ELM is found in many types of programming languages, namely, MATLAB, Python, and R. ELM is a feedforward neural network used to resolve regression, clustering, and featuring learning with the help of a single layer of hidden nodes with the help of above-described relationship; ELM tries to understand the complex relation between variables and gives a good prediction. A pictorial view of ELM is shown in Fig. 2.

Fig. 2
figure 2

Extreme learning machine

For the formation of a classification model from a set of data (W = {mh ∊ Bu} (h = 1, 2, 3, …, t)) having parameters such as t samples and u input, it is simple to write the output of an ELM network with u inputs, d hidden neurons, and y output neurons.

$${Y}_{j}\left(h\right)={M}_{j}^{N}d\left(h\right)$$
(13)

where Mj ∊ Bd (h ∊ {1, 2, 3, … y}) is the weight vector; it touched with the hidden neurons to the jth output neurons; and output vector of hidden neurosis H(j) ∊ Bd, that is given by

$$H\left(h\right)=\left[f\left({x}_{1}^{N}m\left(h\right)+{r}_{1}\right),f\left({x}_{2}^{N}n\left(h\right)+{r}_{2}\right),\dots ,f\left({x}_{d}^{N}\left(h\right)+{r}_{d}\right)\right]$$
(14)

where rv (v = 1, 2, …, d) and xV ∊ Bu are the bias of the vth hidden neurons and weight factor of vth hidden neurons respectively, and f (⦁) indicates the initiation function (sigmoidal function). It was observed that the bias (rk) and weight (xv) vectors are produced in a casual means from which the hidden layer output matrix (T) can be formed. The weight matrix is then planned by the “Moore–Penrose pseudo inverse” method, which is given by:

$$X={\left(T\times {T}^{N}\right)}^{-1}T\times {Q}^{N}$$
(15)

where Q = [e (1), e (2) …, e (t)] shows a matrix of g⨯t dimensions and its hth column is real largest vector e (h) ∊ Bg. The class label for new input parameter can be calculated as:

$$O=argmax\left\{{D}_{j}\right\}\left(j=1,2,\dots ,D\right)$$
(16)

where O represents predicted class label. Other detail working guidelines of ELM can be seen from Zhang et al. (2021) (Fig. 3).

Fig. 3
figure 3

Extreme gradient boosting

Extreme gradient boosting (XGBoost)

XGBoost (extreme gradient boosting) is a powerful computational algorithm which is designed to improve the accuracy of a previous model, and it employs a complicated function to prevent overfitting. Applications of XGBoost are found in many types of programming language such as Python, R, Java, and Julia. Due to its high accuracy and speed, XGBoost is very popular in the field of data science (Chen and Guestrin 2016). It gives best performance in developing prediction models for DNA replications and DNA sequence in general (Do and Le., 2020), forecasts glucose concentration with the help of robust machine learning algorithm (Mekonnen et al. 2020), and resolves Walmart’s sales problem using XGBoost algorithm and meticulous features engineering (Niu., 2020). The exact formula for the objective loss function of XGBoost approach is given below:

$${\text{Objective}}\left(a\right)={\sum }_{j=1}^{q}p\left({Y}_{j},{Y}_{j}^{a}\right)+\Omega \left({f}_{a}\right)+As$$
(17)
$${Y}_{j}^{a}={Y}_{j}^{a-1}+{f}_{a}\left({z}_{j}\right)$$
(18)
$$\Omega \left({f}_{a}\right)=\uprho {X}_{a}+\frac{1}{2}\lambda {\Vert m\Vert }^{2}$$
(19)

where Xa indicates the number of leaves there are on the tree. A is constant, zj is the input vector, and λ are ρ hyperparameters. Yj stands for the mean the actual value and \({Y}_{{\text{j}}}^{{\text{a}}}\) stands for the forecast value of sales, respectively, similarly. P (⦁) and \({f}_{{\text{a}}}\) (⦁) stands for square loss function and a regression tree, respectively. Based on Taylor’s formula, function of an object (⦁) is roughly stated in the following ways:

$${\text{Object}}\left(a\right)\approx {\sum }_{j=1}^{q}[p({Y}_{j},{Y}_{j}^{a-1}+{e}_{j}{f}_{a}\left({z}_{a}\right)+\frac{1}{2}{k}_{j}{f}_{a}^{2}\left({z}_{j}\right)]+\Omega \left({f}_{a}\right)+A$$
(20)

where ej and \(\frac{1}{2}{k}_{j}\) expressed as a first and quadratic term’s coefficients in the Taylor expansion.

Dataset preparation

With the help of mean and standard deviation, input data are generated. Input variables, namely, quartz content (QC), degree of saturation (S), porosity (ɳ), and dry density (γ) of coarse soil are considered to compute the output (thermal conductivity of soil). For this purpose, mean and standard deviation of coarse soil input variable were taken from past research’s Tokoro et al. (2016). The statistical depictions of input parameters are given in Table 1.

Table 1 Statistical depiction of dataset

A total of 110 dataset were taken, and after that, the input (QC, S, η, and γ) and output (thermal conductivity of soil) variables have been normalized. Before the consideration of the machine learning models, there is an important stage, i.e., normalization of datasets. Normalizations is conducted to remove dimensionality influence of the variable. That is why for the generation of any model, the output and input were normalized among 0 and 1 using the following equation:

$${M}_{\text{normalized}}=\frac{M-{M}_{{\text{min}}}}{{M}_{{\text{max}}}-{M}_{{\text{min}}}}$$
(21)

where Mmax and Mmin are the maximum and minimum values of the parameter (M) taken into account, respectively. After that the dataset obtained by normalization is split into two parts, i.e., training (TR) phase and testing (TS) phase. Training phase is done by 70% (77 cases) of the dataset of the normalized value and testing phase is done by 30% (33 cases) of the dataset of the normalized value. The methodology flowchart is represented in Fig. 4.

Fig. 4
figure 4

Flowchart of methodology

Statistical performance parameter

The prediction accuracy of proposed machine learning models such as extreme learning machine (ELM), adaptive neuro fuzzy inference system (ANFIS), and extreme gradient boosting (XGBoost) was examined by calculating numerous statistical performance parameters and evaluated graphically (i.e., scatter plot). We used various trend measuring statistical performance parameters like coefficient of determination (R2), a-20 index, Willmott’s index of agreement (WI), variance account factor (VAF), Nash–Sutcliffe efficiency (NS), and error measuring parameters like mean absolute error (MAE), root mean square error (RMSE), scatter index (SI), RMSE-observation standard deviation ratio (RSR), and weighted mean absolute percentage error (WMAPE).

$${R}^{2}=\frac{{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-\overline{{TC }_{obs}}\right)}^{2}-{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-{TC}_{{\text{pred}},i}\right)}^{2}}{{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-\overline{{TC }_{obs}}\right)}^{2}}$$
(22)
$$a-20{\text{Index}}=\frac{g20}{r}$$
(23)
$$WI=1-\left[\frac{{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-{TC}_{{\text{pred}},i}\right)}^{2}}{{\sum }_{i=1}^{r}{\left(\left|{TC}_{\text{pred}}-\overline{{TC }_{obs}}\right|+\left|{TC}_{obs,i}-\overline{{TC }_{obs}}\right|\right)}^{2}}\right]$$
(24)
$$VAF=\left(1-\frac{var\left({TC}_{obs,i}-{TC}_{{\text{pred}},i}\right)}{var\left({TC}_{obs,i}\right)}\right)\times 100$$
(25)
$$NS=1-\frac{{\sum }_{i=1}^{r}{\left({TC}_{obs}-{TC}_{\text{pred}}\right)}^{2}}{{\sum }_{i=1}^{r}{\left({TC}_{obs}-\overline{{TC }_{obs}}\right)}^{2}}$$
(26)
$$MAE=\frac{1}{r}{\sum }_{i=1}^{r}\left|\left({TC}_{{\text{pred}},i}-{TC}_{obs,i}\right)\right|$$
(27)
$${\text{RMSE}}=\sqrt{\frac{1}{r}{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-{TC}_{{\text{pred}},i}\right)}^{2}}$$
(28)
$$SI=\frac{\sqrt{\frac{1}{r}{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-{TC}_{{\text{pred}},i}\right)}^{2}}}{\overline{{TC }_{obs}}}=\frac{\text{RMSE}}{\overline{{TC }_{obs}}}$$
(29)
$$RSR=\frac{\sqrt{\frac{1}{r}{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-{TC}_{{\text{pred}},i}\right)}^{2}}}{\sqrt{\frac{1}{r}{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-\overline{{TC }_{obs,i}}\right)}^{2}}}=\frac{\text{RMSE}}{\sqrt{\frac{1}{r}{\sum }_{i=1}^{r}{\left({TC}_{obs,i}-{TC}_{{\text{pred}},i}\right)}^{2}}}$$
(30)
$${\text{WMAPE}}=\frac{{\sum }_{i=1}^{r}\left|\frac{{TC}_{obs,i}-{TC}_{{\text{pred}},i}}{{TC}_{obs,i}}\right|\times {TC}_{obs,i}}{{\sum }_{i=1}^{r}{TC}_{obs,i}}$$
(31)

where TCobs,i and TCpred,i are the actual and predicted thermal conductivity of ith values, respectively, \({\overline{{\text{TC}}} }_{{\text{obs}}}\) is the mean thermal conductivity of actual value, r is the total number of training and testing samples, g20 is the number of data that has both observed and predicted value ratios that lie between 0.80 and 1.20. With the help of the above equations, the model having the least value of RMSE, SI, MAE, RSR, and WMAPE and higher value of R2, a-20 Index, VAF, WI, and NS have superior prediction ability.

Results and analysis

Prediction power

In this portion, calculation of various performance parameters is done, which will be used in deciding the best predictive model. In this study, three machine learning models are utilized to predict thermal conductivity (unsaturated condition) of soil. The three machine learning models, namely, ANFIS, ELM, and XGBoost, are used to predict thermal conductivity of soil. Numerous performance parameters were computed to assess these models’ performance. Performance parameter is further divided into two forms, that is, the trend measuring parameters (R2, a-20 index, VAF, WI and NS) and error measuring parameters (RMSE, MAE, SI, RSR, and WMAPE). The results of the performance parameters are listed in Table 2 (TR dataset) and Table 3 (TS dataset). Based on the performance parameters, ANFIS has been found to have more accurate prediction power throughout the training with higher value of R2 = 0.9913, a-20 index = 0.9480, VAF = 99.1554, WI = 0.9978, and NS = 0.9913 and lower value of RMSE = 0.0171, MAE = 0.0050, SI = 0.0538, RSR = 0.0931, and WMAPE = 0.0157, whereas the same is decreased in the testing phase (R2 = 0.9762, a-20 index = 0.7879, VAF = 97.7848, WI = 0.9941, NS = 0.9762,RMSE = 0.0334, MAE = 0.0204, SI = 0.1104, RSR = 0.1543, WMAPE = 0.0674). The performance of XGBoost model (R2 = 0.9804, a-20 index = 0.9351, VAF = 98.0956, WI = 0.9952, NS = 0.9804, RMSE = 0.0257, MAE = 0.0132, SI = 0.0808, RSR = 0.1398, WMAPE = 0.0409) is lower than ANFIS model in the training phase, while in the testing phase, it has higher predicting power (R2 = 0.9901, a-20 Index = 0.8788, VAF = 99.0228,WI = 0.9975,NS = 0.9901, RMSE = 0.0215,MAE = 0.0120, SI = 0.0712, RSR = 0.0995, WMAPE = 0.0032). Based on these performance parameters, we may conclude that XGBoost performs better during testing phase, while ANFIS performs better during training phase. In training stage, ANFIS model outperformed the other models by far, while the XGBoost performed better in the testing phase. The trend is reversed due to lesser data in testing phase, as we know that in machine learning model, the more data you provide to the ML system, the faster that model can learn and improve.

Table 2 Comparison of model prediction power for TR dataset
Table 3 Comparison of model prediction power for TS dataset

Rank analysis

In this section, ranking of the proposed models (ANFIS, ELM, and XGBoost) is done according to the performance indicators. The value which is closer to the ideal value of the performance parameters is ranked 1 and the farthest value ranked 3, as we have used three models. To get the overall rank of the model, ranks of the training phase and testing phase are added. The models which have the lower overall rank will be the best model and model of higher overall rank is worst model. Ranking of the models is shown in Table 4. In the rank analysis, we have found that ANFIS gives best results in training phase (ANFISTR = 12) and XGBoost in the testing phase (XGBoostTS = 11). Overall ranks of the models are as follows: ANFIS = 34, ELM = 47, XGBoost = 38. So, it can be clearly seen that ANFIS model gives the better prediction in comparison of ELM and XGBoost, as ANFIS has the lower rank among the three models that are used in this study. The graphical representation of rank analysis is also represented by radar diagram which is shown in Fig. 5.

Table 4 Rank analysis of the models (training and testing)
Fig. 5
figure 5

Radar diagram

Reliability analysis

Thermal conductivity of soil cannot be considered directly as accurate, as numerous uncertainties can possibly complicate it. This inaccuracy can be related to the soil properties or analytical parameter used. To identify unpredictability and to get a dependable method for predicting the thermal conductivity, reliability analysis has been done. To evaluate the reliability index (β), the first-order second moment (FOSM) approach has been used. In FOSM method, reliability index is computed with the help of average value (µ) and standard deviation (σ) of the performance function and can be computed as:

$$\upbeta =\frac{\upmu -1}{\upsigma }$$
(32)

where µ and σ is the average value and standard deviation of thermal conductivity, respectively. The probability of failure (Pf) is directly dependent on the reliability index (β) on the assumption that all the normal variables are distributed normally. The probability of failure can be determined as:

$${P}_{f}=1-\Phi \left(\upbeta \right)$$
(33)

where Φ (β) denotes the standard normal cumulative probability. By using FOSM technique, the reliability index (β) of the models has been calculated (using Eq. 32), and comparison is done with the reliability index of the model with the actual value. A greater value of β shows a superior performance of the model. The greater value of reliability index and lesser value of the probability of failure (Pf) displays that ANFIS performs best among all three models (as shown in Table 5). Hence, the models are ranked on the basis of β and Pf. The model with greater value of β and lesser value of Pf will be ranked 1 as it shows better performance, and the model with lesser value of β and greater value of Pf will be ranked 3 as it shows worst performance. Also it can be observed that the ANFIS model’s reliability index and probability of failure values closely match the existing value. As a result, the ANFIS model may be trusted to be reliable for analyzing the thermal conductivity of soil.

Table 5 Ranking of models according to the reliability index (β)

Error matrix

An error matrix, also known as a confusion matrix, is a table used to evaluate the performance of the models that are used in this study. It is commonly used in machine learning. It provides a summary of the predictions made by the model with the observed outcomes. By analyzing the error matrix, various evaluation matrices can be derived (accuracy, precision) to check the performance of the models. These matrices can help in understanding the strength and weaknesses of the model. Additionally, it conveys the idea of the maximum and minimum error of the predicted model. The error matrix is generated according to the ideal values of the performance parameters. In this study, ten performance parameters are used, viz., R2, a-20 index, VAF, WI, NS, RMSE, MAE, SI, RSR, and WMAPE, where R2, a-20 index, VAF, WI, and NS are trend measuring parameters and RMSE, MAE, SI, RSR, and WMAPE are error measuring parameters.

$${E}_{e}=\left|\left(1-\left|P\right|\right)\right|\times 100$$
(34)
$${E}_{t}=\frac{\left(I-\left|P\right|\right)}{I}\times 100$$
(35)

where \({E}_{{\text{e}}}\) and \({E}_{{\text{t}}}\) represent the error in error measuring parameters (EMPs) and error in trend measuring parameter (TMP), respectively; I is the ideal value of trend and error measuring parameter; and P is the value of performance parameters estimated for error measuring parameter and trend measuring parameter. Calculation of error for error measuring and trend measuring parameter estimated by using Eqs. 3435. Tables 6, 7 and 8 represent the calculation of error for trend measuring parameter and Table 7, 8 and 9 show the calculation of error for error measuring parameters.

Table 6 Calculation of error for trend measuring parameter (TR dataset)
Table 7 Calculation of error for error measuring parameter (TR dataset)
Table 8 Calculation of error for trend measuring parameter (TS dataset)
Table 9 Calculation of error for error measuring parameter (TS dataset)

In this section, error is calculated for trend measuring parameter (TMP) and error measuring parameter (EMP). Error matrix for trend and error measuring parameter for both TR and TS dataset is shown in Figs. 6 and 7, respectively. TMP are R2, a-20 index, VAF, WI, and NS and EMP are RMSE, MAE, SI, RSR, and WMAPE considered in this study. The lowest error has been shown by green color, moderate error by yellow color, and highest error by red color. As per the error matrix of the thermal conductivity of soil of trend measuring and error measuring parameter, ANFIS model gives less error than the other two models used in this study in the training phase. In the TS phase, XGBoost performs best as it has less error on comparing with other two models.

Fig. 6
figure 6

Error matrix for trend measuring parameters (TR and TS dataset)

Fig. 7
figure 7

Error matrix for error measuring parameters (TR and TS dataset)

Regression curve

Regression curve is drawn among the predicted value and observed value of soil’s thermal conductivity. Regression curve is also known as R-curve. It provides the calculated R-value that is displayed in Table 3 and 4. Observed thermal conductivity of soil is shown on x-axis, and predicted soil’s thermal conductivity is shown on y-axis. Both the datasets of TR and TS phase is shown in the same plot. In Fig. 8, dotted lines indicate ± 10% deviation between the actual line and the projected data. It can be observed from the curve that the observed value and predicted value overlaps in the case of ANFIS and XGBoost model, but slight deviation can be seen in the ELM model.

Fig. 8
figure 8

Scatter plot for models a ANFIS; b ELM; and c XGBoost

William’s plot

William’s plot is drawn between standardized residual and leverage, in which standardized residual is plotted on the ordinate and leverage on the abscissa. In order to determine whether an AI model is capable of producing accurate predictions, it is important to evaluate the applicability domain of three AI models that are considered in this study. By outlining the leverage (h) values of the TR and TS datasets, the three separate AI models’ applicability areas were identified. The William’s plot for both TR and TS dataset is shown in Fig. 9. A squared zone with a leverage threshold h* and a range of ± 3 standard deviations surround the applicability of this domain. The leverage threshold (h*) is normally fixed at 3 p/m, where m is the total number of training compound and p is the number of model parameters plus one (m + 1). The value of leverage threshold is taken as 0.195. From the plot, it can be seen that all the compounds had leverage lower than the threshold leverage of the TR and TS datasets, but despite that the leverage is less than threshold leverage, there is some value in the plot which lies outside the square boundary, i.e., two training and two testing in the case of ANFIS, one training and one testing for XGBoost, and for the ELM four testing and three training datasets.

Fig. 9
figure 9

William’s plot for models a ANFIS; b ELM; and c XGBoost

Sensitivity analysis

The impact on the output (thermal conductivity of soil) due to each input parameters (Qc, η, S, γ) is calculated for all the three models (namely, ANFIS, ELM, and XGBoost). A parameter called strength of relation is calculated to know the impact on thermal conductivity of soil. It can be calculated as:

$$SOR=\frac{{\sum }_{i=1}^{r}{R}_{m,i}{S}_{n,i}}{\sqrt{{\sum }_{i=1}^{r}{\left({R}_{m,i}\right)}^{2}{\sum }_{i=1}^{r}{\left({S}_{n,i}\right)}^{2}}}$$
(36)

where Rm,i denotes the ith value of mth independent variable, r denotes the total number of observation dataset, and m denotes the total number of input parameters considered, and Sn,i denotes the ith value of nth dependent variable. Here, to know the impact of each input parameters, (m = 4, n = 1, r = 110) has been considered. The greater value of strength shows greater influence of input parameters on the output. The strength of relation of numerous input parameters will be different according to their impact on output parameter as shown in Fig. 10. From the figure, it can be observed that the dry density of soil (γ) is the most influential parameter and quartz content (Qc) is the least influential parameter.

Fig. 10
figure 10

Strength of relation of input parameters

Conclusion

In this project, three predictive models were developed to compute the thermal conductivity of soil, namely, ANFIS, ELM, and XGBoost, using four input parameters of soil (Qc, η, S, γ). We compared all three models with several parameters like R-curve, rank analysis, and error matrix, and as a result, it can be seen that ANFIS shows better outcome amongst all three models (ANFIS, ELM, and XGBoost) in the training stage, and XGBoost shows the best prediction ability in the testing stage to determine thermal conductivity of soil amongst all three models. The total rank achieved by ANFIS model in the testing stage is 22 and did better than the two models by a long shot, while the total rank achieved by ELM model is 27 and XGBoost model is 11 in testing phase. But, when evaluation is done according to overall rank (training stage rank + testing stage rank) gained by all models, the uppermost predictive exactness is displayed by ANFIS with the final rank of 34, which is traced by ANFIS, XGBoost, and ELM. Reliability index (β) and probability of failure (Pf) were calculated too for all three models and matched with existing value. A greater value of β and lesser value of Pf is given by ANFIS model amongst all the three models. Other than this, the radar diagram has also drawn to show the appropriateness of model. Sensitivity analysis is also performed to know the influence of each input parameters on the output. For this, strength of relation is calculated to know the impact of each input parameters on the output (thermal conductivity of soil). From sensitivity analysis it can be observed that dry density of soil (γ) is the most dominant parameter among all four input parameters, followed by porosity (η), degree of saturation (S), and quartz content (Qc).