Introduction

Fluid cooling or heating is very important for many industrial applications. Improving the thermal properties of heat transfer fluids can be a method for heat transfer. Nanofluids have attracted the attention of many scientists in recent years [1,2,3,4,5,6,7,8]. For example, small amounts of nanoparticles in working fluids increase the thermal conductivity of these fluids. To improve the thermal conductivity of liquids, researchers have investigated the thermo-physical properties of nanofluids. Recently, ANNs are used in many scientific projects. ANNs are systems that are inspired by the human brain and the process of data is similar to biological neurons. The process of learning in such systems happens by examples. ANNs are often including connected units which are called neurons. The birth of ANNs returns to the 1940s by Warren McCulloch and Walter Pitts. In the 1940’s, the Hebbian learning algorithm was proposed by Hebb was an unsupervised learning method. In 1954, Hebbian networks were simulated by Farley and Clark AND Holland, Rochester, Holland, Habit and Duda worked on computational methods [9,10,11,12,13,14,15]. In many cases, these ANNs can predict the behavior of the systems. One of the most critical aspects of these ANNs is in predicting the behavior of nonlinear systems or complex systems.

Jamal-Abadi et al. [16] optimized the \(k_{\text{nf}}\) of Al2O3/water nanofluid using ANN. Their results show the maximum enhancement of 42% for \(k_{\text{nf}}\). Tahani et al. [17] modeled the \(k_{\text{nf}}\) of GO/water nanofluid using ANN method. Their results indicate that the ANN can predict the \(k_{\text{nf}}\). Khosrojerdi et al. [18] modeled the \(k_{\text{nf}}\) of GO/water nanofluid by MLP-ANN. Their results show high accuracy of ANN modeling for predicting the \(k_{\text{nf}}\). Esfe et al. [19] used an ANN model to predict the \(k_{\text{nf}}\) of MWCNT–water nanofluid. They concluded that ANN could predict the \(k_{\text{nf}}\) more accurately. Aghayari et al. [20] compared the experimental and predicted data for the \(k_{\text{nf}}\) of Fe3O4/water nanofluid using ANNs. Afrand et al. [21] predicted the effects of MgO concentration and temperature on the thermal conductivity of water by using of ANN approach. Comparisons revealed that the ANN approach was accurate. Zhao and Li [22] predicted the \(k_{\text{nf}}\) of Al2O3–water nanofluids using ANNs. They found that the ANN provides an effective way to predict the properties of this type of nanofluids. Esfe et al. [23] evaluated the properties of EG-ZnO-DWCNT with ANN. Their results show the accuracy of ANN in modeling the \(k_{\text{nf}}\). Aghayari et al. [24] modeled the electrical conductivity of CuO/glycerol nanofluids. Kannaiyan et al. [25] modeled the \(k_{\text{nf}}\) of Al2O3/SiO2–water nanofluid using ANNs. They found that the predicted \(k_{\text{nf}}\) is satisfactory. Zendehboudi and Saidur [26] obtained a model to estimate the \(k_{\text{nf}}\) of 26 nanofluids under different situations.

In this study, after generating experimental data points for different volume fraction of nanoparticles \((\phi = 0,0.1,0.5,1,1.5,2,3\;{\text{and}}\;5\% )\) and different temperatures (25, 30, 35, 40, 45 and 50 °C), an algorithm to find the best neuron number in the hidden layer of ANN method is proposed to find the best architecture and then to predict the \(k_{\text{nf}}\) of SiO2/water–ethylene glycol (50:50) hybrid Newtonian nanofluid. Then, using the fitting method, a surface is fitted on the experimental data points. According to the authors’ research, there is no investigation with ANN into the \(k_{\text{nf}}\) of this type of nanofluids.

Experimental

In this study, a mixture of 60 to 40 volumes of water and ethylene glycol was used. The silica nanoparticles were suspended in water and ethylene glycol mixture (made by Germany’s Merk Corporation). The nanofluid is stabilized by combining the chemical and mechanical methods in the different \(\phi\). A minimum of 5 h of ultra-sonication is used to stabilize the nanofluid. This nanofluid is made of 7 different volume fraction of nanoparticles (0, 0.1, 0.5, 1, 1.5, 2, 3 and 5%) and different temperatures (25, 30, 35, 40, 45 and 50 °C). After suspending the nanoparticles, the \(k_{\text{nf}}\) is measured with a KD2 probe. The physical and chemical properties of materials are presented in Tables 1, 2, and 3.

Table 1 Ethylene glycol properties
Table 2 Water properties
Table 3 Properties of Silica nanoparticles

After making nanofluids, each sample was monitored for three days with no deposition and settling. Figure 1 shows the variation of \(k_{\text{nf}}\) versus \(\phi\) at all experimental temperatures. As can be seen, the changes of \(k_{\text{nf}}\) at all temperatures have a similar general shape. It is observed that in a lower \(\phi\), the slope of \(k_{\text{nf}}\) is greater. This behavior is due to the fact that increasing the \(\phi\) increases the probability of localization and consequently decreasing the specific surface area. At 25 °C, by increasing the \(\phi\) from \(\phi = 0.1\) to \(\phi = 5\%\), the \(k_{\text{nf}}\) increases from 3.7 to 38.4% and the highest \(k_{\text{nf}}\) occurs at the highest volume fraction. At 30 °C, by increasing \(\phi\) from 0.1 to 5%, the \(k_{\text{nf}}\) increases from 3.9 to 31% and the highest \(k_{\text{nf}}\) occurs at the highest \(\phi\). At 35 °C, by increasing the \(\phi\) from 0.1 to 5%, the \(k_{\text{nf}}\) increases from 4.1 to 41.9% and the highest \(k_{\text{nf}}\) occurs at the highest \(\phi\). At 40 °C, by increasing the \(\phi\) from 0.1 to 5%, the \(k_{\text{nf}}\) increases from 4.3 to 43.1% and the highest \(k_{\text{nf}}\) occurs at the highest \(\phi\). At 50 °C, by increasing the \(\phi\) from 0.1 to 5%, the \(k_{\text{nf}}\) increases from 5.1 to 50.9% and the highest \(k_{\text{nf}}\) occurs at the highest \(\phi\).

Fig. 1
figure 1

\(k_{\text{nf}}\) versus \(\phi\)

Figure 2 shows the changes in the \(k_{\text{nf}}\) versus temperature at the different \(\phi\). At \(\phi = 0.1\%\), the \(k_{\text{nf}}\) increases 6.6% with increasing temperature from 25 to 50 °C. The highest \(k_{\text{nf}}\) is related to T = 50 °C. At \(\phi = 0.5\%\), with increasing temperature from T = 25–50 °C the \(k_{\text{nf}}\) increases 9.9%. The \(k_{\text{nf}}\) increases by 9.9% compared to the \(k_{bf}\) at T = 25 °C.

Fig. 2
figure 2

\(k_{\text{nf}}\) versus temperature

At \(\phi = 1\%\), with increasing temperature from T = 25–50 °C, the \(k_{\text{nf}}\) increases 14.9%. At \(\phi = 1.5\%\) at 50 °C, the \(k_{\text{nf}}\) increased by 13.7%. At \(\phi = 2\%\), the \(k_{\text{nf}}\) increases 27.8% with increasing temperature from 25 to 50 °C. In \(\phi = 3\%\), with increasing temperature from T = 25–50 °C, the \(k_{\text{nf}}\) increases 8.9%. At \(\phi = 5\%\), with increasing temperature from T = 25–50 °C, the \(k_{\text{nf}}\) increases 45.5%.

As the temperature increases, the movement of the nanoparticles increases, resulting in a higher \(k_{\text{nf}}\). The increase in \(k_{\text{nf}}\) with increasing temperature and \(\phi\) can be attributed to the weakening of the molecular bonds in the fluid layers as well as the increase in the collision between the nanoparticles.

ANN method

ANNs are used for modeling the behavior of nanofluid. ANN is widely used in predicting the behavior of system especially nonlinear systems. But in this paper, an optimized ANN for predicting the \(k_{\text{nf}}\) is reached by changing different neuron numbers in the hidden layer and then comparing the performances and selecting the best neuron number. In fact, the architecture of ANN is modified to obtain the best neuron number for predicting the \(k_{\text{nf}}\). But some terms and descriptions about the basics of ANN has been presented as follows.

Mean square error (MSE) is presented as follows,

$${\text{MSE}} = \frac{1}{N}\sum\limits_{{{\text{k}} = 1}}^{N} {(Y_{\text{k}}^{\text{ANN}} - Y_{\text{k}}^{\exp } )^{2} }$$
(1)

In Eq. 1, \(N\) is the number of data points, \(Y_{\text{k}}^{\text{ANN}}\) is the output and \(Y_{\text{k}}^{\exp }\) is the experimental value. In ANNs, the data are predicted by Eq. 2,

$$y_{\text i} = \varPhi \left( {\sum\limits_{j = 1}^{m} {w_{\text{ij}} x_{\text{j}} + b_{\text{i}} } } \right)$$
(2)

In Eq. 2, \(y_{\text{i}}\) is the output, \(\varPhi\) is activation function, \(w_{\text{ij}}\) is the weighting matrix, \(x_{\text{j}}\) is input and \(b_{\text{i}}\) is the bias. In this designed ANN, the activation function of input and hidden layer is tansig which is introduced in Eq. 3. Also, the activation function of output layer is purelin.

$$\tan sig(n) = \frac{2}{{1 + {\text{e}}^{ - 2{\text n}} }} - 1$$
(3)

This ANN is a feed-forward network with Levenberg–Marquardt or damped least square for the learning algorithm. This algorithm was firstly introduced in 1944. The experimental data points are randomly divided into train, validation, and test parts. In the current work, there are 48 data points. 70% of data set is categorized as train, 15% for validation, and 15% for test data. Train data points are used to train the network; meanwhile, validation data points are used to modify the training process, and at the final step, the test data points are used to calculate the performance of the network. This ANN predicts the \(k_{\text{nf}}\) of aforementioned nanofluid. Therefore, only one neuron is used in the output layer. Obviously, the results of ANN depend on the neuron numbers of the hidden layer. This algorithm tries different neuron number and calculates the performance, and finally, the best ANN is selected as the best answer. In addition, in this algorithm, an inner iteration is defined to increase the reliability of the ANN. In fact, this inner iteration calculates the performance of each neuron number, and then, the algorithm generated another ANN and then calculates the performance, and at the final step on inner iteration, the mean value of performance for that neuron number is considered as the performance. This algorithm is shown in Fig. 3.

Fig. 3
figure 3

Algorithm to find the best neuron number of hidden layer

In this work, the number of inner iterations is considered 10. Different neuron numbers have been tested (from 6 to 35). The sorted neuron numbers based on the performances are presented in Table 4.

Table 4 the sorted performance of ANN’s

Considering Table 4, it can be seen that the ANN with 8 neurons has the best performance. The correlation is defined as:

$$\rho X,Y = \frac{{\text{cov} (X,Y)}}{{\sigma_{\text{x}} \sigma_{\text{y}} }} = \frac{{E\left[ {(X - \mu_{\text{X}} )(Y - \mu_{\text{Y}} )} \right]}}{{\sigma_{\text{X}} \sigma_{\text{Y}} }}$$
(4)

In Eq. 4, \(E\) is the expected value, \(\text{cov}\) is covariance, \(\mu_{\text{X}}\) is the mean value of \(X\) and \(\mu_{\text{Y}}\) is the mean value of \(Y\) and \(\sigma_{\text{X}} ,\sigma_{\text{Y}}\) are standard deviations of \(X,Y\). The correlation coefficients for train, validation, test and all dataset are presented in Table 5. The correlation coefficient indicates that how closely data are along a straight line. The closer that the correlation coefficient to one, the better data points are aligned. Data sets with correlation coefficient close to zero show no first-order line relationship. The results of Eq. 4 for different neuron numbers are presented in Table 5. It can be seen that the ANN with 8 neurons in the hidden layer has the best correlation for overall data and its value is close to one.

Table 5 The correlation coefficients of ANN outputs for different neuron numbers

The correlation coefficient indicates that how closely data are along a straight line. The closer that the correlation coefficient to one, the better data points are aligned. Data sets with correlation coefficient close to zero show no first-order line relationship. The results of Eq. 5 for different neuron numbers are presented in Table 5. It can be seen that the ANN with 8 neurons has the best correlation for overall data and its value is close to one. In many ANNs, the data points are divided into three main categories including train validation and test randomly. Then these data points are used for the network. Train data trains the network and generates masses and biases of the network. The validation data points modify the masses and biases and finally the test data points are used to evaluate the performance of the network. In Table 2 different neuron numbers have been tested and these neuron numbers are sorted based on their test performance. On the top of Table 2 it can be seen that the best neuron number is 8 because it has the best performance.

The train, validation, and test figures of ANN are presented in Figs. 4 to 7. Figure 4 shows ANN train outputs. In ANN train data outputs, MSE is 1.3040e−06, and maximum absolute value of error is 0.0027.

Fig. 4
figure 4

ANN train outputs

Figure 5 shows ANN validation outputs. In ANN validation outputs, MSE is 2.5222e−05 and maximum absolute value of error is 0.0073.

Fig. 5
figure 5

ANN validation outputs

Figure 6 shows ANN test outputs. In ANN test outputs, MSE is 1.9024e−05 and maximum absolute value of error is 0.0084.

Fig. 6
figure 6

ANN test outputs

Figure 7 shows ANN all data outputs. In ANN all outputs, MSE is 7.3763e−06 and maximum absolute value of error is 0.0084. It can be seen that train, validation and test data points are accurately predicted by the ANN. In fact, this network can predict the \(k_{\text{nf}}\) for different temperatures and different \(\phi\). In the next part, another method for predicting the \(k_{\text{nf}}\) has been explained.

Fig. 7
figure 7

ANN all data outputs

Surface fitting

In this part, since there are two inputs (\(\phi\) and temperature), we can fit a surface on experimental data points to predict the \(k_{\text{nf}}\). The fitted surface is shown in Fig. 8.

Fig. 8
figure 8

The fitted surface

Different orders for functions have been tested. But, the third order had significantly better results compared to the second order. Although, the fourth order showed better results but there were no considerable differences between the third-order and the fourth-order results. In addition, the fourth-order function had more coefficients. Therefore, to avoid complexity in the formula, the third-order function is selected. Regarding the experimental data points, a third-order function is obtained. The fitted surface is presented in Eq. 5,

$${\text{Fitresult}}\,(x,y) = p00 + p10*x + p01*y + p20*x^{2} + p11*x*y + p30*x^{3} + p21*x^{2} *y$$
(5)

In Eq. 5, x is purity and y is temperature and \({\text{Fitresult}}\left( {x,y} \right)\) is the fitted surface. The coefficients of the fitted surface are presented in Table 6. In the fitting method, the MSE is 2.7547e−05, and the maximum value of error is 0.0125. The correlation coefficient of the fitting method is 0.9919. In Fig. 9 the experimental data points, ANN outputs and fitting results have been shown.

Table 6 the coefficients of the fitted surface
Fig. 9
figure 9

Experimental, ANN and fitting results

In Fig. 10, the absolute values of errors of ANN and fitting method have been compared. It can be seen that the ANN method has smaller absolute values of errors compared to the fitting method. In addition, the maximum error value of the ANN method is smaller than half of the fitting method. In Fig. 10 it can be seen that the ANN had smaller errors compared to the fitting method.

Fig. 10
figure 10

Errors of ANN and fitting method

Conclusions

Based on the presented results, it can be concluded that the ANN had better ability in predicting the \(k_{\text{nf}}\) for \(\phi = 0,0.1,0.5,1.5,2,3,5\%\) and T = 25, 30, 35, 40, 45, 50 °C. Also, ANN showed better performance and better correlation and also smaller error in most of the predicted data points. Based on the presented results, it can be seen that:

  • The best neuron number is 8.

  • The correlation coefficient of the fitting method is 0.9919.

  • In the fitting method, the MSE is 2.7547e-05, and the maximum value of error is 0.0125.

  • The ANN had ability in predicting thermal conductivity for volume fraction of particles range (0, 0.1, 0.5, 1.5, 2, 3, 5%) and temperature range (25, 30, 35, 40, 45, 50 °C).

  • ANN method has smaller absolute values of errors compared to the fitting method.

  • The maximum error value of the ANN method is smaller than half of the fitting method.

Therefore, it can be said that the designed ANN had showed better results compared to the surface fitting method and the results were reliable and accurate. Lab costs for generating experimental data are high. By using such a network, the thermal conductivity of this nanofluid can be obtained and decreases the experimental costs.