1 Introduction

Urbanisation is an unquestionable trend in developing countries. The construction of infrastructure systems, as well as buildings, is an essential requirement for urbanisation. However, this trend has led to a series of negative effects [1, 2]. From an environmental point of view, a large number of natural aggregates have been exploited for construction purposes [3,4,5,6]. This has a serious negative impact on the environment due to construction materials mining activities, such as blasting, loading/unloading, transporting, crushing, to name a few [7,8,9,10,11,12]. Moreover, the amount of demolition and construction waste has also increased significantly, due to urban development, especially regarding construction waste from old, demolished buildings [13,14,15]. Aiming towards the achievement of sustainable development, many scientists have researched and applied recycled aggregate concrete (RAC) in construction. This has a double benefit, as construction and demolition waste is managed sustainably, while the exploitation of natural construction materials is considerably lower, thus succeeding in lowering the environmental impact of concrete and preventing the degradation of the ecological environment [16,17,18,19,20]. However, the main disadvantages of RAC are low compression load capacity, and low elastic modulus [21,22,23]. Therefore, an accurate prediction of the compressive strength (CS) of RAC is a necessity, which can allow the safe use of RAC in buildings.

To this aim, artificial intelligence (AI) has been extensively studied and applied as a robust computer engineering tool in the service of construction engineering [24,25,26]. For predicting CS of RAC, Duan et al. [27] successfully used an ANN with 168 sets of data, highlighting the excellent potential of the ANN model for forecasting CS of RAC in their study. MLR and nonlinear regression analysis have also been applied for predicting CS of RAC, in a study of Younis, Pilakoutas [28]. The same research revealed that recycled tires’ steel fibres could improve the CS of RAC. Deshpande et al. [29] also used ANN to predict CS of RAC with promising results. ANN is one of the most popular and widely used algorithms; however, it displays also some drawbacks, mainly poor prediction capacity, especially when the range of the testing dataset coincides with the training data. ANN also has a limited prediction performance when the number of datasets, used to develop and train the model, is limited in number. These drawbacks lead to the junction of ANNs with fuzzy logic (FL) and ANFIS algorithm. In another study, Khademi et al. [30] investigated the predictability of three AI techniques in predicting CS of RAC, namely ANN, ANFIS, and MLR. They concluded that the ANN model was capable of predicting CS of RAC more accurately, achieving an R2 of 0.919 and a MSE of 19.768. The ANFIS algorithm, however, displays a weakness in the determination of the weights in the membership function; thus, some optimisation algorithm (or meta-heuristic algorithms) was applied to solve the disadvantages of ANFIS and enhance the model’s performance [31, 32]. Several AI algorithms such as SVM and ELM have been applied; however, these algorithms have some disadvantages (e.g. large time and memory for computing large datasets; slow learning speed, poor computational scalability) [33,34,35]. In one study, Abdollahzadeh et al. [36] used GEP to predict CS of RAC, showing the high applicability of the GEP model for the CS prediction of RAC. Their GEP model was thus introduced as a tool for predicting CS of RAC containing silica fume. However, several issues, such as optimisation of GEP, comparison with the benchmark algorithms, were not considered in their study. The feasibility of deep learning theory (i.e. convolutional neural networks—CNN) in AI was also considered by Deng et al. [37] for predicting CS of RAC; 74 sets of concrete block were investigated to this purpose. Higher generalisation ability, higher efficiency, and higher precision than the traditional methods were the main findings in their study for the CNN model. Analogous techniques for estimating CS of RAC can be found in the following papers [38,39,40,41]. More details of CS prediction, as well as more extended state-of-the-art on AI techniques developed and used for the prediction of CS, can be found in [42,43,44,45,46,47].

The review of the related literature highlights the importance of the issue under examination in construction engineering. Although a number of research have been conducted for predicting CS of RAC [48,49,50,51,52,53,54,55,56,57,58], further research is still necessary to find a robust algorithm capable of revealing the different aspects of its design and which can allow its generalised use. Furthermore, in the present study, four new hybrid AI techniques, including hybrid algorithms of XGBoost, ANN, ANFIS, and SVR with ICA optimisation algorithm called ICA-XGBoost, ICA-ANN, ICA-SVR, and ICA-ANFIS were developed and applied for predicting CS of RAC. The results of this research add further insight into this very interesting and necessary subject in the field of civil and construction engineering.

2 Experimental data

According to our review of the relevant literature, the CS of RAC is influenced by many factors, such as the fine aggregate density (FA) portion used, the water–cement ratio (WCR), the recycled coarse aggregate density (RCA), the water–total material ratio (WTMR), water absorption (WA), and the natural coarse aggregate density (NCA) [37, 41, 59]. Therefore, 209 RAC experimental results aiming to determine the CS of RAC were conducted, based on the influencing mentioned above factors. The experiments were performed in the laboratory, applying different mixing ratios and parameters. Through the experimental procedure implemented, the WA of RAC samples was determined in the range of 0.338 to 21.604%; WCR lay in the range of 0.307 to 0.736; FA lay in the range of 278.1 to 1211.8 kg/m3; RCA lay in the range of 0 to 1324.2 kg/m3; NCA lay in the range of 0 to 1340.3 kg/m3, and WTMR lay in the range of 0.035 to 0.2. Aiming to indicate the 28-day compressive strength (CS) of RAC samples, the servo press machine 2000 kN was used. Strain gauges were used to measure the deformation of RAC samples through pressure sensors. Experimental results were recorded with the CS of RAC lies in the range of 15.01 to 73.08 MPa, with different mixing ratios. In the present study, WA, WCR, FA, RCA, NCA, and WTMR were considered as the input variables, whereas the CS of RAC at 28 days was selected as the output variable. In Figs. 1 and 2, the properties of the RAC datasets used were depicted.

Fig. 1
figure 1

Histograms of the database parameters

Fig. 2
figure 2

Description of the RAC dataset used

For description of the database used, the mean values, maximum, minimum, and standard deviation (STD) values are presented in Table 1. Basically, some of the RAC variables could be dependent on each other. High positive or negative amounts of the correlation coefficient among the input variables can lead to poor efficiency of the approaches. Besides, it can make difficult explaining the influences expository variables on the respond. Subsequently, the correlation coefficients between all possible variables have been specified and presented in Table 2. It should be noted that the values of the correlation matrix are symmetric to its main diagonal (italicized values in Table 2). There are no significant correlations among the independent input variables (Table 2).

Table 1 Statistics of the experimental database utilised herein
Table 2 Correlation matrix of the input variables

3 Background of intelligent techniques used

3.1 Imperialist competitive algorithm (ICA)

The ICA was proposed by Atashpaz-Gargari, Lucas [60] obtained by simulating human social evolution for solving optimisation issues. It is known as one of the evolution algorithms which may decode continuous function with high performance [61,62,63]. Actually, ICA is a global search algorithm that is developed based on imperialist competition and social policies [64]. Therefore, the most powerful empire can overcome different colonies along with their resource. Other realms can compete together to obtain the territory when an empire collapses. The ICA core may be demonstrated by the eight steps below. Figure 3 shows the pseudo-code of the ICA.

Fig. 3
figure 3

The ICA’s pseudo-code

  1. 1.

    Create initial empires and search spaces by randomly;

  2. 2.

    Colony assimilation: the position of colonies is changed according to the location of the countries;

  3. 3.

    Accidental modifications occur in the features of each country as a revolution;

  4. 4.

    Swapping the territory position for the empire. A colony with a better position can rise and control the empire, and it will replace the previous empire;

  5. 5.

    The empires compete to conquer the other’s colonies;

  6. 6.

    The weaker empires will be defeated and eliminated. The entire colonies of the weaker empires will be lost. In this step, natural selection rules are applied;

  7. 7.

    Check the stop criteria. If the stop criteria are satisfied, then stop the competitive process. Otherwise, return to the step of colony assimilation (step 2).

  8. 8.

    End.

3.2 Extreme gradient boosting (XGBoost)

For the first one, Chen, He [65], the XGBoost is an ensemble tree algorithm developed. After that, it is enhanced according to the gradient boosting (GB) Friedman [66] decision. It may deal with both classification and regression issues efficiently because the boosted trees are generated and worked parallel [67]. In XGBoost, an objective function (OA) is defined based on the conditions of gradient boosting conditions. It is taken into account as the core of the XGBoost algorithm, and similar to many different optimisation methods. Like the GB machine and GB decision tree, XGBoost proposes a reliable and fast model for different engineering simulations based on the parallel boosting trees [10]. Actually, in order to increase the precision of estimations, it can symbolise a soft computing library, which can combine novel algorithms together with approach of GB decision tree.

The XGBoost can be described as below:

Let \(D = \left\{ {(x_{i} ,y_{i} )} \right\}\) is a dataset including of n samples as well as m features (\(\left| D \right| = n,x_{i} \in R^{m} ,y_{i} \in R\)). The suggested tree ensemble model uses z additive functions for approximating the system response as:

$$\hat{y}_{i} = \phi (x_{i} ) = \sum\limits_{z = 1}^{Z} {f_{z} (x_{i} ),\,f_{z} \in } \,F$$
(1)

in which F is the space of regression trees. It is defined as:

$$F = \left\{ {f\left( x \right) = \omega_{q\left( x \right)} } \right\}\,(q:R^{m} \to T,\,\,\omega \in R^{T} )$$
(2)

q stands as the tree structure, T and w are the number of leaf nodes and their weights. In addition, the fk term considered a function, which shows to w and q corresponded to an independent tree.

In order to optimise the ensemble tree along with to decrease errors, the OA of XGBoost can be minimised as follows:

$$L^{(t)} = \sum\limits_{i = 1}^{n} {l(y_{i,} \hat{y}_{i}^{(t - 1)} + f_{t} (x_{i} )) + \varOmega (f_{t} )}$$
(3)

l stands as a convex function (i.e. loss function) which is applied to determine the difference between exact and calculated values, \(y_{i}\) is considered as a measured value,\(\hat{y}_{i}\) stands as a predicted value. For minimising the errors, the number of iteration (t) is used, whereas \(\varOmega\) is the penalty factor for the complication of the regression tree approach:

$$\varOmega (f_{k} ) = \gamma T + \frac{1}{2}\lambda \left\| w \right\|^{2}$$
(4)

3.3 Support vector regression (SVR)

For estimating problems, Cortes and Vapnik [68] introduced SVM with the capability of wide application as a benchmark machine learning approach. It has two essential branches (i.e. support vector classification (SVC) and support vector regression (SVR) that SVR is utilised as the most usual figure of SVM [69, 70]. The nature of SVR coming from target values, which detect a \(\varphi (x)\) function for mapping data to flat space aiming to achieve a space as flat as feasible. By considering two forms of nonlinear and linear regression, solving complex problems is l [71].

As can be seen in Fig. 4, optimised and linear regression problems may be performed by a convex optimisation of calculation with constraints and solutions by SVR for the linear regression problems.

Fig. 4
figure 4

Linear SVR

In terms of nonlinear regression problems, optimisation and nonlinear regression problems by SVR can be used with a convex optimisation of calculation along with kernel functions for transforming the dataset in a higher dimensional of the dataset in the feature space. The kernel functions with two different forms that are the most commonly used, including radial and also polynomial basis function are additionally shown in Fig. 5.

Fig. 5
figure 5

Nonlinear SVR

3.4 Artificial neural network (ANN)

ANN has wide applications in different areas. This approach has been introduced since the 1970s [6, 72,73,74,75,76]. ANN is a family of approaches in AI that is constructed based on the learning capability of the human brain. Basically, the ANN model structure has a layer in input and hidden layer(s) as well as a layer in output [77]. The essential parameters of ANN were the neurons or nodes known as their connections and processing elements [78, 79]. In the input layer, the input neurons provide the input signals (i.e. properties of the dataset). Accurately, in the present work, the nodes in input achieve the messages of WA, WCR, FA, RCA, NCA, and WTMR. Then, the hidden nodes get the signal from the input neurons and implementing a computational with weights. They are then delivered to the subsequent nodes for the next calculations [80]. For simple regression problems, an ANN system with only one hidden layer can provide the outcome predictions with the acceptable [81]. The ANN algorithm that has two hidden layers is commonly utilised for more complex problems [82, 83]. Also, the output nodes in the output layer achieve signals from nodes in hidden layers and computational of output amounts. In the present paper, for the ANN model, the CS of RAC is utilised and used as an output variable. Figure 6 shows the architecture of the ANN algorithm that is useful to estimate the CS of RAC.

Fig. 6
figure 6

The ANN structure for estimating the CS of RCA

3.5 Adaptive neuro-fuzzy inference system (ANFIS)

ANFIS is one of the ANN branches, which is developed based on a combination of ANN and a fuzzy system [84]. It was introduced by [85] first in 1993 and was widely applied in many fields [32, 86,87,88,89,90]. In ANFIS, the membership functions are assigned and adjusted by the training capability of ANN. The BP algorithm is used to adjust the parameter of the ANN model until the error is satisfied [91].

In ANFIS, IF–THEN rules are used to predict any problems. Assume x and y are the inputs, and z is the output in a fuzzy inference system. The IF–THEN rules are then applied, as illustrated in Fig. 7.

Fig. 7
figure 7

IF-THEN rules of ANFIS model

In theoretical, ANFIS includes five layers (except input and output layers), as shown in Fig. 8:

Fig. 8
figure 8

Architecture of ANFIS

  • Layer 1: Generating the membership grades (\(\mu A_{1} ,\,\mu A_{2} ,\,\mu B_{1} ,\,\mu B_{2}\)) from the inputs (i.e. a and y) by adaptively act.

    $$O_{1,i} = \mu_{{A_{i} }} (x) \;{\text{for}}\;i = 1,2\;{\text{or}}\;O_{1,i} = \mu_{{B_{i} - 2}} (y)\;{\text{for}}\;i = 3,4$$
    (5)

    where i is the number of input variables.

  • Layer 2: Getting an output using AND/OR rule node, called firing strengths.

    $$O_{2,i} = \mu_{{A_{i} }} (x)\mu_{{B_{i} }} (y),\,i = 1,2, \ldots$$
    (6)
  • Layer 3: Computing the normalised firing strength by an average node.

    $$O_{3,i} = \bar{\varpi }_{i} = \frac{{\varpi_{i} }}{{\varpi_{1} + \varpi_{2} }},i = 1,2, \ldots$$
    (7)
  • Layer 4: Turning the parameters of p, q, and r and showing them as consequent nodes.

    $$O_{4,i} = \bar{\varpi }_{i} f_{i} = \bar{\varpi }_{i} (p_{i} x + q_{i} y + r_{i} )$$
    (8)
  • Layer 5: Computing the total average of output. This layer uses a sum of input signals to calculate output nodes.

    $$O_{5,i} = f = \sum\limits_{i} {\bar{\varpi }_{i} f_{i} }$$
    (9)

4 Framework of the proposed ica-xgboost model

As regarded above, the primary purpose of this work is to present a novel technique of AI for predicting the CS of RAC based on the XGBoost model and the ICA optimisation, called ICA-XGBoost technique. Accordingly, the RAC dataset with different mixing ratios and different CS has been divided into two sections for testing and training purposes, as the first stage. Of the total dataset, 80% of the data (approximate 169 experimental datasets) was used for the development of the ICA-XGBoost model, whereas the remaining 20% (approximate 40 experimental datasets) was used for testing the precision of the developed ICA-XGBoost model. In the second stage, an initial XGBoost model is developed based on the training dataset. Subsequently, the hyper-parameters of the developed XGBoost model are chosen as the main parameters to optimise by the ICA. Before optimising the hyper-parameters of the initial XGBoost model, the settings of ICA need to be established in the third stage. Then, the ICA-XGBoost model was generated in the fourth stage by adjusting the factors of the initial XGBoost model (by the established ICA optimisation). In this way, the accuracy of the XGBoost model can be enhanced. To check the improvement of the ICA-XGBoost model, the fitness of the generated ICA-XGBoost model was evaluated via error values, i.e. RMSE, as the fifth stage. Stopping criteria is checked through the RMSE for the sixth stage. If the model errors are satisfied with the stop condition (i.e. lowest RMSE), the optimisation process by ICA is horned there. Subsequently, the testing dataset with 40 experimental datasets was used to re-check the accuracy/error level of the developed ICA-XGBoost model, as the final stage. The optimisation procedure of the XGBoost model by the ICA for predicting the CS of RAC is shown in Fig. 9.

Fig. 9
figure 9

Optimisation procedure of the XGBoost model by the ICA for predicting the CS of RAC

5 Development of the models

As described in Fig. 8, the original dataset includes 209 experimental results and was divided into two parts (80/20) before proceeding to develop predictive models. Note that the dataset was divided randomly. Also, all the stated models are generated based on the same training dataset.

5.1 ICA-XGBoost model

For the development of the ICA-XGBoost model, an initial XGBoost model was developed with the following hyper-parameters: subsample percentage (\(\varsigma\)), boosting iterations (k), minimum loss reduction (\(\gamma\)), max tree depth (d), shrinkage (\(\eta\)), minimum sum of instance weight (\(\mu\)), and subsample ratio of columns (\(\delta\)). Subsequently, the ICA’s parameters were established for optimising the hyper-parameters of the XGBoost model. In the present work, the ICA’s parameters were set up as follow:

  • The number of initial countries (Ncountry): a trial and error procedure with various Ncountry was conducted with Ncountry of 50, 100, 150, 200, 250, 300, 350, 400, 450, 500, respectively.

  • The initial imperialists (Nimper): 30

  • The lower–upper limit of the optimisation region (L): [− 10, 10]

  • The assimilation coefficient (As): 2.8

  • The revolution of each country (r): 0.6

  • The maximum number of iterations (Ni): 1000 times

After establishing the parameters of the ICA, imperial competition is implemented to find the most substantial empire, corresponding to the most optimal values of the XGBoost model with the lowest RMSE. The competition process was repeated 1000 times to ensure the obtained values are the most optimal. Eventually, an optimal ICA-XGBoost model was found with the following parameters: \(\varsigma\) = 0.427, k = 856, \(\gamma\) = 4.288, d = 3, \(\mu\) = 2, \(\eta\) = 0.022, and \(\delta\) = 0.692. The optimisation process of the XGBoost model by the ICA is shown in Fig. 10.

Fig. 10
figure 10

The optimisation process of the XGBoost model by the ICA for predicting the CS of RAC

5.2 ICA-ANN model

For the development of the ICA-ANN model, the similar techniques and the same training dataset were used. Note that all the parameters of the ICA are the same as those used for the development of the ICA-XGBoost model. However, unlike the XGBoost model, an ANN model with the backpropagation algorithm was selected as the background of the optimisation. Indeed, an ANN model includes two hidden layers (i.e. ANN 6-16-21-1), which was developed for the prediction of the CS of RAC in the Python platform. The min–max scale method was used to transfer the input data in the range of -1 to 1 to avoid overfitting. Herein, the weights and biases of the ANN 6-16-21-1 model were optimised by the ICA. RMSE also utilised to analyse the performance of the optimisation process for the ICA-ANN model, as shown in Fig. 11. Finally, the optimal ICA-ANN model was determined with the optimal weights and biases, as shown in Fig. 12.

Fig. 11
figure 11

The optimisation process of the ANN model by the ICA for predicting the CS of RAC

Fig. 12
figure 12

The ICA-ANN structure for estimating the CS of RAC

5.3 ICA-SVR model

Similar to the ICA-XGBoost and ICA-ANN models, the ICA-SVR model was developed based on the optimisation of ICA for the SVR model. The same training dataset and parameters of the ICA were applied as those used for the previous models (i.e. ICA-XGBoost and ICA-ANN models). In this regard, the radial basis function of the kernel is utilised for the development of the SVR model. Hence, the sigma (\(\sigma\)) and cost (C) are optimised by the ICA to enhance the accuracy of the SVR model. Lowest RMSE was used as the final goal for the development of the ICA-SVR model. Finally, an optimal ICA-SVR model was developed with C of 27.509 and \(\sigma\) 0.015. The optimisation process of the ICA-SVR model for predicting the CS of RAC is shown in Fig. 13.

Fig. 13
figure 13

The optimisation process of the SVR model by the ICA for predicting the CS of RAC

5.4 ICA-ANFIS model

For the ICA-ANFIS modelling, an initial ANFIS model is developed as the first step; after that, the ICA was applied to optimise the developed ANFIS model. The parameters of the membership functions of the generated ANFIS model were optimised/trained by the ICA in this task. The same training dataset and ICA’s parameters were applied as those developed the previous models (i.e. ICA-XGBoost, ICA-ANN, and ICA-SVR models). A fitness function, i.e. RMSE, also utilised to assess the precision of the optimisation process for the ICA-ANFIS model. Ultimately, the optimal ICA-ANFIS model was found with the number of fuzzy terms of 15 and maximum iterations of 24. The training process of the ICA-ANFIS algorithm on the training dataset is shown in Fig. 14.

Fig. 14
figure 14

The optimisation process of the ANFIS model by the ICA for predicting the CS of RAC

6 Statistical criteria for model assessment

Once the models were developed, the testing dataset includes 40 experimental datasets, which were used to evaluate the accuracy in practical engineering of the models. Five statistical criteria, such as RMSE, R2, MAE, MAPE, and VAF, were suggested to evaluate the models’ performances, as follow:

$${\text{RMSE}} = \sqrt {\frac{1}{m}\sum\limits_{i = 1}^{k} {(y_{{i{\text{RAC}}}} - \hat{y}_{{i{\text{RAC}}}} )^{2} } }$$
(10)
$${\text{R}}^{ 2} = 1 - \frac{{\sum\nolimits_{i = 1} {(y_{{i{\text{RAC}}}} - \hat{y}_{{i{\text{RAC}}}} } )^{2} }}{{\sum\nolimits_{i} {(y_{{i{\text{RAC}}}} - \hat{y}_{{i{\text{RAC}}}} )^{2} } }}$$
(11)
$${\text{MAE}} = \frac{1}{n}\sum\limits_{i = 1}^{k} {\left| {y_{{i{\text{RAC}}}} - \hat{y}_{{i{\text{RAC}}}} } \right|}$$
(12)
$$MAPE = \frac{100\% }{k}\sum\limits_{i = 1}^{k} {\left| {\frac{{y_{{i{\text{RAC}}}} - \hat{y}_{{i{\text{RAC}}}} }}{{y_{{i{\text{RAC}}}} }}} \right|}$$
(13)
$${\text{VAF}} = \left( {1 - \frac{{\text{var} \,(y_{{i{\text{RAC}}}} - \hat{y}_{{i{\text{RAC}}}} )}}{{\text{var} \,(y_{{i{\text{RAC}}}} )}}} \right) \times 100$$
(14)

k stands as the number of instances; \(\bar{y}\), \(y_{{i{\text{RAC}}}}\), and \(\hat{y}_{{i{\text{RAC}}}}\) show the average, measured, and modelled amounts of the response variable, respectively.

Furthermore, the following new recently proposed [92,93,94,95], the a10-index, engineering index to the reliability evaluations of the expanded AI models have been used:

$$a10 - {\text{index}} = \frac{m10}{M}$$
(15)

M stands as the number of dataset sample, and also m10 is the number of samples along with a value of experimental rate value/estimated value in the range of 0.90 and 1.10. It is important to note that for a complete predictive approach, the values of a10-index were considered to be unity. The suggested a10-index possesses the advantage. Note that their value showed a physical engineering meaning. It is noted that the amount of the samples which satisfies calculated amounts with a deviation of ± 10% has compared to experimental data.

Also, a ranking method and colour intensity were utilised to classify the developed models.

7 Results and discussion

The performance of the developed soft computing models (ICA-XGBoost, ICA-ANN, ICA-SVR, and ICA-ANFIS) regarding the prediction of compressive strength of recycled aggregate concrete is evaluated quantitatively both in training and testing phase (Table 3) through the six performance indexes (a10-index, RMSE, MAE, MAPE, VAF, and R2) that have been previously presented.

Table 3 Performance of the developed hybrid models both for training and testing datasets

Before the actual assessment of the models, one must first investigate and determine whether the well-known and frequent issue of “overfitting” has occurred (overfitting problem in machine learning). A reliable manner in which this issue can be assessed is related to the comparison of the difference performance indices, among training data and testing data. When overfitting occurs, the performance indices for the training phase are quite satisfactory; however, the performance indices for the validation phase are quite lower. The smaller the difference between the performance indices of these phases, the lower the possibility that overfitting has occurred. Specifically, when the difference in the indices R2 and VAF is less than 5%, the probability of overfitting occurrence is extremely low. Thus, based on the values displayed in Table 3, overfitting of the models developed within this research has been avoided.

Based on the results presented in Table 3 and having ensured that the overfitting problem has not occured, the optimum AI model is the developed ICA-XGBoost model, which ensures, for the case of testing datasets, the optimum values for all the performance indices. In contrast, the ICA-ANFIS model indices indicate that its performance is the lowest among all models.

Considering the parameters (operators) of the proposed ICA-XGBoost, ICA-ANN, ICA-SVR, and ICA-ANFIS models, it can be concluded that although the settings of the ICA are the same, however, the accuracy of the models are different. This finding shows different prediction power of the different algorithm, as well as the ICA, seems more suitable when combined with the XGBoost model for estimating the CS of RAC. Figure 15 illustrates the accuracy of the predicted values by the different hybrid models in determining the CS of RAC in terms of scatter plot. Also, a comparison of measured and predicted values by different models is shown in Fig. 16 in terms of histogram. According to Fig. 16, all applied models have efficiency for predicting CS of RAC.

Fig. 15
figure 15

Illustrating the accuracy of the estimated values by the individual models

Fig. 16
figure 16

A comparison of measured and estimated values of the models

Furthermore, in Fig. 17 the ratio of the experimental values concerning the predicted values is depicted, for the datasets which were used for testing the reliability of the proposed ICA-XGBoost optimum neural network in terms of compressive strength prediction. Noted that all samples utilised for the testing process possess a deviation lower than ± 10% (points are among the two dotted lines in Fig. 15).

Fig. 17
figure 17

Experimental to the predicted values of compressive strength based on the ICA-XGBoost

According to Box-plot finding (Fig. 18), ICA-XGBoost and ICA-ANN algorithms could predict the minimum value of CS of RAC properly, but neither ICA-XGBoost and ICA-ANN nor other applied models could not predict maximum value accurately. The ICA-XGBoost algorithm has a higher performance in terms of median values, followed by ICA-ANN, ICA-SVR and ICA-ANFIS models. ICA-ANFIS and ICA-SVR algorithms have a higher prediction power in predicting third quartile (Q3) and first quartile (Q1), respectively.

Fig. 18
figure 18

Box-plot for the models’ performance in the testing phase

Based on the result of the Taylor diagram (Fig. 19), the proposed ICA-XGBoost model outperforms other models (correlation coefficient higher than 0.99) followed by ICA-SVR, ICA-ANN, and ICA-ANFIS, respectively. This can be related to computing capability of different algorithms, and as each model have advantages and disadvantages, thus different models should be applied and the best one selected for future studies. Recently applied hybrid algorithms have been rapidly increased, and most of the literature review shows that hybrid algorithm can enhance the prediction power of the standalone algorithms. Khosravi et al. [96] applied standalone ANFIS as well as ANFIS hybrid with genetic, imperialist competitive, and differential evolution algorithms for reference evaporation estimation. Finally, they stated that all hybrid algorithms have a higher prediction power than standalone algorithms. Khozani et al. [97] applied four standalone algorithms of RF, RT, REPT, and M5P as well as a hybrid algorithm of bagging-M5P model for apparent shear stress prediction. They finally stated that hybrid algorithm outperforms others. XGBoost as one of the flexible models has some advantages, such as it can work on both regression and classification problems, parallel processing, and it can work effectively with large and multidimensional datasets [8, 98, 99].

Fig. 19
figure 19

Taylor diagram for models’ evaluation and comparison

Generally, as this kind of research such as compressive strength of recycled aggregate concrete prediction including the relationship between input variables and between inputs and output are not simple and have a nonlinear relationship and simple and empirical models do not have sufficient accuracy. Thus, the more nonlinear and flexible models, the higher prediction power. AI algorithms with nonlinear structure especially hybrid models are more flexible and robust than standalone models [31]; therefore, hybrid algorithm can enhance the prediction power of standalone algorithms and completely proper to prediction of phenomena with complex process.

8 Conclusion

Recycled aggregate concrete is a promising material which could replace typical concrete. Its extensive use can contribute, not only towards the improvement of economic efficiency, but also towards sustainable development through the reduction in concrete’s environmental impact. However, due to the influence of mortar and cement remnants from the original concrete on the surface of the recycled aggregates, its 28-days compressive strength is often inferior to that of typical concrete. Therefore, an accurate prediction of the 28-days compressive strength is necessary to optimise this concrete material and to ensure its safe application for building purposes.

In the present work, four different AI models have been trained and developed for the prediction of compressive strength of recycled aggregate concrete. Among these models, the ICA-XGBoost model is proposed as the optimum. Namely, based on the newly proposed performance a10-index, all samples utilised for the testing process possess a deviation lower than ± 10% in relation to the actual experimental values, proving the developed model as a useful tool for researchers, engineers, as well as for supporting not only teaching, but also interpretation of the mechanical behaviour of recycled aggregate concrete. Furthermore, based on the proposed ICA-XGBoost technique, recycled aggregate concrete can be used safely for construction purposes, when specific parameters are fulfilled and may thus, in the future, serve as an essential, environmentally friendly building material.