1 Introduction

Nowadays, concrete has an undeniably special position in the building industry and is one of the most widely used building materials in the world with its ever-increasing widespread application. The reason for such a special position is the fulfilment of the technical, economic and environmental requirements of human societies [1]. The implementation of concrete structures as the cheapest and most durable structures has driven the researchers to use concrete with higher strength, which has brought great successes in this area [2]. The most important factor that determines the strength of High-Strength (HS) concrete is the porosity in three concrete phases (aggregates, cement paste and interfacial transition zone (ITZ)) [3, 4]. To produce HS concrete, the contrasting effect of water and cement content on the consistency and strength of concrete cannot be counteracted without the use of water-reducing additives. For this reason, in the past decade, the use of super-plasticizers has played an important role in the production of HS concrete. Basically, the ITZ is the weakest part of the concrete with an ordinary weight made of resistant aggregates with the maximum size of 12–20 mm and the water to cement ratio of 0.4–0.7. The concrete strength can be greatly increased by reducing the maximum size of coarse aggregates for a given water-to-cement ratio, as this will improve the ITZ strength [5,6,7,8]. Therefore, in the mix design of HS concrete, the maximum aggregate size is usually considered 19 mm or less. The advantage of using HS concrete is the relatively low hydration heat per unit strength, which reduces the possibility of thermal cracking and minimizes the column dimensions and thus, ensures the beauty of the structure [9,10,11]. HS concrete columns can hold more weight and therefore, be made slimmer than OPC columns, which allows for more usable space, especially in the lower floors of buildings [10]. The use of pozzolans in the production of composite cement has reduced the total cost of concrete, but the unavailability of various types of pozzolans for the construction of structures and their contribution to the mix design of HS concrete has made it challenging to use the pozzolans [11, 12]. Therefore, in this study, a mix design containing four main components of concrete (water, cement, coarse and fine aggregates) and super-plasticizer was used to present computational models and estimate the compressive strength of concrete.

Up to date, various applied machine learning models have been recruited to simulate the mechanical properties of concretes including Gene Expression Programming (GEP) [13], Random Forest (RF) [14], Model Tree (MT) [15], Artificial Neural Network (ANN) [16], Adaptive Neuro-Fuzzy Inference Systems (ANFIS) [17], Support Vector Machines (SVM) [13, 18] and Multivariate Adaptive Regression Splines (MARS) [19, 20]. It is worth noting that ML techniques have also been used to model several other properties of concrete [21,22,23,24,25]. For example, Ashrafian et al. presented an integrative predictive formula-based model to indicate relationship of mixture and hardened strength of foamed concrete [26]. Compressive strength prediction of plastic concrete using four computational intelligence and statistical approaches was studied by Amlashi et al. [27]. Sun et al. developed an enhanced SVM method to evaluate CS and permeability coefficient [28]. Shahmansouri et al. performed GEP technique to estimate properties of eco-environmentally concrete containing zeolite [29]. Feng et al. applied machine learning method named an adaptive boosting to predict the hardened strength [30]. Iqbal et al. implemented collected data records to propose goodness of fit model for formulation of the mechanical properties of eco-friendly concrete inspired by GEP method [31]. Asteris et al. utilized artificial computing systems to predict CS of self-compacting concrete as an alternative predictive models in terms of MARS predictive strategy [32]. Behnood et al. used decision tree approach for mechanical properties of ordinary concretes [33]. In case of hybrid evolutionary MARS application in civil engineering, Al Sudani et al. [33] developed MARS Integrated with differential evolution algorithm to forecast streamflow pattern in semi-arid region. Rezaie-Balf et al. [34] predicting daily solar radiation using MARS model trained by crow search algorithm. Tien Bui et al. [35] presented new intelligent model based on GIS-based MARS and particle swarm optimization to estimate flash flood susceptible areas.

Applied regression-based soft computing methods do not consider any clear knowledge about the relationship between the inputs and corresponding target variables. It is a weakness of all machine learning models that the independent instances of the estimative database should fall within the range of predictors of calibration dataset. It seems that machine learning techniques could be optimistically applied but due to their dependence on the initial control parameters their use is limited [35]. Aiming at improving the performance capacity of soft computing models in terms of accuracy and error, the optimization algorithms can be employed. To this end, soft computing models should be hybridized with evolutionary algorithms to create intelligent solutions called integrated evolutionary models. Based on this concept, the MARS user-defined parameters are optimized using Water Cycle Algorithm (WCA), Crow Search Algorithm (CSA) and Cat Swarm Optimization (CSO). The main parameters for tuning of MARS are maximum interaction between variables (Imax), the maximum number of basis functions (Mmax) and smooth parameters (c). The main characteristics of MARS model as they are the controlling elements of the complexity and generalization in the modeling process as these parameters reported by Andalib and Atry [36]. Thus, predictability of MARS model highly depends on the selected parameters quality. Friedman has selected these parameters among widespread ranges which their actual values are dependent upon the database [37]. Selection of optimum values of the parameters is a challenge in SC methods, and it is similar to an optimization effort.

In this paper, the application of WCA, CSA and CSO search engines for optimization of MARS parameters was investigated. In terms of WCA, to prove the concept of evolutionary optimization, the proposed algorithm selected to integration process based on high convergency and stability of the final solutions. The performance capability of WCA to fully explore the search space more accurately was implemented using the process of water cycle in nature. This may be utilized to search the local solution space, while the computational procedure of flowing can be utilized for the global search for enhancement of the convergence speed. Besides, population-based optimization algorithm including CSA and CSO which was applied to control the diversity due to application of limited and simple adjustable parameters. Also, aforementioned evolutionary algorithms have developed optimal performances for designing reliable solution for the mathematical test functions and operations of the systems [38,39,40]. Many researchers have proven superiority of WCA, CSA and CSO algorithms over other available algorithms [36,37,38,39]. It is an advantage for proposed meta-heuristic algorithms that they can be combined with MARS to detect hyper-parameters such as Mmax, Imax and c.

As mentioned previously, classical machine learning models are trustworthy means for prediction the concrete properties but they could not generate a relationship between the main concrete properties and the mixture design codes. The main goal of this research is to develop a new evolutionary formula-based models for predicting the Compressive Strength (CS) of HS concrete through adoption of a self-adaptive approach which is based on MARS integrated WCA (MARS-WCA), MARS integrated CSA (MARS-CSA) and MARS integrated CSO (MARS-CSO). Besides, another motivation of this research is to investigate the effect of uncertainties of proposed SC models for HS concrete mixtures. For this aim, by considering the influential input parameters as random variables, the effects of uncertain parameters on the reliability analysis results are investigated for different values of target CS using the Monte Carlo Simulation (MCS) method and external validations to satisfy model’s verification.

This research is presented in four sections. Section one introduces the topic of this research. Section two deals with the applied MARS and evolutionary algorithms, theoretical framework and HS concrete database details. Section 3 describes the standalone MARS and the proposed evolutionary models together with the evaluation process. Section 4 summarizes this the key-findings of this study.

2 Materials and Methods

2.1 Theoretical Framework and Data Description

In Construction of buildings sector, to optimize the size of columns in the lower floors HS concrete can be utilized specially in in high-rise buildings. One of the most advantages of HS concrete is for deflection reduction due to a higher elastic modulus [41]. In addition, it provides a higher tensile strength, greater durability, reduced creep than traditional concrete. By using HS concrete, it can be concluded that the cost of construction of buildings can be decreased with the currently prevailing prices of concrete and steel and make the members slender for more flexible structure [4, 11, 41]. This type of concrete can also be used to reduce slab depths and, therefore, a building’s overall height. This can result in significant cost savings while slimmer high-strength concrete columns can increase the overall net area.

In this study, 324 experimental mix design aiming to determine the compressive strength of HS concrete based on the influencing factors was gathered from the Al-Shamiri et al. [42] for constructing accurate models. In the mentioned research, the extreme learning machine (ELM) and ANN approaches was developed to estimate the CS of HS concrete. The developed ELM model requires more hidden neurons than ANN model, but its training speed is extremely fast as no iteration is required during the learning phase. The results shown that the developed model was reliable for predicting the CS of HS concrete, but they do not have capabilities to extract equation from their predicted values for future studies. Recently, Hameed et al. [43] used high-order response surface model as the prediction model to accurately predict the compressive strength of HS concrete based on dataset presented by Al-Shamiri et al. [44]. They also used dimensional variables to develop proposed model for predicting of compressive strength of HS concrete.

The descriptive measures including mean, minimum, maximum, standard deviation, Skewness and Kurtosis are summarize in Table 1 to introduce the characteristics of explained dataset. This data set can represent the whole population or a sample of it. Statistical measures can be classified as measures of central tendency and measures of spread. Skewness assesses the extent to which a variable’s distribution is symmetrical. If the distribution of responses for a variable stretches toward the right or left tail of the distribution, then the distribution is referred to as skewed. Kurtosis is a measure of whether the distribution is too peaked. Based on the statistical analysis presented in Table 1, it can be deduced that CS and cement content ranges \([37.5-73.6 \mathrm{MPa}\)] and [\(284-600 \mathrm{kg}/{\mathrm{m}}^{3}\)], respectively.

Table 1 Statistical analysis of input and output variables

Principally, some of admixture materials which is considered as the input parameters for the development of AI models could have a high positive and negative dependency to each other, which can result in poor efficiency and accuracy of the modeling. In addition, extracting the most influential independent variables on the compressive strength could be difficult to explain for modeling. In this regard, the correlation matrix which depicts the correlation between all the possible independent variables is proposed in this study (Table 2). According to Table 2, no significant correlations are observed between the independent input variables for the prediction of CS of HS concrete.

Table 2 Correlation matrix of the input variables against target variable

For developing of experimental records, three scenarios including 70–30, 75–25, and 80–20 percentages of the training and testing groups, respectively, were considered. To do so, linear regression method was performed to evaluate proposed scenarios. Table 3 presented results of proposed scenarios that investigated in this study. According to Table 3, optimal scenario for development of AI methods was 75–25 splitting percentage in terms of statistical metrics.

Table 3 Data splitting evaluation

One of the important steps in machine learning models is randomization in which dataset should randomly mixed to prevent overfitting, and then, processed data are categorized into two groups, training (to build up the predictive model) and testing (to evaluate the constructed model). In this study, about 75% of the experimental data (243 data samples) is for development (training phase) and the remainder (81 data samples) should be kept unseen for evolutionary MARS models. For experimental details on material specimens, mixture proportion and curing conditions of these specimens, the readers are referred to [44]. The descriptive measures and marginal plot of the input parameters for modeling of the CS of HS concrete are presented in Table 2 and Fig. 1. The marginal plot allows studying the relationship between 2 numeric variables and the distribution of x and y axes variables using a histogram. This helps to visualize the distribution intensity at different values of variables along both axes.

Fig. 1
figure 1

Marginal plot of each input variable vs. compressive strength along with distribution of each variables (unite of compressive strength, super-plasticizer, and other variables are MPa, percentage of cement, and Kg/m3, respectively)

2.2 Multivariate Adaptive Regression Splines (MARS)

MARS is a nonparametric paradigm consisting splines for nonlinear modeling between the independent and target variables of a knowledge system. For this purpose, it utilizes a regression-based intelligence algorithm [37]. The Basis Function (BF) based on slope of regression line which extracted in final model (one spline to more) has changed. MARS being a flexible method in assessing nonlinear procedure in predictor and target variables, in comparison with other common method applied in regression-based data intelligent approaches [43, 45]. MARS could search probable situation between in all degrees and finding solution for variable's interactions due to extract complicated data structures from multi-dimensional datasets [37]. The general equation for MARS function is as follows:

$$f\left( x \right) = \beta _{0} + \sum\limits_{{_{{m = 1}}^{{}} }}^{M} {\beta _{m} \lambda _{m} \left( x \right)}$$
(1)

Here, \(f\left(x\right)\) denotes the predicted response corresponding to the predictor variable x. \({\beta }_{0}\) and \({\beta }_{m}\) are the predicted constant coefficients in order to attain the best data fit and \({\uplambda }_{m}\) is a spline function. M is the number of basis functions. In MARS, the knots locations are selected through an adaptive regression algorithm and BFs are created using a stepwise search [46]. Two steps are adopted in MARS for optimization namely the forward stage and backward stage. To reduced data over-fitting, a large number of BFs are utilized in the forward stage. MARS adopted the generalized cross-validation (GCV) algorithm to remove pruned BFs, which as presented in Eq. 4, helps with preventing over-fitting in the backward stage [47]. The generalized cross-validation formula is as follows:

$$\mathrm{GCV}=\frac{\frac{1}{\mathrm{N}}{{\sum }_{\mathrm{i}=1}^{\mathrm{N}}\left[{\mathrm{y}}_{\mathrm{i}}-\mathrm{f}({\mathrm{x}}_{\mathrm{i}})\right]}^{2}}{{\left[1- \frac{\mathrm{C}(\mathrm{B})}{\mathrm{N}}\right]}^{2}}$$
(2)

In the above equation, N is the number of training data records, C (B) indicated the effective number of parameters in BFs, C denotes the penalty factor corresponding to each basis function and B is the number of BF. C (B) could be presented as:

$$C\left(B\right)=\left(B+1\right)+\mathrm{d}B$$
(3)

where B is the number of spline function and \(d\) is the penalizing parameter. After obtaining the MARS model, overall relationship is acquired using combination of all BFs including intercept BF and pairwise BF with interaction.

2.3 Water Cycle Algorithm (WCA)

WCA has been a nature-inspired optimizing solution proposed based on the observations from nature phenomena such as river, sea, mountain cloud and rain. Within the hydrologic cycle or water cycle, the clouds are formed by evaporation, also they are generated through photosynthesis by plants and trees and transfer to atmosphere. A precipitation when happen that the atmosphere gets colder and clouds get condense [38]. Thus, underground water is filled with melted snow or rainfalls. Then, the water moves under the ground as the surface water which flows over the ground. Therefore, the cycle continues by evaporation from the rivers and streams [39].

Similar to other meta-heuristic algorithms, each raindrop is taken as an initial population in this method. Initially, we assume we have rain or precipitation. So, the best individual or rainfalls are chosen as a sea and some other good rainfalls are assumed as a river. Thus, in a Nvar dimensional optimization procedure a raindrop would be an array of 1 × Nvar which is calculated by the follow [48]:

$$\mathrm{Raindrop}= \left[{x}_{1},{x}_{2},{x}_{3},\dots. {x}_{N}\right]$$
(4)

To compute the evolutionary process, alternative procedure which is raindrop’s matrix for the size of Npop \(\times\) Nvar is created which is the raindrop’s population:

$$\mathrm{Population\,\,of\,\,raindrops}= \left[\begin{array}{c}{\mathrm{Raindrops}}_{1}\\ \begin{array}{c}{\mathrm{Raindrops}}_{2}\\ \begin{array}{c}{\mathrm{Raindrops}}_{3}\\ \vdots \end{array}\end{array}\\ {\mathrm{Raindrops}}_{{N}_{\mathrm{pop}}}\end{array}\right]$$
(5)

The remained raindrops are considered as streams which flow into the rivers and sea. Here, we assume a parameter named Nsr, which is the sum of rivers plus a single sea (Eq. 6). The remained population is computed using Eq. 7:

$${N}_{\mathrm{sr}}=\text{Number of Rivers}+1 (\mathrm{sea})$$
(6)
$${N}_{\mathrm{Raindrops}}= {N}_{\mathrm{pop}}- {N}_{\mathrm{sr}}$$
(7)

Now, each river adsorbs the stream water in proportion with the flow intensity. Therefore, the amount of water entering the river from each stream is different from the other streams. Eq. 8 is used to designate/assign raindrops to rivers and sea, on the basis of flow intensity:

$$\begin{aligned}{\mathrm{NS}}_{n}&=\mathrm{round} \left\{\left|\frac{{\mathrm{cost}}_{n}}{\sum_{i=1}^{{N}_{\mathrm{sr}}}{\mathrm{cost}}_{i}}\right|\times {N}_{\mathrm{Raindrops}}\right\},\\&\quad n=\mathrm{1,2}, \dots, {N}_{\mathrm{sr}}\end{aligned}$$
(8)

In the above equation, NSndenotes the number of streams flowing to a certain sea or rivers. The flowchart of the water cycle optimization algorithm is shown in Fig. 2.

Fig. 2
figure 2

Flowchart of water cycle optimization algorithm

2.4 Crow Search Algorithm (CSA)

Crows have several characteristics: one of them is the memorization of faces, the second is communication via advanced ways and third is their capability in hiding and recovering the food when needed. These features of crows help them in discovering what others have hidden and steal them after leaving of the owner. Askar Zadeh [39], considering the abovementioned characteristics of crows, proposed an evolutionary algorithm which himself called it CSA for solving of the complicated engineering challenge.

As other meta-heuristic algorithms, in CSA we assume a dimensional environment with a number of crows. The population size which is the crow number here is denoted by N. The crow position i at time (iteration) within the search environment is defined by a vector \(x^{{i,{\text{iter}}}} = [x_{1}^{{i,{\text{iter}}}},x_{2}^{{i,{\text{iter}}}}, \ldots,x_{d}^{{i,{\text{iter}}}} ]\) where \(i=\left(1, 2, \dots, N\right)\) and \(iter=(1, 2, \dots, {\mathrm{iter}}_{\mathrm{max}})\). Here, itermax represents the maximum number of iterations.

The parameter of awareness probability (AP) in CSA helps with balancing between the intensification and diversification [33]. By decrease in the AP value, CSA tends to apply the local region searching, so a small AP value increases intensification. Also, by increase in the AP value, the searching probability of the neighborhood of current ideal solutions reduces, leading to increase in diversification. The steps for implementing CSA are as follows [49]:

  • Step 1: Definition of the optimization problem, decision variables and constrains and valuation of CSA user-defined parameters including the flock size, maximum number of iterations (itermax), Flight Length (FL) and AP.

  • Step 2: Initializing the memory of crows in a d-dimensional search space using Eqs. 9 and 10.

    $$\mathrm{Position}=\left[\begin{array}{c}\begin{array}{ccc}{x}_{1}^{1}& {x}_{2}^{1}& \dots \\ {x}_{1}^{2}& {x}_{2}^{2}& \dots \\ \vdots & \vdots & \vdots \end{array} \begin{array}{c}{x}_{d}^{1}\\ {x}_{d}^{2}\\ \vdots \end{array}\\ \begin{array}{cc}{x}_{1}^{N}& {x}_{2}^{N}\end{array} \begin{array}{cc}\dots & {x}_{d}^{N}\end{array}\end{array}\right]$$
    (9)
    $$\mathrm{Memory}=\left[\begin{array}{c}\begin{array}{ccc}{m}_{1}^{1}& {m}_{2}^{1}& \dots \\ {m}_{1}^{2}& {m}_{2}^{2}& \dots \\ \vdots & \vdots & \vdots \end{array} \begin{array}{c}{m}_{d}^{1}\\ {m}_{d}^{2}\\ \vdots \end{array}\\ \begin{array}{cc}{m}_{1}^{N}& {m}_{2}^{N}\end{array} \begin{array}{cc}\dots & {m}_{d}^{N}\end{array}\end{array}\right]$$
    (10)
  • Step 3: Evaluation of the fitness function so that per each crow, the quality of its position is determined based on decision variable and objective function values.

  • Step 4: Generating new positions in this way that if crow i wants to generate a new position, it should select randomized of the flock crows (for example j) and attempt to find the position of foods that are hidden by this crow (mj)

  • Step 5: Checking capability of the new location for each crow.

  • Step 6: Assessing the objective function of new position per each crow

  • Step 7: crows' memory will be updated using Eq. 11.

    $$m^{{i,{\text{Iter}} + 1}} = \left\{ {\begin{array}{*{20}l} {x^{{i,{\text{iter}} + 1}} \,\,\,f(x^{{i,{\text{iter}} + 1}} )\,\,is\;\;better\,\,than\,\,\,f(m^{{i,{\text{iter}}}} )} \\ {m^{{i,{\text{iter}}}} \,\,\,otherwise} \\ \end{array} } \right.$$
    (11)

In the above equation, f(.) represents the value of objective function. Mi, iter denote the memory of crow i in iteration iter and xi, iter is the position of crow i in iteration (iter). In the final step, the termination criterion is checked by repeating steps 4–7 to reach itermax. Figure 3 presents flowchart of the CSA optimization method.

Fig. 3
figure 3

Flowchart of crow search algorithm

2.5 Cat swarm Optimization (CSO)

Although resting behavior of cats is generalized, they are highly alert about their environment and any moving object. This inherent behavior helps cats with finding their preys. The cats spend very little time for chasing prey with respect to their resting time. Chu and Tsai proposed their CSO method which was inspired by hunting pattern of the cats in two modes namely the "seeking mode" (for the resting time of cats) and "tracing mode" (for the time of chasing preys) [41].

In CSO, the number of populations comprised of cats is implemented and prepared randomized distribution in the N-dimensional solution space. This population is divided into two sub-groups of "seeking mode" where cats are resting and are looking around their surroundings. And the "tracing mode" where they begin to move around and chase their preys. The mixture of these two modes helps with global solution in the N-dimensional space. As the cats spend too little time in the tracing mode, their number in this subset is small. For this purpose, the Mixture Ratio (MR) is utilized which has a low value. After grouping the cats into these two modes, new positions and objective functions could be calculated, and the cat with the optimistic searching criteria is keep in the memory. The above stages are repeated till the stopping criteria are met [50]. Flowchart of cat swarm optimization algorithm is depicted in Fig. 4. For more information about the computational procedures of CSO [41] study have considered.

Fig. 4
figure 4

Flowchart of cat swarm optimization algorithm

2.6 Description of Proposed Evolutionary-Based MARS Trained by Optimization Algorithm

In present study, an evolutionary-based predictive methods were implemented and proposed. One of the main motivations of this study was to achieve the optimum values for three hyper-parameters of MARS model in the prediction of CS of HS concrete using evolutionary-based solutions. Three hyper-parameters of MARS that can affect the model's predictivity in terms of accuracy are Mmax, c, Imax. Among various evolutionary optimization techniques such as random search, Nelder–Mead search, genetic algorithms, grid search, pattern search, heuristic search, this study recruited two bio-inspired algorithms based on the strategic behavior of cats and crows, namely CSO and CSA, and a natural-inspired optimization algorithm (e.g., WCA)-based water cycle production to calculate the optimum hyper-parameters. The four main steps of the proposed adaptive machine learning techniques are described as following:

  • Step 1: The accuracy or performance of the hybrid MARS models was compared with evolutionary algorithm's greedy selector to assess the fitness function. In the present research, the MSE metric (objective function) was used as the stopping criteria. Hence, Eq. 12 was proposed as follows to calculate the summation of the training and testing RMSE for controlling model fitness function.

    $${\text{F}}\text{=}{\text{E}}_{\mathrm{Calibrate}}+{E}_{\mathrm{Validate}}$$
    (12)

Where in the above equation, Ecalibrate and Evalidate denote the calibration and validation procedure's error, respectively. In Eq. 12, mean square error is incorporated as the index for prediction error. In Eq. 12, the fitness function represents the balance achieved between the model complexity and generalization.

  • Step 2: The initial values of three hyper-parameters (i.e., Mmax, c, Imax) of MARS model for prediction of CS of HS concrete were randomly generated based on a uniform distribution, as given in Eq. 13. The initial values of the abovementioned hyper-parameters are expressed in the following formula.

    $$X=(\mathrm{Ub}-\mathrm{Lb})\times \mathrm{rand}+\mathrm{Lb}$$
    (13)
  • Step 3: The evolutionary algorithm's population during the optimization process included a large number of the variables which were the MARS’s hyper-parameters. At each generation, MSE as the candidate objective function of each set of hyper-parameters was calculated. The three algorithmic approaches, CSO, CSA, and WCA, discarded inferior set of hyper-parameters and gradually guided the population toward regions featuring good set of hyper-parameters. According to this mechanism, better solutions of bfs and c can be employed for modeling CS of HS concrete in the training phase of MARS model.

  • Step 4: The optimization process would stop when the considered rmse as the stopping criteria was achieved. Finally, the optimal prediction of CS with fittest parameter values was determined when the stop criterion was reached. The pseudocode of hybrid mars-based three evolutionary algorithms in predicting CS of HS concrete is proposed in Fig. 5.

Fig. 5
figure 5

Pseudo code of proposed evolutionary algorithms for MARS model optimization

2.7 Statistical Measures

In this study, the effectiveness of the WCA, CSO, and CSA algorithms to optimize MARS model is evaluated several performance metrics (Eqs. 1419) as follows,

  1. (1)

    Pearson correlation coefficient (R):

$$\begin{aligned}R&=\frac{{\sum }_{i=1}^{M}\left({\mathrm{CS}}_{\mathrm{Obs}}-\overline{{\mathrm{CS} }_{\mathrm{Obs}}}\right).\left({\mathrm{CS}}_{\mathrm{Pre}}-\overline{{\mathrm{CS} }_{\mathrm{Pre}}}\right)}{\sqrt{{\sum }_{i=1}^{M}{{(\mathrm{CS}}_{\mathrm{Obs}}-\overline{{\mathrm{CS} }_{\mathrm{Obs}}})}^{2}{\sum }_{i=1}^{M}{\left({\mathrm{CS}}_{\mathrm{Pre}}-\overline{{\mathrm{CS} }_{\mathrm{Pre}}}\right)}^{2}}},\\&\quad \mathrm{Range }= (0, 1);\mathrm{ Ideal\,\,value }= 1\end{aligned}$$
(14)
  1. (2)

    Nash–Sutcliffe Efficiency (NSE):

$$\begin{aligned}\mathrm{NSE}&=1-\frac{{\sum }_{i=1}^{M}{\left({\mathrm{CS}}_{\mathrm{Pre}}-{\mathrm{CS}}_{\mathrm{Obs}}\right)}^{2}}{{\sum }_{i=1}^{M}{\left({\mathrm{CS}}_{\mathrm{Obs}}-\overline{{\mathrm{CS} }_{\mathrm{Obs}}}\right)}^{2}},\\&\quad \mathrm{Range }= (-\infty, 1);\mathrm{ Ideal\,\,value }= 1\end{aligned}$$
(15)
  1. (3)

    Mean Absolute Error (MAE):

$$\begin{aligned}\mathrm{MAE}&=\frac{{\sum }_{i=1}^{M} \left|{\mathrm{CS}}_{\mathrm{Pre}}-{\mathrm{CS}}_{\mathrm{Obs}}\right|}{\mathrm{N}}, \\&\quad \mathrm{Range }= (-\infty, +\infty );\mathrm{ Ideal\,\,value }= 0\end{aligned}$$
(16)
  1. (4)

    Wilmot’s Index of agreement (WI):

$$\begin{aligned}\mathrm{WI}&=1-\frac{{\sum }_{i=1}^{M}{\left({\mathrm{CS}}_{\mathrm{Obs}}-{\mathrm{CS}}_{\mathrm{Pre}}\right)}^{2}}{{\sum }_{i=1}^{M}{\left(\left|{\mathrm{CS}}_{\mathrm{Pre}}-\overline{{\mathrm{CS} }_{\mathrm{Obs}}}\right|+\left|{\mathrm{CS}}_{\mathrm{Obs}}-\overline{{\mathrm{CS} }_{\mathrm{Obs}}}\right|\right)}^{2}},\\&\quad\mathrm{ Range }= (0, 1);\mathrm{ Ideal\,\,value }= 1\end{aligned}$$
(17)
  1. (5)

    Root Mean Square Error (RMSE):

$$\begin{aligned}\mathrm{RMSE}&=\sqrt{\frac{1}{N}{\sum }_{i=1}^{M}{\left({\mathrm{CS}}_{\mathrm{Pre}}-{\mathrm{CS}}_{\mathrm{Obs}}\right)}^{2}},\\&\quad\mathrm{Range }= (-\infty, +\infty );\mathrm{ Ideal\,\,value }= 0\end{aligned}$$
(18)
  1. (6)

    Legates–McCabe’s Index (LMI):

$$\begin{aligned}\mathrm{LMI}&=1-\left[\frac{{\sum }_{\mathrm{I}=1}^{\mathrm{M}}\left|{\mathrm{CS}}_{\mathrm{Pre}}-\overline{{\mathrm{CS} }_{\mathrm{Pre}}}\right|}{{\sum }_{\mathrm{I}=1}^{\mathrm{M}}\left|{\mathrm{CS}}_{\mathrm{Obs}}-\overline{{\mathrm{CS} }_{\mathrm{Obs}}}\right|}\right],\\&\quad \mathrm{Range }= (-\infty, 1);\mathrm{ Ideal\,\,value }= 1\end{aligned}$$
(19)

In the above equations, \({\mathrm{CS}}_{\mathrm{Obs}}\) and \({\mathrm{CS}}_{\mathrm{Pre}}\) denote the experimental and prediction of CS values, respectively. \(\overline{{\mathrm{CS} }_{\mathrm{Obs}}}\) and \(\overline{{\mathrm{CS} }_{\mathrm{Pre}}}\) are the average values of experimental and prediction, respectively. Also, M represent the total number of the samples.

3 Application Results and Discussion

3.1 MARS Formulation for CS Prediction

In this study, Jekabsons open source code which presented [47] was employed for development of piecewise-linear MARS. The best performance of MARS model for CS of HSNC possessed 8 basis functions, while the maximum interaction level was set in second-order interaction. In the present research, smooth parameters investigated in default value (1–8) recommended by Jekabsons and best value was defined cbest = 1 based on minimum mean square error. Total number of effective parameters was 18.5, and GCV was 4.102. The details of BFs are reported in Table 4. Moreover, in this study, cross-validation method was utilized for prevention of model capability assessing bias. Table 5 presented Analysis of Variance (ANOVA) that is known for assessing the interactions of predictors and recognizing the most influential ones using GCV evaluation metric. The interpretable nonparametric formula-based model was defined as following:

Table 4 Basis functions of developed MARS method
Table 5 Analysis of variance for input variables of MARS
$$\begin{aligned}{\mathbf{C}\mathbf{S}}_{\mathbf{M}\mathbf{A}\mathbf{R}\mathbf{S}} &= 57.768 -0.058\times \mathrm{BF}1 -0.852\times \mathrm{BF}2 \\&\quad+0.967\times \mathrm{BF}3 -0.107\times \mathrm{BF}4 +0.073\\&\quad\times \mathrm{BF}5 -0.102\times \mathrm{BF}6 +0.111\times \mathrm{BF}7\end{aligned}$$
(20)

3.2 Evolutionary-MARS Formulation for CS Prediction

In computer sciences, accurate selection of parameters corresponding to the soft computing methods is the very important procedure in achieving high performance. For example, in ANN model weights, neurons, layers and biases are among the crucial control parameters that should be determined. Even if in addressing a problem, the model provides good results but an incorrect selection of parameters could result into undesirable outcome. One technique to find the appropriate parameters is combination of the previous experiences with a restricted heuristic search of optimistic solutions. But this may take a lot of time and not yield a result.

MARS model is dependent on the parameters included in it such as the smooth parameter c, maximum basis function Mmax, and maximum interaction Imax. On the other hand, finding optimal parameters simultaneously is very difficult due to the many available choices, but finding appropriate parameters can significantly improve MARS capability in predicting the procedure although suitable values may fall outside the suggested ranges. Therefore, in this study, evolutionary MARS model is incorporated to help users to face this challenge.

At the first stage, MARS is incorporated to handle the underlying function. Then, a new MARS method is built per each set of parameters that are provided by meta-heuristic algorithm (e. g., WCA, CSA and CSO). The greedy selector in aforementioned algorithms will compare the capability of proposed formula-based model with respect to the objective function assessment. After the end of train step, MARS model is incorporated to verify the data rows. It is obvious that due to good training of data we may face over-fitting. Therefore, combination of calibration error and validation error can result into an optimally balanced model with minimum errors. At the second stage, meta-heuristic algorithms applied to search for the best parameter setting value such as Mmax, Imax, and c. When the stopping condition is met, the optimization would be terminated. In this study, the generation number is depicted to present the convergence rate of each optimization algorithm up to 100 generations. Ultimately, when the stopping criterion is satisfied, the optimal predictive model with the fittest parameter setting is obtained. So, evolutionary MARS has accomplished the calibrating process and is now prepared to simulate new input patterns. The related equations of BFs of evolutionary MARS are indicated in Table 6 for formulation of CS. The interpretable evolutionary MARS predictive formulas were developed as following:

Table 6 Basis functions of developed evolutionary MARS methods
$$\begin{aligned}{\mathbf{C}\mathbf{S}}_{\mathbf{M}\mathbf{A}\mathbf{R}\mathbf{S}-\mathbf{W}\mathbf{C}\mathbf{A}} &= 52.721 -0.070\times \mathrm{BF}1 -1.549\\&\quad \times \mathrm{BF}2 -0.056\times \mathrm{BF}3 -0.087\\&\quad\times \mathrm{BF}4 +0.015 \times \mathrm{BF}5 -0.005\\&\quad \times \mathrm{BF}6 -0.003\times \mathrm{BF}7 +0.004\\&\quad \times \mathrm{BF}8 +0.0002 \times \mathrm{BF}9 +0.003\\&\quad\times \mathrm{BF}10 +0.005 \times \mathrm{BF}11 -0.001\\&\quad\times \mathrm{BF}12 +0.001 \times \mathrm{BF}13 -0.003\\&\quad\times \mathrm{BF}14 +0.0007\times \mathrm{BF}15 +0.001\\&\quad \times \mathrm{BF}16 +0.0006\times \mathrm{BF}17 +0.0007\\&\quad \times \mathrm{BF}18 -0.002 \mathrm{BF}19 +0.033\\&\quad \times\times \mathrm{BF}20 -0.071 \mathrm{BF}21 -0.004\\&\quad \times \mathrm{BF}22 -0.012 \times \mathrm{BF}23 +0.003\\&\quad\times \mathrm{BF}24 +0.022 \times \mathrm{BF}25 -0.057\\&\quad\times \mathrm{BF}26 +0.070 \times \mathrm{BF}27 +0.055\\&\quad\times \mathrm{BF}28\end{aligned}$$
(21)
$$\begin{aligned}{\mathbf{C}\mathbf{S}}_{\mathbf{M}\mathbf{A}\mathbf{R}\mathbf{S}-\mathbf{C}\mathbf{S}\mathbf{A}} &= 59.491 -0.073\times \mathrm{BF}1 -1.339\\&\quad\times \mathrm{BF}2 +1.173\times \mathrm{BF}3 -0.114\\&\quad\times \mathrm{BF}4 +0.053\times \mathrm{BF}5 -0.088\\&\quad\times \mathrm{BF}6 +0.004\times \mathrm{BF}7 +0.003\\&\quad\times \mathrm{BF}8 -0.005\times \mathrm{BF}9 -0.002\\&\quad\times \mathrm{BF}10 +0.006\times \mathrm{BF}11 +0.0002\\&\quad\times \mathrm{BF}12 +0.003\times \mathrm{BF}13 +0.006\\&\quad\times \mathrm{BF}14 -0.001\times \mathrm{BF}15 +0.001\\&\quad\times \mathrm{BF}16 -0.004\times \mathrm{BF}17 +0.0008\\&\quad\times \mathrm{BF}18 +0.001\times \mathrm{BF}19 +0.0007\\&\quad\times \mathrm{BF}20 +0.0007\times \mathrm{BF}21 -0.002\\&\quad\times \mathrm{BF}22 -0.001\times \mathrm{BF}23 -0.002\\&\quad\times \mathrm{BF}24 -0.003\times \mathrm{BF}25\end{aligned}$$
(22)
$$\begin{aligned}{\mathbf{C}\mathbf{S}}_{\mathbf{M}\mathbf{A}\mathbf{R}\mathbf{S}-\mathbf{C}\mathbf{S}\mathbf{O}} &= 58.646 -0.067\times \mathrm{BF}1 -1.029\\&\quad\times \mathrm{BF}2 +1.091\times \mathrm{BF}3 -0.111\\&\quad\times \mathrm{BF}4 +0.066\times \mathrm{BF}5 -0.095\\&\quad\times \mathrm{BF}6 +0.084\times \mathrm{BF}7 +0.002\\&\quad\times \mathrm{BF}8 +0.003\times \mathrm{BF}9 +0.003\\&\quad\times \mathrm{BF}10 -0.004\times \mathrm{BF}11 -0.003\\&\quad\times \mathrm{BF}12 +0.005\times \mathrm{BF}13 +0.0002\\&\quad\times \mathrm{BF}14 +0.0001\times \mathrm{BF}15\end{aligned}$$
(23)

Overall, the proposed approach of this paper is presented new insight into the investigating of the compressive strength of the HS concrete using the experimental mixture variables. To do this, two-step procedure is implemented; (1) compute the BFs of the problem as per Table 6. The interpretable MARS method generates many of BFs for experimental variables. A model is then learned from the output of each of BFs with the target variable. This means that the output of each BFs is weighted by a coefficient. A prediction is made by summing the weighted output of all of the BFs in the model. (2) Compute the CS applying Eqs. 2023 depending on the MARS, MARS-WCA, MARS-CSA and MARS-CSO developed methods.

3.3 Results Comparison of Evolutionary and Benchmark Methods

The predictability and suitability of evolutionary nonparametric multivariate paradigm named MARS-WCA, MARS-CSA, MARS-CSO and standalone MARS was evaluated for compressive strength of HS concrete formulation. The six-performance metrics (R, NSE, MAE, WI, RMSE, and LMI) of the developed machine learning models on the formulation of CS is represented in Table 6. Before evaluating the results extracted by standalone and hybrid models, the overfitting issue for every model should be checked and controlled. One of the simple approaches that can assess this issue is the difference of models' performance in calibration and validation stages. When overfitting occurs, the performance metrics of the predictive models are satisfactory in the calibration stage; however, the significant weakness can be seen in the validation stage. For instance, when the difference in the indices R, NSE, and WI is less than 5%, the probability of overfitting occurrence is quite low. Therefore, overfitting of the proposed models based on the values represented in Table 7 has been avoided.

Table 7 Performance capability of proposed soft computing models

Based on the results, MARS-WCA model (R = 0.994 and WI = 0.997, NSE = 0.988) and (R = 0.991 and WI = 0.995, NSE = 0.981), outperformed other machine learning methods in both calibration and validation stages, respectively. MARS-WCA has the lowest RMSE (0.991 MPa) and the highest LMI (0.906); it enhanced the precision of the validation stage to formulate CS in terms of LMI of the MARS-CSA, MARS-CSO and standalone MARS by 1.1%, 4.9% and 7.1%, respectively. Moreover, for further evaluation of the capacity of proposed evolutionary and standalone MARS, objective function (OBJ) index was computed. The optimistic predictive formula-based models are chosen based on OBJ through the best model’s predictability. An OBJ (with range of (0, + ∞) and optimum value of zero) is measured as follows:

$$\begin{aligned}\mathrm{OBJ}&=\left(\frac{{\mathrm{No}.}_{\mathrm{training}}-{\mathrm{No}.}_{\mathrm{testing}}}{{\mathrm{NO}.}_{\mathrm{all}}}\right)\frac{{\mathrm{MAE}}_{\mathrm{training}}}{{R}_{\mathrm{training}}^{2}}\\&\quad+\frac{{2\mathrm{No}.}_{\mathrm{testing}}}{{\mathrm{NO}.}_{\mathrm{all}}}\times \frac{{\mathrm{MAE}}_{\mathrm{testing}}}{{R}_{\mathrm{testing}}^{2}}\end{aligned}$$
(24)

The comparison between the simulated and observation CS values has indicated that the developed MARS-WCA model (with OBJ = 0.932) presented more efficiency beside another models (OBJMARS = 1.734, OBJMARS-CSA = 1.038 and OBJMARS-CSO = 1.379).

Figures 6 and 7 illustrated the comparison of the predicted CS values that are modeled with standalone MARS model and three hybrid MARS models coupling with WCA, CSA and CSO algorithms. According to these figures, the MARS-WCA model has better visual agreement between actual and predicted values and the highest accuracy in the formulation of CS during the train and test steps based on y = x (black and dotted) lines. Figure 8 shows that hybrid MARS model coupled with WCA could predict the local maximum and minimum slightly higher than other standalone and hybrid MARS models. It can trustworthy express that the compressive strength that predicted by evolutionary models were in robustness coherence with the actual data points.

Fig. 6
figure 6

Scatter plots of experimental and formulated CS for training performance

Fig. 7
figure 7

Scatter plots of experimental and formulated CS for testing performance

Fig. 8
figure 8

Curve fitting plots of experimental and formulated CS of HS Concrete

The estimation errors of CS models of HSNC in Fig. 9 presented significantly improvement of error production in model-based WCA integrated with MARS compared with other proposed models. Distribution plot of the relative errors of the standalone and evolutionary MARS models of the CS of HSNC to propagate the magnitude is presented in Fig. 10. The histogram values relative error of MARS-WCA, MARS-CSA, MARS-CSO and standalone MARS models in the interval of [-3%, 3%] account for 87.9%, 84.50%, 83.00% and 58.30% of the total samples. The number of relative error values estimated by the proposed models in the range of [-6%, 6%] further reduce. Through error interval analytical process, it is found that the estimation errors of MARS-WCA (mean = -0.004 and Standard Deviation (StDev) = 2.224) lower than other evolutionary and standalone models.

Fig. 9
figure 9

Graphical comparisons of error in model’s development; a: Residual, b: Relative error for validation stage

Fig. 10
figure 10

Relative error histograms of proposed models

Figure 11 revealed the mean square error values as an objective function against the iteration cycles for the MARS integrated with meta-heuristic algorithms. Through Fig. 11, as the algorithm searching and optimizing the MARS parameters, the obtained results satisfied the constraints, while the value of constraint violation decreased. In this part, the CS of HS concrete estimated by the MARS-WCA, MARS, CSA and MARS-CSO models. As those evolutionary application for the hybrid models, the WCA, CSA and CSO were utilized to optimize the Mmax, Imax and c. The parameters of WCA (e.g., Number of rivers + sea (Nsr) and Evaporation condition constant (dmax)), CSA (e. g., Flight Length (FL) and Awareness Probability (AP)) and CSO (e. g., Seeking Memory Pool (SMP) and Seeking Range of the selected Dimension (SRD)) are also needed to be tune up before optimization procedure of MARS method.

Fig. 11
figure 11

Meta-heuristic algorithms performance to optimize MARS parameters

Based on Fig. 11, the speed of convergence of WCA after 100 generations is very high in probability of searching the optimum solution (decision variables) compared to CSA and CSO algorithms in same populations [46]. It is obvious that the integration of MARS with WCA also provides efficient and reliable prediction for CS of HS concrete, and the well-known optimizer in terms of computational effort (Nsr = 4 and dmax = 1e-16) that are challenging for regularizing the ability of the WCA algorithm [51]. As can be seen from Fig. 8, the selected hyper-tuning parameters is reported FL = 3 and AP = 0.2 (for CSA), and SMP = 5 and SRD = 1 (for CSO). Finally, for further evaluation of influence of input variables on the CS of HS concrete proposed model, 3D surface diagrams for MARS-WCA (optimistic model) are presented in Fig. 12.

Fig. 12
figure 12

3D surface diagrams of MARS-WCA model for evaluation of inputs

To compare the results of the proposed MARS-WCA model with other AI models for CS prediction of HS concrete, the best performance model of Al-Shamiri's study [44], Back-Propagation Network (BPN) was selected. Due to the proximity of the modeling results in terms of R, RMSE, NSE and MAE, a composite metric (i.e., Performance Index (PI)) that involves two R and RMSE criteria was considered to combine and generate a single result for better comparison. The PI statistical metric changes from zero to one, with smaller values, presented a better prediction and is calculated as following:

$$\mathrm{PI}=\frac{1}{\left|\overline{{\mathrm{CS} }_{\mathrm{obs}}}\right|}\frac{\mathrm{RMSE}}{(R+1)}$$
(25)

Comparing the results of integrated MARS-WCA with BPN, it was found that the performance of both models is similar in terms of PI where the values for MARS-WCA and BPN are 0.109 and 0.104, respectively, for the prediction of CS of HS concrete. Moreover, the evaluation metrics indicate that applying optimization algorithms like WCA would improve the models’ performances remarkably. The most important point in comparison of MARS-WCA and Al-Shamiri's method was the applicability of the BPN model which cannot be used by researchers in the specific field of study. In other words, in this study, an explicit formulation was presented by integrated MARS-WCA which can be apply in order to predict and determine the compressive strength of high-strength concrete.

3.4 Validation of Formula-Based Models

3.4.1 External Analysis of the Predictive Models

Validation is a crucial aspect of any quantitative structure–activity relationship modeling and a special emphasis on statistical significance and predictive ability of those models as their most crucial characteristics. Checking the robustness of the predictive models just with coefficient determination (R2 > 0.5) could not provide the ultimate proof of the model's ability or capability for prediction of phenomena, and it needs an approach to stablish the model robustness by randomization of response (i.e., compressive strength) [52]. External analysis as a reliable analytical method is used for comparison between the results of estimated and observation event data. Golbraikh and Tropsha [53] have adopted the new external validation criteria for evaluation of the estimation capacity of the models corresponding to the performance of testing subset based on coefficient of determination. External validation means assessing the model performance with independent samples [52]. In this method, minimum one of the coefficient of determination regression line gradients that passes through the source for estimated values against experimental CS or vice versa should be close to unity.

$$K=\sum_{i=1}^{n}\frac{{\mathrm{CS}}_{\mathrm{obs}}\times {\mathrm{CS}}_{\mathrm{pre}}}{{\mathrm{CS}}_{\mathrm{pre}}^{2}}$$
(26)
$${K}^{\mathrm{^{\prime}}}=\frac{{\mathrm{CS}}_{\mathrm{obs}}\times {\mathrm{CS}}_{\mathrm{pre}}}{{\mathrm{CS}}_{\mathrm{obs}}^{2}}$$
(27)

\({\mathrm{CS}}_{\mathrm{obs}}\) and \({\mathrm{CS}}_{\mathrm{pre}}\) represent the experimental and formulated CS values, respectively. The determination coefficients passing through the source between the predicted and experimental values (\({\mathrm{R}}_{0}^{2}\)) and conversely (\({\mathrm{R}}_{0}^{\mathrm{^{\prime}}2}\)) are derived using the following equations:

$${R}_{0}^{2}=1-\sum_{i=1}^{n}\frac{{\mathrm{CS}}_{\mathrm{pre}}^{2}{\left(1-\mathrm{k}\right)}^{2}}{\sum_{i=1}^{n}{{\mathrm{CS}}_{\mathrm{pre}}-\overline{{\mathrm{CS} }_{\mathrm{pre}}} )}^{2}}$$
(28)
$$R_{0} ^{{{\prime }{\text{2}}}} = {\text{1}}-\sum\limits_{{{\text{i = 1}}}}^{{\text{n}}} {{\text{CS}}_{{{\text{obs}}}}^{{\text{2}}} } ({\text{1}-\text{k}}^{{\prime }} )^{{\text{2}}} {\text{/}}\sum\limits_{{{\text{i = 1}}}}^{{\text{n}}} {({\text{CS}}_{{{\text{obs}}}} {{-}}\overline{{{\text{CS}}}} _{{{\text{obs}}}} )^{{\text{2}}} }$$
(29)
$$m=\left({R}^{2}-{R}_{0}^{2}\right) / \, {\text{R}}^{2}$$
(30)
$$n = {\text{(}}R^{2} - R_{0} ^{{{\prime }2}} {\text{)/}}{\mkern 1mu} {\text{R}}^{{\text{2}}}$$
(31)

As the result of verification metric through the analysis based on slope of the regression (\({R}_{m})\) should be greater than 0.5.

$${R}_{m}={R}^{2}\times \left(1-\sqrt{\left|{R}^{2}-{R}_{0}^{2}\right|}\right)>0.5$$
(32)

The validation measures and the related performance of CS formulation obtained by various models are reported in Table 8. Based on the results, the MARS-WCA models for compressive strength which yielded Rm = 0.852 satisfied the conditions with best validation with respect to other used evolutionary and standalone SC methods such as MARS-CSA, MARS-CSO and standalone MARS models. Therefore, it is found that MARS-WCA has the highest validity for predicting CS of HS concrete and the computed correlations had not been accidentally.

Table 8 External analysis measures of predictive models

3.4.2 Monte Carlo Uncertainty Analysis

In this sub-section, the randomness of model’s uncertainty is performed by Monte Carlo Simulation (MCS). This procedure was first applied in military projects for simulation of the probabilistic events [51]. In this content, the target variable of the presented study (i.e., compressive strength) comprises several uncertainties like parameters uncertainty of the models, uncertainty of input variables, etc. Therefore, quantitative uncertainty of the MARS-WCA, MARS-CSA, MARS-CSO and standalone MARS models that associated with estimated CS is investigated. Based on this technique, the individual error of the predictions is calculated for all datasets (Eq. 31). The mean \((\overline{e })\) and standard deviation \({(S}_{e})\) of the formulation error is calculated by Eqs. 32 and 33, respectively [48]:

$${e}_{i}={\mathrm{log}}_{10}({{\mathrm{CS}}^{\mathrm{pre}}}_{i})-{\mathrm{log}}_{10}({{\mathrm{CS}}^{\mathrm{obs}}}_{i})$$
(33)
$$\overline{e }=\sum_{i=1}^{n}{e}_{i}$$
(34)
$${S}_{e}=\sqrt{\sum_{i=1}^{n}\left(\frac{{\left({e}_{i}-\overline{e }\right)}^{2}}{n-1}\right)}$$
(35)

Also, \(\pm 1.96 {S}_{e}\) yields 95% confidence level around predicted Pi as follows:

$$\left.\left\{{P}_{i}\times {10}^{-\overline{{\varvec{e}} }-1.96{{\varvec{S}}}_{{\varvec{e}}}}\right.,{P}_{i}\times {10}^{-\overline{{\varvec{e}} }+1.96{{\varvec{S}}}_{{\varvec{e}}}}\right\}$$
(36)

The uncertainty percentage and Mean Absolute Deviation (MAD) as the two main factors in assessing the uncertainty of the proposed models are reported in Table 9. Based on the result, the positive mean prediction error indicates that the predicted CS calculated by all these methods is higher than the observations. In addition, MARS-WCA and MARS-CSA models for CS yielded the minimum (15.26% and 15.67%) bandwidth uncertainties, respectively. Moreover, in other developed models, MARS-WCA satisfied bandwidth criteria and had lowest uncertainty.

Table 9 Monte Carlo result for uncertainty of models

3.5 Variable importance

In this study, to determine the difference of predictor’s value that will affected on target, analysis of variables importance (AVI) technique is recruited. The AVI% is calculated for each independent variable as follows [54]:

$${L}_{i}={\mathrm{CS}}_{\mathrm{max}}\left({x}_{i}\right)-{\mathrm{CS}}_{\mathrm{min}}({x}_{i})$$
(37)
$${\mathrm{AVI}}_{i}=\frac{{L}_{i}}{\sum_{j=1}^{M}{L}_{i}}\times 100$$
(38)

where \({\mathrm{CS}}_{\mathrm{max}}\) and \({\mathrm{CS}}_{\mathrm{min}}\) = maximum and minimum of the predicted compressive strength over the ith input domain, where other independent variable values are equal to their average values. Figure 13 indicated the variable importance results for the simulation of CS of HS concrete using MARS-WCA model that was selected as the best model among other hybrid and standalone models. The bar plot presented that the most effective variable in CS of HS concrete is the cement content (30.7%) and water content (20.5%).

Fig. 13
figure 13

Results of analysis of variables importance for corresponding compressive strength

4 Conclusion

Achieving a high-accurate model to predict compressive strength is very important especially, in the field of concrete technology. This paper evaluates the ability of the evolutionary algorithms, WCA, CSA and CSO, for optimizing hyperparameter of MARS model in order to extract new accurate formulas for the compressive strength of HS concrete. Through the results of this research, the following conclusion can be presented:

  • New closed-from evolutionary nonparametric multivariate regression-based formula is proposed for the compressive strength of HS concrete. The overall nonlinear formulations given by non–parametric paradigm are presented with 29, 26, 16 and 8 BFs (Eqs. 2022 and 18) for MARS-WCA, MARS-CSA, MARS-CSO and standalone MARS, respectively.

  • Proposed CS developed models manifested that the MARS-WCA model presented more accurate prediction in comprising with the other three evolutionary and standalone techniques, with respect to R, NSE, RMSE, MAE, WI, and LMI indicators for calibrating and validating subsets. As the results of the evolutionary and original models were similar, OBJ as the composite metric is recruited to illustrate the superiority of the model. The integrated MARS-WCA model with OBJ = 0.932 presented more efficiency compared to other models (OBJ of MARS = 1.734, OBJ of MARS-CSA = 1.038 and OBJ of MARS-CSO = 1.379).

  • The proposed evolutionary models as novel machine learning tools confirmed all of the required criterion of the statistical regression line gradients for external validation. Results indicated that MARS-WCA which yielded K and k' near 1 and Rm equal to 0.852 satisfied the conditions with best validation with respect to MARS-CSA, MARS-CSO and standalone MARS models.

  • The Monte Carlo uncertainty procedure for implemented evolutionary models was employed. MARS-WCA with 15.26% and MARS-CSA with 15.67% bandwidth uncertainties yielded the minimum uncertainty for predicting CS of HS concrete. The robustness of the proposed evolutionary machine learning techniques was verified in this study. Moreover, sensitivity analysis of variable importance considered the highest importance variables influenced on the compressive strength of HS concrete to be the cement content and water content with 30.7 and 20.5%.