1 Introduction

Portland cement-based concrete is the second most-used material after water to accomplish rapid construction throughout the world (Feng et al. 2020). Portland cement manufacture produces large emissions in addition to requiring a lot of energy. In other words, the production of one tonne of cement produces roughly one tonne of carbon emissions (Alsalman et al. 2021). These vast carbon and other hazardous substances emission into the atmosphere creates various environmental concerns (Oliveira et al. 2019, 2020). To check on these environmental issues created by Portland cement production, various remedial measures have been explored by several researchers (Ahmed et al. 2022b, d, 2023b; Qaidi et al. 2022b, c; Smirnova et al. 2021; Unis Ahmed et al. 2022; Pradhan et al. 2022d, 2023a, b) and among those use of supplementary materials, use of alkali-activated materials, and geopolymer concrete are notable. Geopolymers are inorganic polymeric materials produced using some alumino-silicate materials as precursors and some reagents, which were discovered in the 1970s (Davidovits 1976). The world has seen enormous research since the 1990s on the development of geopolymer concrete (GPC) using various sources of precursors such as natural (clays, kaolin, metakaolin), industrial wastes (fly ash, ground granulated blast furnace slag or GGBFS, red mud), and agricultural wastes (sugarcane ash, rice husk ash, etc.) and various alkaline as well as acid solutions as chemical regents (Ahmed et al. 2023a, c; Das et al. 2023; Dwibedy and Panigrahi 2023; Pradhan et al. 2023a, b, 2024, Qaidi et al. 2022a, c; Unis Ahmed et al. 2023). To overcome the difficulty in the workmanship of GPC and alkali-activated concrete, a new form of GPC has also been explored which has a self-compacting nature and is termed Self-compacting geopolymer concrete and SC-AASC (Pradhan et al. 2022c). The SC-AASC can fill the dense reinforced formwork without segregation (EFNARC 2002). SC-AASC usage ensures a reduction in construction cost, labor, and energy consumption so it is highly beneficiary for modern-day infrastructural development. The GPC and alkali-activated concrete have been established as a substitute for Portland cement concrete because of their mechanical and durability behaviors (Pradhan et al. 2022a). However, it is not in use in practical construction fields yet. The primary reason is the variations in the properties of starting materials as it creates difficulty in controlling the GPC properties. Prediction of strength characteristics from its constituting material has been practiced so that wastage of time, materials, energy, and repetition of tests can be avoided (Ahmed et al. 2021, 2022a, b, c, d, e, f; Faraj et al. 2022a). The factors governing the properties are vast (Pradhan et al. 2022b) and simple linear regressions can’t provide an acceptable relation for the prediction of compressive strength. Machine Learning and Deep Learning techniques are extensively employed for tasks such as database classification and regression analysis. This is primarily due to their capacity to grasp intricate patterns within data (Ongsulee 2017). Various researchers have applied machine learning to predict concrete properties (Ahmed et al. 2022c, d, 2023b; Faraj et al. 2022a, b). Nhat-Duc (2023) used a deep learning approach to predict the compressive strength of GGBFS-based concrete. Zhang et al. (2023) used hybrid algorithms for compressive strength prediction of ultra-great workability concrete. Hu (2023) used a novel hybrid SVR model for compressive strength evaluation. Recently, various metaheuristic techniques have been used for hyperparameter optimization of machine learning models for concrete properties prediction (Dash et al. 2023; Parhi and Panigrahi 2023; Parhi and Patro 2023a, b; Singh et al. 2023). Compressive strength is the most important criterion in any type of concrete (Parhi et al. 2023). Compressive strength prediction of ecofriendly plastic sand paver blocks using gene expression programming (GEP) and multi-expression programming (MEP) has been done by various researchers (Iftikhar et al. 2023a, b). Chen et al. (2023a) employed GEP, decision tree (DT), multilayer perceptron neural network (MLPNN), and support vector machine (SVM) for predicting the reduction of compressive strength due to acid attack in egg-shell and glass waste incorporated cement-based composite. Zou et al. (2023) used GEP and MEP for slump and strength prediction of alkali-derived concrete. Utilization of GEP was further found in the strength prediction of waste-derived self-compacting concrete by Chen et al. (2023b). Alsharari et al. (2023) used a decision tree (DT), Adaboost (ADB), and bagging regressor (BR) approach for strength estimation of concrete at elevated temperatures. Random forest (RF) and GEP were used by Qureshi et al. (2023) for compressive strength prediction of preplaced aggregate concrete.

This research aims to meet stringent performance requirements in modern construction infrastructure by producing sustainable, green, and noise-free concrete. To facilitate these unique characteristics for a wide range of applications a method was proposed to develop HS-self-compacting alkali activated slag concrete in this manuscript. Critical sustainable factors contributing to the workability and strength of concrete have been identified and varied to obtain high-performance concrete. Further different machine learning techniques were used to optimize the mix proportion for developed high-performance concrete by predicting the strength properties of its constituent materials.

1.1 Research significance and essence of the manuscript

The manuscript presents a pioneering study in the field of concrete technology and machine learning, focusing on the development of High-Performance Self-Compacting alkali-activated slag concrete. This research holds significant importance and contributes to both the concrete industry and the domain of predictive modeling. The development of High-Performance SC-AASC represents an innovative approach to sustainable construction materials. This has the potential to reduce carbon emissions and environmental impact, addressing critical concerns in the construction industry. From a base, M25 SC-AASC mix design, M45, and M60 high strength SC-AASC were prepared by exploring the variation of critical strength and workability factors. Understanding how these factors impact the concrete's performance is essential for optimizing its use in real-world construction applications. Metaheuristic Spotted Hyena-optimized Machine Learning models were used for predicting the strength properties of SC-AASC. The manuscript adopts advanced computational techniques to enhance the accuracy of strength predictions, potentially revolutionizing quality control and design processes in the construction industry. The research also employs global sensitivity analysis methods, such as Sobol and FAST, to identify the parameters that influence the strength of SC-AASC. This analysis provides invaluable insights into the underlying mechanisms governing the material’s performance. The research aligns with global efforts to promote sustainability in construction. The development of a more environmentally friendly concrete mix and the use of advanced modeling techniques contribute to reducing the ecological footprint of construction projects. The manuscript bridges the gap between materials science, construction engineering, and data science, showcasing the potential of interdisciplinary research to address complex challenges in the construction sector.

2 Materials and methodology

2.1 Material collection

For the preparation of high-strength workable concrete, Ground Granulated Blast furnace slag (GGBFS) was used as the sole binder. GGBFS of specific gravity 2.8 was supplied by TATA Steel, Jajpur, Odisha. Table 1 represents the oxide composition of GGBFS. For the preparation of an alkaline activator solution (AAS), a combination of sodium hydroxide (SH) and sodium silicate (SS) of specific gravity 2.13 and 1.4 was used. Locally available natural coarse aggregate and fine aggregate were collected. SIKA VISCOFLOW 4005NS of specific gravity 1.11 was used as a superplasticizer (SP) for the production of higher-strength workable concrete. To maintain workability clean tap water was used as extra water (EW).

Table 1 Oxide composition of GGBFS (% wt)

2.2 Mix proportion

In this study, the mix design procedure for self-compacting concrete as per IS 10262:2019 has been adopted. From the previous studies to prepare higher strength workable concrete, some parametric values have been identified like molarity of SH as 12 M, AAS ratio (SS/SH) as 2.5 (Nuruddin et al. 2011; Pradhan et al. 2022a, b, c) with a variation of percentage of superplasticizer as 6% and 7%, (Jithendra and Elavenil 2019; Memon et al. 2012) and extra water as 21% and 24% (Muraleedharan and Nadir 2021; Patel and Shah 2018). During the progress of high strength, mix designs were prepared from standard base SC-AASC i.e., M25 to M45 to M60 with the variation of two principal strength contributing parameters like AAS content and AAS/Binder ratio. Nine M25 mix designs were prepared with successive reductions of AAS content i.e., 220–210 kg/m3, and AAS/Binder ratio i.e., 0.47–0.43. After getting workable M25 concrete, similarly, nine numbers of M45 and M60 mix designs were prepared with regular reduction of AAS content i.e., 210–200 kg/m3 and 200–190 kg/m3 along with a decrement of AAS/Binder ratio i.e., 0.41–0.39 and 0.37–0.36 respectively. Figure 1 represents the mix-design progress from M25 to M60. The detailed mix proportions of M25, M45, and M60 are shown in Tables 2, 3, and 4.

Fig. 1
figure 1

Mix design progression from M25 to M60

Table 2 Details of mix proportions of M25 SC-AASC
Table 3 Details of mix proportions of M45 SC-AASC
Table 4 Details of mix proportions of M60 SC-AASC

2.3 Preparation, casting, and curing

For the preparation of fresh SC-AASC, materials like GGBFS, and aggregates were weighted properly. Aggregates were collected in a saturated surface dry condition. A dry mix of Ground Granulated Blast Furnace Slag (GGBFS) and these aggregates was prepared by mixing them for 2–3 min. A Sodium Hydroxide (SH) solution was prepared 24 h before casting. Two hours before casting, the Sodium Silicate (SS) solution was added to the SH solution to create an Alkali-Activated Slag (AAS) solution. The AAS solution was then added to the dry mix of GGBFS and aggregates. Extra water and a superplasticizer were added to the mixture concerning the mix's workability. The materials were thoroughly mixed for 3–5 min to ensure uniformity and homogeneity. Freshly prepared SC-AASC was tested for workability. SC-AASC was cast into cubes with sides measuring 150 mm and cylinders with a diameter of 150 mm and a height of 300 mm. The specimens were cured at 14-, 28-, and 56-day ambient temperature for mechanical property study. The detailed procedure for the production of SC-AASC is represented in Fig. 2.

Fig. 2
figure 2

Production procedure of self-compacting alkali-activated slag concrete

2.4 Test methodology

The freshly prepared concrete was tested using four fresh property tests. The Slump Flow Test was used to evaluate the workability and flow characteristics of self-compacting AASC by measuring the spread of the concrete when it flows freely. The V-Funnel Test assesses the filling ability of SC-AASC by measuring the time it takes to flow through a V-shaped funnel, indicating its flow properties. The J-Ring Test measures the ability of SC-AASC to pass through obstructions, assessing its flowability in the presence of reinforcement. The V-Funnel at T5 Minutes Test accesses the resistance to segregation during pouring. The compressive Strength Test was used to determine the maximum load SC-AASC can withstand under axial compression, indicating its overall strength, and split-tensile strength was used to evaluate its resistance to cracking. Figure 3 represents all the fresh and hardened tests conducted along with casted specimens.

Fig. 3
figure 3

Fresh and hardened tests conducted on specimens

2.5 Experimental test results

The freshly prepared concrete was tested for filling ability test (slump flow, V funnel test), Passing ability test (J ring), and resistance to segregation test (V-funnel at T5minute) of workability as per EFNARC guidelines. After 14, 28, and 56 days of ambient curing, the specimens were tested for mechanical properties of compressive and split tensile tests. To avoid any misleading results, for each mix, three specimens were tested. The results of fresh and mechanical properties are shown in Table 5.

Table 5 Fresh and mechanical results of all the mixes

It is well observable that as the mix grade changes from M25 to M60, the slump flow generally decreases, indicating that higher-strength mixes are less flowable but still within an acceptable range for self-compacting concrete. The v-funnel test evaluates filling ability, Similar to slump flow results higher mix grades show slightly longer times (in seconds), indicating reduced filling ability with increasing strength. The J-ring test measures the ability of freshly prepared concrete to pass through the tight obstruction. It is a field test so can be performed onsite. The results show a slight increase in J-Ring values with higher mix grades, suggesting better resistance to segregation in stronger mixes. V-funnel test at T5 minutes provides insights into segregation resistance, stability, and viscosity. Similar to other workability test results this test also indicates reduced flowability over time for higher mix grades but within acceptable ranges. All the mix grades successfully pass the flowability and filling ability criteria for self-compacting concrete. The compressive strength and split-tensile strength results exhibited increasing mechanical strength with an increase in curing period and mix grades. The strong chemical bond formed due to the chemical activation of slag contributes to higher compressive strength compared to normal Portland-based concrete. Alkali-activated slag sometimes contains unreacted slag particles which makes the microstructure porous and prone to cracking (Sakulich et al. 2010). Due to this reason, a low value of split-tensile strength can be visualized in some mixed proportions. Also, self-compacting alkali-activated slag may require longer curing periods to achieve its optimal strength. In comparison, Portland cement concrete often gains strength more rapidly in the early stages. The results indicate satisfactory performance of SC-AASC for all the mix grades, hence the approach used to obtain the high-performance SC-AASC was found to be competent.

3 Database and exploratory data analysis

3.1 Database for model development

An adequate and dependable database holds equal significance alongside effective computational techniques. When simulating a model, the database must encompass a diverse array of input data. Variability in the experimental database arises from both epistemic and aleatoric errors, contributing to high variance. To serve the goals of research endeavors, a coherent database with 135 observations was generated through the Python interface of Google Colab. Refer to Table 6 for a summary of the descriptive statistics about the SC-AASC database, which was employed to construct models.

Table 6 Summary of experimental database used for modeling

Machine learning (ML) approaches are used to extract data information, unknown patterns, and the relationship between the dataset. ML has two kinds of techniques for prediction and modeling; one is a single separate model and another is known as an ensemble (Chou and Pham 2013). For the prediction of strength characteristics of M25, M45, and M60 SC-AASC; 12 inputs were given i.e., along with all the constituent parameters that produce SC-AASC and the important fresh property results which defy the essential characteristic of SC-AASC i.e., flowability was also taken as input data while compressive strength and split tensile strength were taken as output data. Using both single and ensemble ML algorithms compressive and split tensile strength were predicted.

3.2 Exploratory data analysis

Data analysis is an important step before proceeding with the creation of a model. Exploratory data analysis (EDA) helps in the initial investigation of data i.e., to identify patterns, get meaningful insights, check inconsistency and irregularity of data, and to check the hypothesis and parametric features with the assistance of statistics and graphs (Morgenthaler 2009). EDA was done before proceeding into ML model creation. Data analysis is an important step before proceeding with the creation of a model. Exploratory data analysis (EDA) helps in the initial investigation of data i.e., to identify patterns, get meaningful insights, check inconsistency and irregularity of data, and to check the hypothesis and parametric features with the assistance of statistics and graphs (Morgenthaler 2009). EDA was done before proceeding into ML model creation. The dataset containing missing values and outliers significantly decreases the accuracy of the models. Outlier mining is a data analysis process associated with the detection and removal of outliers in a dataset (Angiulli and Pizzuti 2005). In a dataset outliers are data points situated at an uncommon distance or having significant deviation from well-structured data (Wang et al. 2019). Commutation of system, mechanical and software faults, and human and instrumental errors give emergence to outliers (Hodge and Austin 2004). Outliers can be viewed by plotting the attributes in a box plot which shows the distribution and range of data for each attribute (Petrovskiy 2003). If an outlier is present for any attribute that can be viewed with a dot mark in a box plot. Figures 4 and 5 show both datasets of compressive and split-tensile having no outliers.

Fig. 4
figure 4

Box plot of the compressive model dataset

Fig. 5
figure 5

Box plot of the split-tensile model dataset

A density plot gives a fair assessment of the outlook of the data and if it is normally distributed or not so that the applied algorithm will not require any assumption about the organization of data which may cause consumption of time while running the model (Chen 2017). KDE is a probability density function that assists in the evaluation of the density plot (Kim and Park 2014). A steady curve was drawn at every data point and all the curves were summated to make a single smooth curve (Terrell and Scott 1992). The distribution of the data in the density plot should be like a Gaussian bell-shaped. If the data contains irregularity, then shape and distribution will vary. Figure 6 shows the perfect Gaussian distribution of all the input attributes of the dataset. The plot having the region with a higher peak resembles the density of maximum data points. Figure 7 shows the correlation matrix of all the features in the dataset. The Pearson correlation matrix indicates the relationship between all the attributes is linear or non-linear. As it can be visualized from the figure the strength of SC-AASC shows a non-linear to moderately linear relationship with other parameters.

Fig. 6
figure 6

Density-based distribution plot of the input dataset

Fig. 7
figure 7

Co-relation matrix

4 Machine learning models and model evaluation parameters

Most of the studies as documented in Sect. 1 employ ANN, GEP, Tree-based algorithms, boosting, and bagging regressors for strength prediction. In this study, for a comprehensive comparison between different types of AI-based models, six ML algorithms were employed. The hyperparameters of these models were optimized using a metaheuristic optimization algorithm. The ML algorithms include modified linear regression (LASSO), support vector regressor, bagging, and boosting regressors. These are very efficient ML algorithms that were used for strength prediction in this study.

4.1 LASSO regressor

The Least Absolute Shrinkage and Selection Operator (LASSO) is a modified linear regression method used as an ML algorithm for modeling purposes. It uses the shrinkage method to penalize the model. The values are shrunk towards a central point as the mean and the model is penalized for the sum of absolute values of the weights (Kang et al. 2021). LASSO uses regularisation and selection of variables or features of the dataset for enhancement of model accuracy i.e., introducing penalty factor to constraint the co-efficient and when the co-efficient is shrunk towards the mean of zero less important features are eliminated before the model proceeds to train (Mangalathu and Jeon 2018). The mathematical equation of the LASSO regressor is given below.

$$ L\left( \beta \right) = \mathop \sum \limits_{i = 1}^{n} \left( {y_{i} - \mathop \sum \limits_{j} x_{ij} \beta_{j} } \right)^{2} + \lambda \mathop \sum \limits_{j = 1}^{p} \left| {\beta_{j} } \right|. $$
(1)

Penalization factor i.e., the amount of shrinkage denoted by λ. An increase of λ has a direct proportion effect on the bias but is indirectly proportioned toward variance. The LASSO regression method reduces the multi-collinearity of the dataset and simplifies the model. It also prevents overfitting.

4.2 Support vector regressor (SVR)

A support vector machine (SVM) is one of the popular ML algorithms used for classification and regression problems proposed by Cortes and Vapnik (1995). The fundamental thought behind SVM is that it takes a set of input data and using a non-linear mapping function transforms the data into high-dimensional feature space (Boser et al. 1992). SVM Regression referred to as SVR designated using the sparse solution, kernels, decision boundaries, and several support vectors; can predict the value for a non-linear database. SVR gives mobility to interpret the amount of acceptance of error in the model and find a suitable line i.e., a hyperplane in a higher dimension to fit the data. SVR trains utilize a symmetrical loss function which similar to SVM penalizes high and low misestimates (Awad and Khanna 2015). SVR’s regression equation in the feature space can be approximated as

$$ Z\left( {a,w} \right) = \left( {w.\phi \left( a \right) + c} \right). $$
(2)

The weight vector defined by w and ϕ(a) is the feature function. By the incorporation of Lagrangian multiplier β and β*, using kernel function K(xi, xj), and applying Karush–Kuhn–Tucker conditions (Parveen et al. 2016); the SVR function can be obtained as:

$$ Z\left( {a,\beta_{i} ,\beta_{i}^{*} } \right) = \mathop \sum \limits_{i = 1}^{{N_{sv} }} \left( {\beta_{i} - \beta_{i}^{*} } \right)K\left( {a,a_{i} } \right) + c. $$
(3)

4.3 Random Forest Regressor (RFR)

The Random Forest (RF) algorithm stands out as a versatile and user-friendly machine-learning technique. It excels in accuracy, often eliminating the need for intricate hyperparameter tuning. RF operates by constructing an ensemble of decision trees (DTs) referred to as a "Forest." These trees are generated from diverse dataset samples and trained using the bagging technique. The final output is obtained by amalgamating the outputs of these trees through averaging, yielding a dependable and precise prediction (Dou et al. 2019). Each tree within the RF is constructed independently, utilizing distinct subsets of attributes. This approach leads to a reduction in the dimensionality of the feature space. While developing these decision trees, RF doesn't prioritize the search for critical factors in node splitting; instead, it seeks the optimal feature among a subset of features. Consequently, it generates smaller DTs, thereby mitigating the risk of overfitting (Géron 2019).

$$ \hat{y} = \frac{1}{B}\mathop \sum \limits_{d = 1}^{B} f_{d} \left( x \right). $$
(4)

4.4 Bagging regressor (BR)

A bagging regressor or bootstrap aggregating regressor is a bagging ensemble ML meta-estimator used for modeling that fits or resamples the original training dataset with replacement (Sutton 2005). Some data may be occurring multiple times while others left out in the resampled dataset. It is based on a powerful statistical method known as a bootstrap for estimating quantity from a data sample. As previously discussed in Sect. 5.2 this method decreases the variance of the black-box estimator by initiating randomization and then creating an ensemble. If for a given training set B = {(x1, y1) …. (xn, yn)} the bagging regressor converts it to quasi-replica training set T by replacement and the final regression equation can be given as

$$ \overline{f}\left( x \right) = \mathop \sum \limits_{i = 1}^{T} f_{i} \left( x \right). $$
(5)

4.5 AdaBoost regressor (AB)

Adaptive Boosting, also referred to as AdaBoost, represents a meta-learning approach within the realm of machine learning (Wu et al. 2008). This pioneering boosting algorithm marked a significant advancement in enhancing the effectiveness of binary classification tasks, originally denoted as AdaBoost.M1. The fundamental principle behind AdaBoost involves its capacity to transform the errors made by weaker classifiers into strengths (Schapire 2013), thereby mitigating bias in supervised learning contexts. While contrasting with the Random Forest (RF) model, wherein Decision Trees (DTs) lack a predetermined depth and encompass an initial node connected to numerous leaf nodes, AdaBoost employs concise DTs featuring only one level or a solitary split. Such DTs with a solitary split are commonly referred to as 'Decision stumps' or simply 'stumps'. The algorithm's workflow entails constructing a model that initially attributes equal weights to all data points within a dataset. Subsequently, these weights are reassigned to favor weaker learners in each iteration, thus granting them higher importance in the subsequent rounds. This iterative training process persists until errors are minimized, leading to heightened accuracy.

$$ H\left( x \right) = {\text{sign}}\left( {\mathop \sum \limits_{t = 1}^{T} \alpha_{t} h_{t} \left( x \right)} \right). $$
(6)

4.6 XGBoost regressor (XGB)

XGBoost, short for Extreme Gradient Boosting, represents the pinnacle of ensemble machine learning algorithms, boasting both remarkable speed and formidable power. This algorithm provides a structured approach to harnessing the combined predictive prowess of multiple individual learners. Its unparalleled capacity for cache optimization renders it the swiftest performer in program execution. At its core, XGBoost embodies a distributed gradient-boosted decision tree (GBDT) framework for machine learning. Widely recognized as the vanguard of meta-learning algorithms, it commands the field of regression, classification, and ranking quandaries (Dong et al. 2020). Drawing upon the foundations of supervised machine learning, XGBoost leverages the principles of the classification and regression tree (CART), seamlessly integrating gradient boosting, stochastic gradient boosting, and regularized gradient boosting algorithms. These techniques orchestrate the execution of decision trees at each data point sequentially, while simultaneously training them in parallel. This strategy imparts more weight to weaker learners, which are subsequently fed into other decision trees. The cumulative output of these trees harmonizes to yield the ultimate result (Duan et al. 2021). XGBooster introduces a pivotal augmentation by integrating the predicted residuals and errors from preceding models via the gradient descent algorithm, thereby minimizing loss. The mathematical formulation encapsulating the essence of XGBoost is as follows:

$$ \hat{y} = \mathop \sum \limits_{k = 1}^{k} f_{k} \left( {x_{i} } \right),f_{k} \varepsilon F. $$
(7)

4.7 Spotted Hyena optimization (SHO)

Dhiman and Kumar (2017) introduced a novel bio-inspired metaheuristic optimization technique known as the Spotted Hyena Optimizer (SHO). This approach draws its inspiration from the social behaviors observed in spotted hyenas, particularly in their hunting activities. The SHO algorithm is designed to replicate these natural interactions and utilize them for optimization tasks. To validate its effectiveness, the SHO algorithm was put to the test using practical engineering design problems that involve more than four variables (Dhiman and Kumar 2019). The results obtained from these experiments showcase the superior performance of SHO compared to other competing algorithms, particularly when dealing with real-world problems. The SHO algorithm revolves around four core components: searching, encircling, hunting, and attacking. These components correspond to the steps spotted hyenas take during their hunting process. The primary objective is to emulate the behavior of the hyenas when they work together to catch their prey. In the algorithm, the targeted prey is akin to the ideal route for encircling the prey, and other search agents or hyenas adjust their positions based on this optimal solution. This optimal solution serves as a guide for the rest of the group, enabling efficient movement towards achieving the optimization goals. The initial focus of the SHO algorithm lies in executing the hunting strategy. This strategy involves generating an ideal response cluster for the most proficient search agent and adjusting the positions of the other search agents. This process is elucidated through a set of subsequent equations. Similar to a pack of spotted hyenas compelled to charge towards their prey, the algorithm orchestrates this movement. The efficacy of an algorithm's exploration capability is intricately linked to its search mechanism. This capability is harnessed by utilizing variables denoted as "E" and random numbers falling either above or below 1. Within the scope of this study, our efforts were directed at optimizing the hyperparameters inherent to all the proposed machine learning methodologies. These optimizations were geared towards enhancing the predictive accuracy for both compressive strength and split-tensile strength.

4.8 Model evaluation parameters

Cross-validation (CV) stands as a crucial technique to enhance the reliability of model outcomes (Xiong et al. 2020). The core concept behind CV involves partitioning the training set into two distinct subsets: one serves as the training dataset, facilitating model training, while the other functions as the validation or testing dataset, employed to assess the model's performance. The model's performance on the validation dataset acts as an estimate of its generalization error, becoming a pivotal metric in the process of model selection. In this specific study, a preference is given to k-fold cross-validation. This methodology involves splitting the dataset into k non-overlapping subsets of equal size. In each iteration, one of these subsets is utilized for validation, while the remaining k − 1 subsets are combined for training. This sequence is reiterated k times, each time utilizing a different subset for validation, allowing for extensive utilization of the dataset for training. The choice of k typically resides in the range of 3–10, with this study adopting a k value of 10. For the creation and assessment of prediction models utilizing machine learning algorithms, the k-fold cross-validation technique is applied in conjunction with the curated SC-AASC database. The evaluation of model accuracy is executed through efficiency metrics like the coefficient of determination (R2) and adjusted R2. Moreover, error metrics such as Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), Mean Squared Error (MSE), and Root Mean Squared Error (RMSE) are employed. R2 gauges the variance proportion, adjusted R2 offers insights into reliability and correlation, MAE computes the average deviation, MAPE addresses size-dependent errors not captured by MAE, while MSE and RMSE function as risk measures that quantify the deviation between data points and the fitted model. Additionally, two other methods namely SI (Scatter index) and StI (Structural index) were also used to access the machine learning models. An excellent model would typically have a StI value close to 0 or a very low StI value, indicating that the model's structural complexity is well-suited to the data and that it generalizes well. When considering the SI parameter, it's evident that a model's performance is classified as poor when SI exceeds 0.3. In cases where the SI falls within the range of 0.2–0.3, the performance is considered fair. Furthermore, a model can be deemed to exhibit good performance when its SI ranges from 0.1 to 0.2. Lastly, an exceptional level of performance is achieved when the SI is less than 0.1.

$$ R^{2} = 1 - \frac{{\mathop \sum \nolimits_{i} \left( {y_{i} - \hat{y}_{i} } \right)^{2} }}{{\mathop \sum \nolimits_{i} \left( {y_{i} - y^{\prime}} \right)^{2} }} $$
(8)
$$ {\text{Adjusted}}\;R^{2} = 1 - \frac{{\left( {1 - R^{2} } \right)\left( {n - 1} \right)}}{n - p - 1} $$
(9)
$$ {\text{MAE}} = \frac{1}{n}\mathop \sum \limits_{i = 1}^{n} \left| {y - y^{\prime}} \right| $$
(10)
$$ {\text{MAPE}} = \frac{1}{n}\mathop \sum \limits_{i = 1}^{n} \left| {\frac{{y^{\prime} - y}}{y}} \right| \times 100 $$
(11)
$$ {\text{RMSE}} = \sqrt {\frac{1}{n}\mathop \sum \limits_{i = 1}^{n} \left( {y^{\prime} - y} \right)^{2} } $$
(12)
$$ {\text{SI}} = \frac{{{\text{RMSE}}}}{{t^{\prime}}} $$
(13)
$$ {\text{StI}} = 1 - R^{2} . $$
(14)

5 Sensitivity analysis

Sensitivity assessments include changing a system's inputs to determine exactly how each input parameter affects the output. In this study, two global sensitivity methods were utilized. The method of allocating the uncertainty in outputs to the uncertainty in each input element over their whole range of interest is known as global sensitivity analysis. When all the input factors are changed at once and the sensitivity is assessed across the whole range of each input factor, the analysis is said to be global. Sobol and FAST global sensitivity analysis were used in this study.

5.1 Sobol sensitivity analysis

The goal of a Sobol sensitivity analysis is to identify the proportion of each input parameter's variability in the model output that is dependent on that parameter alone or on its interaction with other factors (Saltelli et al. 1999). The classical analysis of variance in a factorial design is used as the basis for the dissection of the output variance in a Sobol sensitivity analysis. It should be understood that the goal of a Sobol sensitivity study is not to determine what causes input variability. It merely indicates the effect and magnitude it will have on the model's output. This method is thought to decompose any given integrable function f(x) into the sum of 2n orthogonal terms.

$$ f\left( x \right) = f_{0} + \mathop \sum \limits_{s = 1}^{n} \mathop \sum \limits_{{i_{1} < \cdots < i_{s} }}^{n} f_{{i_{1} \ldots i_{s} }} \left( {x_{{i_{1} }} , \ldots ,x_{{i_{s} }} } \right). $$
(15)

5.2 Fourier amplitude sensitivity testing (FAST)

One of the most well-liked tools for uncertainty and sensitivity analysis is the Fourier Amplitude Sensitivity Test (FAST). It breaks down a model output's variance into partial variances supplied by various model parameters using a periodic sampling strategy and a Fourier transformation. Cukier et al. (1973) developed FAST in 1973. The FAST technique, which provides a measure of the impact of each of the input qualities on the model's output, was used in this work to numerically illustrate each feature's contribution. The effect on strength increases as the parameter's relevance value rises.

6 Result and discussion

6.1 Hyperparameter optimization

After a given number of iterations, the objective function's inputs are examined to locate the global optimum point using the metaheuristic SHO method. When there is a lot of domain knowledge that could sway or bias the optimization process, spotted hyena optimization enables the identification of novel solutions. The appropriate hyperparameter values for each machine-learning method that makes use of the SHO algorithm are listed in Table 7.

Table 7 Optimal hyperparameters for machine learning algorithms determined by SHO

6.2 LASSO regressor

The prediction of strength properties through this model was done appropriately with training and testing of data. The optimum hyperparameter value of the lasso regressor is depicted in Table 7. The training regression plot for compressive data is shown in Fig. 8 and that for split-tensile strength is shown in Fig. 10. Likewise, Figs. 9 and 11 represent the regression plots in testing for both datasets. The testing accuracy of the compressive model is achieved at a somewhat 63% confidence level with a mean absolute error of 6.77 KN and that of the split-tensile model is at an 83% confidence level with and mean absolute error of 0.22 KN. As SC-AASC possesses a non-linear relationship between constituent material and strength it doesn’t show a great level of accuracy in this modified linear regression model. Additional details of all the statistical parameters are listed in Tables 9 and 10.

Fig. 8
figure 8

Regression analysis in the training phase of CS models: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

Fig. 9
figure 9

Regression analysis in the testing phase of CS models: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

Fig. 10
figure 10

Regression analysis in the training phase of STS models: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

Fig. 11
figure 11

Regression analysis in the testing phase of CS models: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

6.3 Support vector regressor (SVR)

Support vector regression is a supervised learning algorithm that is used to predict discrete values. Support Vector Regression uses the same principle as the SVMs. The basic idea behind SVR is to find the best-fit line. Figures 8 and 10 demonstrate the training regression plot for compressive and split-tensile strength data, respectively. Similarly, Figs. 9 and 11 show the regression plots tested on both datasets. The testing accuracy of the compressive strength model is at 85%, and that for split-tensile is at 87% confidence level. In SVR the use of a non-linear kernel resulted in better performance than the linear regression model. More information about each statistical parameter is listed in Tables 9 and 10.

6.4 Random forest (RF)

This is an advanced version of the bagging ensemble regressor and creates DTs at every data point which decreases variance and when ensembled gives better accuracy. For this reason, the model predicts compressive strength at an 88% confidence level and split-tensile strength at a 91% confidence level in the testing phase. The ideal RFR hyperparameter value is shown in Table 6. The fitting curve shown in Figs. 8, 9, 10, and 11. Tables 9 and 10 represent details of all the additional model performance parameters.

6.5 Bagging regressor (BR)

This bagging ensemble model predicts compressive strength at an 80% confidence level and predicts split-tensile strength at a 91.5% confidence level in the testing phase. The hyperparameters optimized by SHO are listed in Table 6. The fitting curve is shown in Figs. 8, 9, 10, and 11 which includes the training and testing phase for both datasets. It shows greater performance than prior models as it decreases the variance of the data. Details of all the additional model performance parameters are shown in Tables 9 and 10.

6.6 AdaBooster (AB)

This boosting ensemble model uses fix-shaped decision stumps at every data point which reduces bias and simplifies the model ultimately increasing accuracy. This model forecasts compressive strength at an 84.5% confidence level and split-tensile strength at a 90.4% confidence level in the testing phase achieving a generalized MAE value. The fitting curve, which includes the training and testing phases for both datasets, is depicted in Figs. 8, 9, 10, and 11.

6.7 XGBooster (XGB)

Spotted hyena-optimized hyperparameters of XGB are shown in Table 6. XGB is the most advanced boosting ensemble model and using CART and gradient boosting methods increases the predictive power of the model. The model estimates compressive strength at a 97% confidence level and split-tensile strength at a 96% confidence level; performing with less MAE, MAPE, and RMSE. For both datasets, the training and testing phases of the fitting curve are depicted in Figs. 8, 9, 10, and 11. It uses self-compatible regulatory functions i.e., under-sampling and column shrinkage that makes it more reliable and outperforms other ML models. Tables 8 and 9 include all the statistical indices of the XGB model.

Table 8 Training and testing scores of all the CS models
Table 9 Training and testing scores of all the STS models

6.8 Comparison between the machine learning models

This section presents a thorough evaluation of the model's robustness and generalizability by contrasting each model with the others. A comparison of the models was made by plotting the errors of each data point, analyzing the maximum errors, and performance metrics scores for each model. A 3D plot is used to visualize the errors of each data point in correspondence to the experiment and predicted values. Figure 12 represents the 3D error plot for the compressive dataset while Fig. 13 represents the split-tensile dataset. From the plot, it can be seen that XGB shows the least error per data point indicating superior prediction performance. Ensemble ML models also perform better in comparison to single ML models. The maximum error for each model is listed in Table 10. Lasso performs least accurately with the highest maximum error while XGB outperforms all. Likewise, for both datasets among all the model's Lasso regressor model shows the least R2 and adjusted R2 value while the XGBoost model shows the highest R2 and adjusted R2 score. Ensemble ML models show good generalizability in comparison to single ML models and compared to bagging ensemble models boosting ensemble models showed good reliability.

Fig. 12
figure 12

3D plot for error in prediction for CS models: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

Fig. 13
figure 13

3D plot for error in prediction for STS models: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

Table 10 Maximum error of each model

Accuracy and error statistics of all the models in training and testing are shown in Tables 8 and 9. Figures 14 and 15 represent the line plot for actual and predicted values in the testing phase. For both CS and STS models, the Lasso model with a higher value of error metrics indicates lower generalization. SVR shows the least error among the single ML models as it shows effectiveness in modeling. Overall boosting ensemble models show fewer errors in comparison to the bagging model as ensemble ML models perform at higher accuracy than single ML models as it prioritizes weak learners but as previously discussed RF model shows the best error statistically than AdaBooster for the split-tensile dataset. XGB model which can be labeled as the best ML model shows excellent generalizability and robust predictive capability.

Fig. 14
figure 14

Line plot between actual and predicted CS models in testing: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

Fig. 15
figure 15

Line plot between actual and predicted STS models in testing: a Lasso; b SVR; c RFR; d BR; e AB; and f XGB

Figure 16 represents a Taylor diagram of all the models tested on both the CS and STS datasets. The radial distance in the Taylor diagram represents the standard deviation and the angular axis represents the co-relation. The performance of all the models was visualized with a reference. It can be seen that the SHO-XGB performs well compared to all the models for both datasets.

Fig. 16
figure 16

Taylor diagram of all the models: a CS dataset, b STS dataset

6.9 Comparison among previously published results

The spotted hyena-optimized XGBoost was found to be performing exceptionally well in both the CS and STS datasets. In this section, a comparison is drawn between previously published research articles that have worked on self-compacting geopolymer concrete and alkali-activated slag concrete. Awoyera et al. (2020) used ANN and GEP for the prediction of compressive strength of SCGC. ANN model was found to be more accurate. Shahmansouri et al. (2020) used GEP for CS prediction of GGBS-based alkali-activated slag concrete. Faridmehr et al. (2021) used Bat-optimized ANN for SCGC CS estimation, and Basilio and Goliatt (2022) utilized gradient-boosting hybridization for CS prediction. Figure 17 represents a comparison of SHO-XGB with all the models. It can be ascertained that the developed model gives satisfactory results.

Fig. 17
figure 17

Comparison of SHO-XGB with published models

6.10 Sensitivity analysis

Two global sensitivity approaches were used to evaluate the relative influence of each input parameter on output. Figure 18 represents the total order of Sobol indices of all the input parameters. The total sampling size of the Sobol sensitivity method was determined to be 77,784. From Fig. 18, it can be seen that AAS/binder ratio, AAS content, binder content, and curing days greatly affect the compressive strength of alkali-activated slag concrete. Similar observations have been made earlier using slag as binder material (Kumar Dash et al. 2023; Saini and Vattipalli 2020; Shahmansouri et al. 2020). So, the alignment of the current study with earlier published results establishes the importance of these factors in controlling the strength development of alkali-activated concrete. The AAS content and AAS-binder ratio which resembles the water content and water-cement ratio are the two most sensitive factors governing strength development (Fig. 18). So, we can adopt the approach of controlling these two factors to have the desired strength properties of alkali-activated slag concrete. The FAST sensitivity analysis (Fig. 19) was used to overcome the cost-effectiveness and time factor associated with the Sobol method. The FAST method takes less time while giving almost equal results. The sampling size in the FAST method was 77,848. AAS/binder ratio, binder content, and AAS content are also presented with maximum index. The fresh property results which were taken as input scored minimum sensitivity value as they possess’ a linear relationship with strength.

Fig. 18
figure 18

Sobol total order index of input parameters

Fig. 19
figure 19

FAST First order index of input parameters

Figures 20 and 21 represent the effect of various parameters on the compressive strength and split-tensile strength of SC-AASC. It's noted that as the content of GGBFS increases, the compressive and split-tensile strengths of SC-AASC also increase. This suggests that GGBFS is contributing positively to the strength properties of the concrete mix. GGBFS is known for its pozzolanic properties, which can enhance the strength and durability of concrete. Conversely, reducing the content of AAS also results in higher strength. AAS is used as an activator in alkali-activated concretes, and a lower content implies that the mixture might be more optimized in terms of activator dosage. It can also be observed that an increase in GGBFS content and a decrease in AAS content led to SC-AASC with higher strength. GGBFS content in the range of 500–540 kg/m3 and AAS in the range of 190–200 kg/m3 resulted in higher strength. A lower AAS/binder ratio resulted in higher strength while a higher content of coarse aggregate also increased the strength. This implies that reducing the amount of activator relative to the total binder content is beneficial for higher strength development, however, the increase in coarse aggregate has a negative effect on the flow property of SC-AASC. Lower fine aggregate content seems to have a positive impact on both compressive and split-tensile strength while 7% of superplasticizer dosage increased the strength of SC-AASC. 21% of extra water dosage and longer curing period observed to be giving maximum strength.

Fig. 20
figure 20

Effect of various parameters on compressive strength

Fig. 21
figure 21

Effect of various parameters on split-tensile strength

7 Practical application of the research

The research on predicting mechanical strength properties in high-performance SC-AASC using metaheuristic optimized machine learning models has multiple practical applications that can offer advantages to various stakeholders in the construction industry. This research has explored the development of High-Performance SC-AASC formulated with GGBFS can be utilized as an eco-friendly alternative to traditional concrete in various construction projects. Its reduced carbon footprint and efficient use of industrial byproducts make it an ideal choice for environmentally conscious builders. Its self-compacting properties can simplify the construction of complex structures, such as bridges, tunnels, and high-rise buildings, where ensuring proper compaction is traditionally challenging. This can lead to faster infrastructure development. The optimization of High-Performance SC-AASC mixtures enhances their durability and longevity, reducing the need for frequent maintenance and repair of concrete structures. This can result in significant long-term cost savings for infrastructure owners and operators. The major drawback is the cost associated with the production of this type of concrete as there are no specific codes and regulations available to optimize its performance. A prediction model can help decrease the cost of the concrete by optimizing various aspects of the manufacturing process like mix design optimization, process efficiency, quality control, and materials selection. The application of metaheuristic-optimized machine learning models for predicting its strength properties offers a powerful tool for engineers and construction professionals. They can use these models to assess material performance quickly and accurately during the design phase, leading to better-informed decisions. The use of predictive modeling reduces the reliance on extensive and time-consuming laboratory testing, speeding up material characterization. This not only accelerates the research and development process but also lowers testing costs, benefiting both researchers and construction projects for the utilization of SCGC and SC-AASC. The utilization of GGBFS as a primary binder encourages collaboration between the construction and steel industries, promoting the repurposing of industrial byproducts and fostering sustainable practices.

8 Conclusions

The high-performance SC-AASC represents an eco-conscious alternative to conventional cement-based composites. By utilizing industrial waste materials and emitting lower quantities of carbon dioxide, AASC emerges as a sustainable construction material, effectively minimizing its environmental influence. Predicting the mechanical strength in SC-AASC presents a notable hurdle, necessitating the use of advanced techniques for precise estimation. This study tackles this challenge through the creation of innovative machine-learning models optimized with metaheuristic methods. The subsequent section presents the concluding insights derived from this research endeavor.

  • The study proposes a method of production of high-strength SC-AASC from a base M25 mix grade to an M45 and M60 mix grade.

  • The AAS content varied from 220 to 190 kg/m3 and the AAS/binder varied from 0.47 to 0.36 to produce high-performance SC-AASC. The percentage of SP varied between 6–7% while the percentage of extra water kept at 21% and 24%.

  • AAS content of 200 kg/m3 and AAS/binder of 0.39 yielded greater strength for M45 graded SC-AASC while M60 grade SC-AASC showed higher strength with AAS content of 190 kg/m3 and AAS/binder of 0.36.

  • SP of 6% and extra water content of 24% facilitated good workability of M45 and M60 grade SC-AASC.

  • A database containing 135 datapoints was developed for the creation of machine learning models by taking 12 input parameters and compressive strength and split-tensile strength as output.

  • Data analysis resulted in the database having no outlier but having non-linear co-relation.

  • Six machine learning models were developed consisting of two single supervised algorithms and four ensembled supervised algorithms.

  • The hyperparameters of all the machine learning models were optimized using the metaheuristic bio-based spotted hyena optimization method.

  • The ensembled ML algorithm (Bagging and Boosting) showed better accuracy and tolerable error in comparison to the single ML algorithm (Lasso and SVR). Error metrics showed boosting ensemble models (AdaBooster and XGB) showed better accuracy than bagging ensemble models (BR and RF).

  • Non-linear single ML models (SVR) showed better performance than modified linear single ML models (Lasso; which can be termed as the worst-performing model). The optimized XGB was found to be best best-performing model.

  • Sobol and FAST sensitivity methods were used for the assessment of relative influence. The AAS/binder, AAS content, binder content, and curing days were found to be the primary influencing factors concerning the strength properties of high-strength SC-AASC.

9 Limitations and future scope of the study

9.1 Limitations

The following are the limitations of this study.

  • The research primarily focuses on specific mix grades (M25 to M60) and may lack broader applicability to other concrete grades or different environmental conditions.

  • While the study emphasizes reduced environmental impact, the broader environmental implications can be derived using life cycle assessments or environmental footprint analysis.

  • Although the study explores variations in critical parameters like AAS content, AAS/binder ratio, and curing days, the scope might not cover all potential combinations or extreme ranges of these parameters. This limitation might restrict the model's predictive capacity in scenarios not fully represented in the dataset.

  • The size of the dataset needs to be enhanced with more parametric variation and experimental testing for better ML performance.

  • The sensitivity analysis, while valuable, might not encompass the entire spectrum of potential influencing factors. Factors not included in the analysis might have an impact on the material properties, affecting the comprehensiveness of the conclusions drawn.

9.2 Future scope

The future scope of the study includes

  • Researching novel materials or hybrid systems that combine various sustainable elements, such as recycled aggregates, nano-materials, or bio-based additives, could lead to the development of even more advanced and sustainable construction materials.

  • Conducting a comprehensive life cycle assessment (LCA) or environmental footprint analysis to evaluate the holistic environmental impact of SC-AASC, including its production, use, and disposal stages, would provide valuable insights into its sustainability.

  • Expanding the study to encompass a wider range of parameters could provide a more comprehensive understanding of their impact on the properties of high-strength SC-AASC.

  • Conducting extensive external validation by applying the developed models to real construction projects or experimental testing would validate the models’ performance in practical scenarios, ensuring their reliability and applicability.

  • Investigating optimization strategies beyond metaheuristic methods, such as evolutionary algorithms or reinforcement learning, could enhance the precision of parameter optimization for high-strength SC-AASC production.

  • Collaborating with industry stakeholders and standardization bodies to develop guidelines or standards for the production and application of high-strength SC-AASC would facilitate its widespread adoption in the construction sector.