Introduction

As documented by various standard codes, a typical RCC structure design must be checked for satisfying strength and durability or serviceability criteria. Strength criteria must satisfy flexure, compression, shear, and torsion requirements, whereas serviceability criteria must satisfy deflection and cracking [1,2,3,4]. Each of these requirements is the function of the compressive strength of the concrete directly or indirectly. This makes it of utmost importance to ensure that the concrete used for construction achieves the desired compressive strength used while designing the structure. For this purpose, testing and estimating the compressive strength of in situ concrete is imperative.

The sampling frequency is defined in several codes, and a general value can be considered at least 1 sample for each 1–5 cubic meters of concreting work. Every batch of concrete that has been prepared is taken for testing [1]. Each concrete sample must be of a minimum of 0.02 cubic meters in volume and be taken from different points from the concreting work [5]; thus, the compressive test must be carried out multiple times throughout the project. The compressive test must be carried out as outlined in several codes [4, 6, 7]. In addition, the compressive strength of concrete can also be determined by non-destructive tests [8]. Although non-destructive tests do not require sampling and are quick to perform, these methods have higher dispersion and deviation from the true compressive strength [9]. Thus, the conventional techniques of compressive testing require expensive equipment, skilled labour, and a dedicated testing facility. These requirements are challenging to follow in a low-cost construction due to the non-availability of funds, critical and short project spans, and the non-availability of a dedicated testing area.

The compressive strength of concrete is not a stochastic property, but it is dependent on various factors like water-cement ratio [10], cement strength class [11], aggregates and admixtures [12], shape and size [9], to name a few factors, apart from these the conditions under which the concrete is cast like the ambient environment temperature [14], site quality control and various other factors. This makes it a function of some pre-defined parameters between which a relationship can be established. Based on the past literature, it has been seen that soft computing techniques are well suited to finding relationships between parameters to get the desired output in multiple fields [15,16,17,18,19]. It is also seen that conventional modelling techniques of concrete fall short for complex concrete mixes. Soft computing techniques have been proven to predict the compressive strength of a simple concrete mix [20] and a complex of steel fibre-reinforced concrete [21]. The conventional modelling techniques were difficult to model the compressive strength of the complex mix model due to a large number of forecasting variables; soft computing techniques proved useful.

This paper highlights the research in this domain using AI and ML, predominantly bifurcating it into four methods. The first is the conglomerate of algorithms based on rules implemented, the second is the classical ML methods, including the regression and classification techniques, the third is the shallow neural networks, and the fourth is deep machine learning. The approaches reviewed in this study range from simple parametric methods to various AI and ML techniques like shallow neural network models, deep neural networks, and also computer vision-based methods in which the image of the surface of concrete [22] or microtomographic images of the concrete [23]. The computer vision-based methods employ convolutional neural network models to identify patterns in the images and pass them through a deep neural network or classical ML techniques to deduce a satisfactory conclusion.

Qualities a successful prediction should have are scalability of the model, i.e., the results should apply to a large-scale concreting as well; secondly, it should be repeatable; and thirdly, it should conform to evaluation metrics. The performance of soft computing techniques is checked by various evaluation metrics like root mean square error (RMSE), mean absolute error, R2 value, and mean absolute percentage error (MAPE). The same shall be discussed in brief in the upcoming section.

This paper aims to critically review different soft computing techniques used to predict the compressive strength of concrete. Figure 1 shows the percentage of soft computing techniques utilized by various literature reviewed in this study. The rest of the paper is organized as follows. In section II, the working principles of some of the popular AI algorithms which are encountered in various literature are described. Section III describes the popular evaluation metric used. Section IV describes methodologies used by researchers and their performing capabilities to predict the compressive strength of concrete. Section V describes the conclusion and the future research scope.

Fig. 1
figure 1

Pie chart of different algorithms in this study, all regression models are grouped as ‘REGRESSION’, all tree models (DT, RF, Boosting Trees and other similar algorithms) are grouped as ‘TREE MODELS’, neural network models except DNN and CNN are grouped as ‘NN’, Bio-inspired algorithms, GEP, GP are grouped as ‘BIO INSPIRED’

Artificial intelligence algorithms

Algorithms based on rules

Adaptive neuro-fuzzy inference system

Fuzzy networks can be optimized using backpropagation and other training functions implemented in a neural network system. This system is called a neuro-fuzzy system. One of the methods for neuro-fuzzy inference is the Mamdani inference system, as shown in Fig. 2. In the Mamdani inference system, a crisp input is applied to a fuzzy set, resulting in a fuzzy output set, which results in a crisp output by de-fuzzification. Mamdani fuzzy inference system has five layers, namely:

  1. i.

    Fuzzification layer: This layer is used for calculating the membership functions. There are different types of membership functions like trapezoidal function, bell function, triangular function, Gaussian function, and left–right function, to list a few. For most cases, the Gaussian membership function is used.

  2. ii.

    Inference layer: The inference layer is the reasoning mechanism of the neuro-fuzzy system. A fuzzy rule can be defined as a relation between its antecedent (input) and the consequence (output). Based on this, fuzzy reasoning can be defined as a single rule with a single antecedent, a single rule with multiple antecedents, or multiple rules with numerous antecedents. Weights in the inference system are found using the T-norm operator.

  3. iii.

    Implication layer: The consequent membership functions are calculated based on the inference layer.

  4. iv.

    Aggregation layer: The implied results are summed up to give a fuzzy output.

  5. v.

    Defuzzification layer: Defuzzification is the process of converting fuzzy output to a crisp output. In Mamdani Inference System, the centroid of the fuzzy output is calculated, and that value is the crisp output.

Fig. 2
figure 2

Mamdani Fuzzy Inference System

Fuzzy logic

A Boolean logic decision-making system takes only two values of true and false or 1 and 0 [24]. These values can be put through several logic gates to get the desired results. However, this logic system does not encompass any condition between true and false, for example, partially true, partially false, or true to some extent, and various other linguistic syntaxes like negations, hedges, and connectives. These scenarios are well captured by fuzzy logic. The decision making by fuzzy logic includes a fuzzification module in which membership values are assigned to specific criteria or conditions. Membership values are like weights that describe to what extent a particular condition is valid. A relationship is formed between membership values using IF-THEN rules, linguistic rules, or conditional rules, which give fuzzy logic an outlook from the perspective of humans. An inference system provides a de-fuzzified output from the relation established to get a crisp value.

Classical ML and clustering

Support vector machine (SVM)

Support vector machine is a classification algorithm used for applications like the classification of images, pattern recognition, and text categorization. The data is classified by constructing a hyperplane of (n-1) dimensions, where ‘n’ is an n-dimensional vector used to describe a data point.

The best hyperplane is the one that ensures maximum margin distance from the data points involved in classification. Figure 3a shows a hyperplane and margin distance in SVM for classification purposes. The data points closest to the hyperplane are called support vectors in this process. The kernel or classifier selection is essential for precise classification using SVM. Different types of kernels in SVM are the polynomial kernel, sigmoid kernel, and radial basis kernel. These three are the non-linear kernels, and the fourth is the linear kernel classifier. SVM has the capability of working with linear as well as non-linear data.

Fig. 3
figure 3

a Support Vector Machine (SVM) algorithm. b Support Vector Regression (SVR) algorithm. c K-nearest neighbour (KNN) algorithm

Support vector regression (SVR)

In simple regression problems, the final aim is to minimize the square of the error to ensure the best fit; however, in support vector regression, the main concern is with error reduction only to a certain degree. It is accepted if the predicted value, including the error, falls within the acceptable range. In SVR, the hyperplane is constructed as the best fit line, and the hyperplane is offset by an error tolerance ‘ε’ on both sides. The aim of SVR is that all data points should fall within this tolerance region. Introducing a slack variable minimizes their deviation for values outside the tolerance region. A slack variable ‘ξ’ denotes the deviation of a data point from the margins of the tolerance region. Figure 3b shows a schematic diagram for SVR. SVR and SVM work on similar principles of creating hyperplanes using the kernel; SVM is used for classification, whereas SVR gives a numerical answer.

Multi-linear regression

Multi-linear regression is a multivariate algorithm for supervised machine learning. It aims at establishing the relationship between multiple input variables and a dependent output variable. This algorithm is validated by the least square error between the actual and predicted values. A general form of multi-linear regression is given in equation (1) [25].

$$y={\beta }_{0}+\sum_{i=1}^{n}{\beta }_{i}{x}_{i}+\varepsilon$$
(1)

where ‘y’ is the dependent variable on ‘n’ different independent variables, β0 is the constant, βi is the regression vector, and ε is the error.

Non-linear regression

Non-linear regression is a multivariate higher degree algorithm for supervised machine learning. This algorithm establishes a relationship between the dependent variable and a polynomial multi-variable equation. Non-linear regression is used when the relationship between a dependent and independent variable is complex and not linear. A general form of non-linear regression is given in equation (2) [25].

$$y = \beta _{0} + \sum\limits_{\begin{subarray}{l} i = 1 \\ m \in R \end{subarray} }^{n} {\beta _{i} x_{n}^{m} + \varepsilon }$$
(2)

where ‘y’ is the dependent variable on ‘n’ independent variables with a degree ‘m’, m belongs to real number \({\mathbb{R}}\), β0 is the constant, and ε is the error.

Gaussian process regression

Gaussian process regression is based on the fact that infinite functions could fit the control data points for a particular distribution of data points. A non-parametric regression model calculates the probability distribution for all functions that fit the data points. The Gaussian process uses a kernel-based probabilistic model. The Gaussian process for an input ‘x’ and corresponding output ‘y’ is given by Eq. (3) [25].

$$y=g{\left(x\right)}^{T}\alpha +f\left(x\right)+\varepsilon f(x)\sim N(\mu ,K)$$
(3)

where g(x) is a vector function that relates the input variables to a multi-dimensional space, α is coefficient to g(x), f(x) constitutes for Gaussian process with µ being the mean vector matrix and K being the covariance matrix. The covariance matrix mathematically can be expressed as in equation (4) [25].

$$K=\left[\begin{array}{ccc}K\left({x}_{1}, {x}_{1}\right)& \dots & K\left({x}_{1},{x}_{n}\right)\\ \vdots & \ddots & \vdots \\ K\left({x}_{n},{x}_{1}\right)& \dots & K\left({x}_{n},{x}_{n}\right)\end{array}\right]=\left[\begin{array}{ccc}1& \dots & K\left({x}_{1},{x}_{n}\right)\\ \vdots & \ddots & \vdots \\ K\left({x}_{n},{x}_{1}\right)& \dots & 1\end{array}\right]$$
(4)

Extreme gradient boosting (XGBoost)

Extreme Gradient boosting is a decision tree based supervised learning technique. It can be used for classification and regression purposes. It employs a gradient boosting framework along with distributed processing or parallel processing. Gradient boosting alone suffers from problems like overfitting and more computation time; hence the extreme Gradient boosting method was evolved as a modified gradient boosting method. The training time of XGBoost is lesser than the Gradient boosting method. This method can model the non-linear interactions between various features used in machine learning.

K-nearest neighbour (KNN)

K nearest neighbour assumes that a data point has the same characteristics as that of maximum data points in its vicinity. It is a supervised machine learning technique used for classification and regression. A search space of potential solutions is searched locally within a radius such that ‘k’ nearest neighbours are selected. Figure 3c shows a KNN algorithm for k=6 and k=12, highlighting the significance of the ‘k’ value. The value of ‘k’ is thus an important parameter that needs to be determined first; it should not be too large or too small. Once the value of ‘k’ is decided, the data point is classified based on its proximity to its ‘k’ neighbours. This degree of proximity is quantified by finding the Euclidean distance and then sorting the distance from the most minor to the most significant order. Euclidean distance is given in equation (5) [25].

$$d=\sqrt{{\left({x}_{2}-{x}_{1}\right)}^{2}+{\left({y}_{2}-{y}_{1}\right)}^{2}}$$
(5)

The class label of the first ‘k’ neighbours is considered, and the label that occurs a maximum number of times is assigned as the class of the data point in consideration.

Decision tree algorithm

The decision tree is a supervised learning algorithm that shows existing patterns between predictors and a dependent variable using a combination of predictable rules. The classification technique generates decisions based on the information fed [26]. The schematic representation of rules and their outcomes looks like a tree diagram. The decision tree algorithm is well suited for classification as well as regression tasks. It typically has three types of nodes: root, decision, and leaf node. The root node is the base node and typically the starting point of the algorithms.

Figure 4a shows various decision tree components and bifurcation. The decision node is the node that determines the course path for subsequent outcomes. A leaf node is a terminal node having the outcome of a particular rule. Decision trees are sensitive to the data on which they are trained; hence if the training data is changed, the structure of the decision tree alters. An advantage of the decision tree algorithm is that it can also be trained for noisy data. This algorithm is easier than others but prone to overfitting. More complex tree-based algorithms like random forest and boosting trees are often preferred over decision tree algorithms.

Fig. 4
figure 4

a Decision Tree algorithm. b Random Forest algorithm

Random forest

A random forest is an ensemble of different decision tree algorithms to form a forest and reduce the disadvantage of individual trees. Random forests are an improvement over bagged decision tree.

The bagged decision tree algorithm creates multiple sub-trees for multiple subsets from the original data. A classification or regression task for a data point is done by taking the average value of all the individual decision trees. Figure 4b shows the classification process of the random forest algorithm.

Bagged decision tree models work better when the correlation between subtrees is less. This is not ensured in the bagged decision tree model; however, random forest ensures this by restricting the learning algorithm to a random feature instead of a learning algorithm searching through all data points in the bagged tree model.

Genetic programming

Genetic programming is a guided random search algorithm based on evolutionary techniques used to search solution space for potential solutions. These algorithms are inspired by nature and are generally used for optimization problems. Genetic programming can solve unconventional problems requiring great computational effort when followed by the traditional approach. A common framework of genetic programming starts by defining the genes which contain chromosomes. Chromosomes are a piece of information contained in a gene. Genetic programming aims to search for a better solution than the previous one. For this purpose, the genes are altered by crossover, mutation, and best gene selection. Genetic programming can also be coupled with other neural network systems to optimize specific parameters and yield better output. Leung et al. [27] tuned the structure of neural networks using an improved genetic algorithm.

Shallow neural network

Artificial neural network (ANN)

The artificial neural network is based on mimicking and learning the data with the process of actual neurons in the vertebrates. In actual neurons, the impulses flow from the dendrites to the cell body, which contains a nucleus. Figure 5a shows a typical artificial neuron. The input information from different features in ANN is similar to dendrites, which are processed in a unit containing an evaluation function and an activation function. The input neurons and output neurons are linked through weights and biases. The simplest form of ANN is a perceptron with two input neurons and one output neuron. Figure 5b shows a simple ANN architecture. Multiple layers of perceptron can be used to create a shallow neural network with a hidden node layer between input and output nodes. A typical epoch consists of updating weights, and the iterations are continued till the final value of weights are achieved such that a weighted sum of information at nodes gives the desired result.

Fig. 5
figure 5

a A typical artificial neuron. b ANN architecture with three input neurons in the input layer, two hidden neurons in the hidden layer and one output neuron

Deep machine learning

Convolutional neural network

The convolutional neural network is a deep neural network with high usage in computer vision for image recognition and pattern detection tasks. The convolutional neural network has three layers: convolutional, pooling, and fully connected. The convolutional and pooling layers can be in multitudes, but the end layer is always fully connected.

The convolutional layer consists of input data, a filter, and a feature map. For an image, the input can be the colour saturation value in each pixel, and the entire image is represented as a matrix of pixel information. The filter is a kernel, typically an array of weights that strides along with pixel information. These weights are multiplied with matrix value and added to form an output matrix. This output matrix is called a feature map or a convolved feature.

The pooling layer reduces dimensions by reducing the number of parameters. Pooling is primarily of two types max pooling and average pooling. In max-pooling, a filter is passed to the input through which the maximum feature value is extracted, and an output layer is formed. The average value of features is taken in average pooling, and the output layer is formed.

In the convolutional and pooling layers, a filter restricts the connectivity of all input features to output features; thus, these layers can be described as partially connected layers. The input nodes get translated through an activation function to an end output layer in the fully connected layer.

Deep belief network

A deep belief network suits an unsupervised and supervised machine learning task. A stack of Restricted Boltzmann Machine forms the deep belief network. Individual RBM performs non-linear operations to create an output that is utilized as input by the next RBM in line. RBMs are utilized for unsupervised learning; every layer of RBM is pre-trained to learn the features in an unsupervised manner. Further fine-tuning is done by the backpropagation technique, which utilizes gradient descent; this is used for classification and regression purposes.

Training algorithms

Backpropagation

Backpropagation is a popular supervised training algorithm technique for the neural network. A single iteration in the forward direction consists of taking a weighted sum and passing through an activation function to find a value at a hidden or output neuron. The backpropagation algorithm updates weights while traversing back from the output to the input in the same iteration. The weight updating equation between the output and the hidden neuron is given by equation (6) [19].

$$\Delta {V}_{jr}=\eta {O}_{j}{\delta }_{r} \;{\delta }_{r}={O}_{r}\left(1-{O}_{r}\right)*\left({t}_{r}-{O}_{r}\right)$$
(6)

where ‘j’ denotes the hidden node, ‘r’ corresponds to the output node, ∆Vjr is the weight change, Oj and Or are activation values at hidden neurons and output neurons, respectively, tr is the target value at the output neuron, and η is the learning rate. The weight updating equation between the hidden and input neuron is given by equation (7) [19].

$$\Delta {W}_{ij}=\eta {x}_{i}{\delta }_{j}\; {\delta }_{j}={O}_{j}*\left(1-{O}_{j}\right)\sum {V}_{jr}*{\delta }_{r}$$
(7)

where ‘i’ denotes input neurons, xi is the input value, ∑Vjr is the updated weight between hidden neuron ‘j’ and output neuron ‘r’. These iterations are carried on until the target value error is minimized.

Levenberg–Marquardt (LM)

The Levenberg-Marquardt was developed in the early 1960s for solving non-linear least-square problems. The loss function is the least square error in neural network training problems. For the given data points xi, yi, a model curve with parameters xi, βi is to be found such that the sum of the square of the error between the actual data and the model curve is minimum. The minimization is an iterative process that requires an initialization value. The Levenberg-Marquardt algorithm combines the Gauss-Newton method and the gradient descent method. The gradient descent reduces the sum of squared error by updating the parameter along the steepest descent. In the Gauss-Newton method, the squared error sum is reduced by assuming the error function is quadratic at the local point and minimized. When the parameter of the model curve is near the optimum value, then LM acts more like gradient descent, and when the model curve parameter is away from its optimum value, LM works more like the Gauss-Newton method.

Restricted Boltzmann machine (RBM)

The restricted Boltzmann Machine is a training algorithm generally used in deep neural networks. It is derived from Boltzmann Machine. In RBM, the neurons are bifurcated into two groups: visible neurons, where input is given, and hidden neurons. In RBM, the visible and hidden neuron groups are forbidden to have any connection within the group; they only have an inter-group connection. This is in contrast to Boltzmann Machine, where all neurons are connected, making RBM’s easier to implement and more efficient to train. The hidden neuron group is used to capture the probability of features. Random weights are initialized, and then Gibbs sampling is performed, simultaneously updating the weights of all neurons in a layer.

Performance and evaluation metrics

Applying an ML algorithm to solve a specific problem depends on how well it can compute the outcome. To quantify this performance, specific evaluation metrics are defined. This section summarises the top ten evaluation metrics used in all the literature studied for this paper. The performance of the literature survey in the subsequent sections is also highlighted on these ten-evaluation metrics; other scarcely used evaluation metrics are excluded.

A-20 INDEX

The A-20 index is the ratio of the M20 index to the total number of samples in the dataset.

$$A20=\frac{M20}{M}$$
(8)

where M20 is the number of samples whose ratio of experimental value to the predicted value by algorithm does not deviate more than ±20% from 1, thus M20 is the number of samples whose ratio of experimental value to predicted value is between 0.8 and 1.2. ‘M’ is the total number of data samples. For a perfect performance evaluation, the A20 index should be equal to 1.

Mean absolute error (MAE)

The nomenclature for subsections from B to J remains as follows: \({x}_{i}\) = Actual value of the ith data. \({\widehat{x}}_{i}\) = The predicted value of the ith data. \({x}_{mean}\) = The average value of the actual data of all the samples. n=The total number of data samples.

The mean absolute error (MAE) is the statistical mean of the absolute difference between the predicted and actual values, i.e., the error. The mean absolute error can be described as given in equation (9) [25].

$$MAE=\frac{1}{n}\sum_{i=1}^{n}\left|{\widehat{x}}_{i}-{x}_{i}\right|$$
(9)

For a perfect correlation, the MAE should be 0.

Mean absolute percentage error (MAPE)

It is the statistical mean of the percentage of error with respect to the actual value of the data [25], given as follows.

$$MAPE=\frac{1}{n}\sum_{i=1}^{n}\left|\frac{{\widehat{x}}_{i}-{x}_{i}}{{x}_{i}}\right|*100$$
(10)

Mean squared error (MSE)

Mean squared error is the mean value of the error squared; this can be considered the second moment of the error about its origin. This helps in understanding the variance and bias of the estimates [25].

$$MAE=\frac{1}{n}\sum_{i=1}^{n}{\left|{\widehat{x}}_{i}-{x}_{i}\right|}^{2}$$
(11)

Regression coefficient (R)

The regression coefficient indicates the correlation between the expected and actual outcomes. The regression coefficient lies between − 1 to +1, where -1 regression indicates an asynchronous correlation, whereas +1 represents a perfect correlation.

Root mean squared error (RMSE)

Root mean squared error is a non-negative value describing the predictive performance of the model. RMSE is useful when large errors are undesirable since the errors are squared before taking their average. RMSE value of 0 indicates the predicted and actual values coincide. It is the square root of the mean square error and is given as follows [25].

$$RMSE= \sqrt{\frac{1}{n}\sum_{i=1}^{n}{\left|{\widehat{x}}_{i}-{x}_{i}\right|}^{2}}$$
(12)

RMSE-standard deviation ratio (RSR)

RSR is the ratio of root mean squared error to the standard deviation of the original data [25].

$$RSR= \frac{RMSE}{\sqrt{\frac{1}{n}{\sum_{i=1}^{n}\left({x}_{i}-{x}_{mean}\right)}^{2}}}$$
(13)

R-squared

The R-squared value determines to what extent the predicted and actual values fit. An R2 value of 1 indicates the model can perfectly predict the outcome, whereas an R2 value of 0 means the model fails to predict [25].

$${R}^{2}=1-\frac{\sum_{i=1}^{n}{\left({x}_{i}-{\widehat{x}}_{i}\right)}^{2}}{\sum_{i=1}^{n}{\left({x}_{i}-{\widehat{x}}_{mean}\right)}^{2}}$$
(14)

Variance account for (VAF)

Variance Account For is the ratio of the variance of the error to the variance of actual data. It is often expressed as a percentage. The ideal value of VAF should be around 100% [25].

$$VAF\%=\left(1-\frac{var\left({x}_{i}-{\widehat{x}}_{i}\right)}{var\left({x}_{i}\right)}\right)*100$$
(15)

Weighted mean absolute percentage error (WMAPE)

WMAPE is a variant of MAPE in which the error value of ith data is intensified appropriately according to the weightage of ith data. Thus, an error for a higher magnitude data has more significant weightage in the total error computation. In this, the weighted mean is taken in which the actual value is multiplied by the error and divided by the summation of actual data [25].

$$WMAPE= \frac{\sum_{i=1}^{n}\left|\frac{{x}_{i}-{\widehat{x}}_{i}}{{x}_{i}}\right|*{x}_{i}}{\sum_{i=1}^{n}{x}_{i}}*100$$
(16)

AI for compressive strength prediction

This section bifurcates the algorithms into four subsections and discusses the individual techniques' methodologies, results and performance evaluation. The number of data points affects the model's performance depending upon the type of machine learning technique employed. Figure 6 shows the number of times different prominent features were used as input parameters in the AI-ML studies.

Fig. 6
figure 6

Bar graph showing prominent features used in all studies. The water-reducing admixtures (superplasticizer, plasticizer, high-range water-reducers, and similar) are grouped as ‘WRA’. The non-destructive tests are grouped as ‘NDT’, chemical properties of cement are grouped as ‘Cement chemical properties’. Microstructure studies, including microstructural imaging, are grouped as ‘Microstructural study’. Images other than microstructural images are grouped as ‘Surface Images’

Algorithms based on rule

In algorithms based on the rule, the inputs are defined as suitable membership functions to which rules are applied and relationships are generated. The output of this relationship is de-fuzzified, and a crisp output is generated. This section discusses various papers utilizing fuzzy logic and a combination of fuzzy logic and neural network.

Iqtidar et al. [28] predicted the compressive strength of concrete blended with rice husk using ANFIS architecture, ANN, MNLR, and LR. A Gaussian membership function was used to fuzzify the total of 192 input data. The input data was collected from Google Scopus, comprised of rice husk ash, age, superplasticizer, water, cement amount, and aggregates; this data was sent to nodes that multiplied the input with certain factors to generate weights for the next layer. The output weights were normalized to get the final weights; these weights were then multiplied to adaptive square nodes, a function of the input parameters. This was then sent to an aggregation layer to produce crisp output. ANFIS performed better than LR and MNLR in this experiment, with an R2 value of 0.89.

Tayfur et al. [29] predicted the compressive strength of high strength concrete using fuzzy logic and ANN. A total of 60 sets of data were produced using different binder contents prepared from the various percentages of silica fumes. Input data of silica fumes content, binder content, and age, and the output of compressive strength was fuzzified using the triangular membership function. Twenty-four fuzzy rules were defined, and the minimum operation was performed for inference to find the fuzzy output set.

The centroid method aggregated and de-fuzzified the outputs to get a crisp output. It was seen that fuzzy logic performed adequately better than ANN with a regression coefficient of 0.95. FL also outperformed ANN in MAE and RMSE evaluation metrics.

Abolpour et al. [30] compared the efficiency of fuzzy logic defined with the triangular membership function and Gaussian membership function in predicting the compressive strength of concrete. A data set of 1030 concrete mixtures from the University of California was used as the input for fuzzification. Input variables include weight percent of cement, water, blast furnace slag, fly ash, superplasticizer, fine and coarse aggregate, and concrete age in days, the output function being the concrete compressive strength. 897 fuzzy rules and five linguistic values were defined. It was seen that the triangular membership function had a MAE of 11.72% and performed better than the Gaussian membership function, with a MAE of 13.27%.

Sarıdemir et al. [31] predicted the long-term effect of GGBFS on the compressive strength of the concrete using fuzzy logic and ANN. The compressive strength was obtained from the literature data and was verified using fuzzy logic and ANN. The input data include the specimen’s age, water, aggregate, Portland cement, and GGBS. The input data was fuzzified using the triangular membership function. The product method was used as the inference operator, and the weighted average method was used for de-fuzzification. The FL correlated with the actual compressive strength with an R2 value of 0.99 and 0.97 for testing and training.

Özcan et al. [32] compared ANN and fuzzy logic for predicting the compressive strength of silica fume concrete. Forty-eight concrete mixtures were produced using four different water-cement ratios, three different cement dosages, and three partial silica fume replacement ratios. A triangular membership function fuzzified the inputs. Nine rules for compressive strength were formed. Mamdani’s inference system was used to get the crisp output. The experimental compressive strength was compared with the predicted compressive strength. The fuzzy logic showed promising outcomes for predicting the compressive strength of silica fume concrete with an R2 value of 0.93.

Topcu and Sarıdemir [33] used ANN and fuzzy logic to predict the compressive strength of concrete blended with different fly ashes having high and low lime percentages for 7, 28, and 90 days. The triangular membership function fuzzified the inputs. One hundred eighty experimental data were used, and Sugeno-type fuzzy rules were established to infer input and output. The product method was used as an inference operator. The crisp output showed a very high correlation with the experimental output, with an R2 value of 0.99 for training and testing.

Akkurt et al. [34] developed a fuzzy logic model to predict 28th-day cement compressive strength and compared it with the ANN method. The data used for the training algorithm was taken from a cement plant in Izmir, Turkey. A triangular membership function fuzzifies fifty sets of input data. The product operator defined the inference system using the Mamdani fuzzy rules. ANN was constructed with 4-4-1 architecture, suggesting four neurons in the input layer, 4 in the hidden layer, and 1 in the output layer. The average error for the fuzzy logic model was 2.69%; this could have been improved by having more than four input parameters. ANN model performed better with an RMSE value of 1.7. Table 1 summarises the rule-based algorithms reviewed in this paper with respect to the number of samples used for training and the input features used in training. Table 2 summarises the rule-based algorithms with respect to the evaluation metrics and their performance.

Table 1 Summary of Rule-based algorithms w.r.t features and data points
Table 2 Summary of Rule-based algorithms with respect to the evaluation metrics

Classical ML and clustering

In the classical machine learning and clustering section, classification and regression algorithms like support vector machine (SVM), support vector regression (SVR), various regression techniques, optimization techniques like genetic algorithms, particle swarm optimization, ant colony optimization and clustering techniques like the random forest, tree-based methods and various other similar techniques are included. The description of the methods, the features used, the number of data samples and the data source is summarised in Table 3. The summary of performance evaluation is summarised in Table 4.

Table 3 Summary of Classical ML and Clustering algorithms w.r.t features and data points
Table 4 Summary of Classical ML and Clustering-based algorithms with respect to the evaluation metrics

Güçlüer et al. [35] compared the performance of ANN, DT, SVM and LR in predicting concrete compressive strength. They utilized results from the non-destructive test of ultrasonic pulse velocity to find the compressive strength and utilized the outputs for training the algorithms. NDT was used to test 522 concrete cubes of standard 150mm size with compressive strength ranging from 35.07 MPa to 53.44 MPa. The M5P decision tree developed by Wang and Witten was used in this study. DT algorithm provided the best results with a regression coefficient of 0.86, MAE of 2.59 MPa and RMSE of 3.77 for 28 days of compressive strength.

Salami et al. [36] predicted the compressive strength of concrete having ternary-blend using the least square support vector machine (LSSVM) coupled with simulated annealing. A comparison was made with the genetic programming (GP) method used as a benchmark for performance. The dataset from the University of California dataset repository, Irvine, was used, and 1030 data values were used to train and test algorithms. The ternary blend consists of cement, blast furnace slag and fly ash. LSSVM coupled with simulated annealing outperformed GP with R2 of 0.982 and 0.954 for training and testing, respectively, while GP had R2 of 0.86 and 0.89 for training and testing, respectively.

Kang et al. [21] predicted steel-reinforced fibre concrete's compressive strength and flexural strength using 11 machine learning techniques and compared their performance. These techniques were the Gradient boosting regressor, XGBoost, RF regressor, AdaBoost, DT, KNN, LR, Ridge, SVR, Lasso, and MLP. The data for training and testing the algorithms were taken from different works of literature. It was seen that the water-cement ratio and the silica fumes used in concrete were the most prominent factors affecting the prediction of compressive strength. In contrast, the volume of steel fibre and the content of silica fume in concrete was more responsible for the prediction of flexural strength. It was seen that the extreme Gradient boosting (XGBoost), RF regressor and DT regressor performed better than the other soft computing techniques.

Ahmad et al. [37] used supervised machine learning techniques of bagging, GEP, AdaBoost and DT and compared their performance to predict the compressive strength of concrete blended with cementitious material. Statistical analysis was performed, and evaluation metrics RMSE, R2, MAE and MSE were determined. Along with this, the validity of algorithms was validated using K-fold cross-validation. The data of 1030 sets of values were distributed randomly and split into ten groups. Of the ten groups, nine were used for training and one for validation. This was repeated ten times, and the average of evaluation metrics was taken to get the performance of the algorithms. Bagging performed the best in both performance evaluation methods, followed by AdaBoost, GEP and DT.

Ahmad et al. [38] compared the performance of supervised machine learning algorithms to predict the compressive strength of concrete at high temperatures. Concrete subjected to high temperatures affects its internal structure and compressive strength significantly. The ML techniques adopted in this paper are bagging, DT, ANN and Gradient boosting. This study used nine input parameters: fine aggregates, coarse aggregates, temperature, silica fumes, cement, fly ash, nano-silica, water, and superplasticizer. DT and ANN were used as individual algorithms for forecasting the compressive strength at high temperatures, while Bagging and Gradient boosting were used as an ensemble. The statistical evaluation and K-fold cross-validation determined that bagging performed better than other algorithms with an R2 value of 0.9, followed by Gradient boosting, DT and ANN, with an R2 value of 0.89, 0.84 and 0.82, respectively.

Kova et al. [39] developed a model for predicting the compressive strength of self-compacting rubberized concrete using MLP-ANN, MLP-ANN ensemble, bagging, RF, Boosting trees, SVR and GPR. The total number of inputs was 10, whereas the output was single (compressive strength). For the bagging method, max trees were limited to 500; the minimum leaf size was kept between 2 and 15. The learning curve for bagging was seen to be saturated at 269 trees. For the RF method, the maximum trees were limited to 500, and the value of subsets for splitting was kept between 2 and 9. In boosting trees, the maximum number of trees was limited to 100 due to overtraining. The SVR model was analysed using linear, RBF and sigmoid kernel; the RBF kernel performed the best out of the three. In the GPR method, the standardization was done using a Z-score in which the mean value was made 0, the variance was made 1, and constant base function models were analysed.

The MNLR and LR technique by Iqtidar et al. [25] was used to predict rice husk ash concrete compressive strength was below average compared with other models used; LR performed the worst with an R2 value of 0.63 and MNLR having an R2 value of 0.7.

Khursheed et al. [40] predicted the compressive strength of fly ash concrete using MPMR, RVM, GP, ENN and ELM. For MPMR, the Gaussian noise was kept at 0.002, and the other tuning parameter was kept at 1.2; these were determined by the trial and error method. In the RVM model, the width of the radial basis kernel was determined using posterior modelling scrutiny, and the free parameter in the RBF kernel was determined as 0.04. For GP, the size of the population was kept at 800, the number of generations was kept at 400, frequency of generations was 50. The anxiety parameter in the ENN model was kept at 0.34, and the confidence parameter was kept at 0.66. The ELM model worked optimistically at six hidden neurons in a single layer, and the radial basis function was used as an activation function. The MPMR performed the best, followed by RVM, GP, ENN and ELM in decreasing order.

Salimbahrami and Shakeri [41] compared the performance to predict the compressive strength of recycled aggregate concrete using SVM and ANN. These were further compared with MLR using K-fold cross-validation. In this study, natural aggregate concrete, the control specimen and recycled aggregate concrete (with and without fibre) are produced in the lab and tested for 7 and 28-day strength. Workability was improved using a poly-carboxylic ether-based superplasticizer. In natural aggregate, the fine aggregate is procured from river sand, and coarse aggregates are crushed gravels with a maximum of 19mm in size. The influence parameter of the RBF kernel for SVM was kept at 0.008, the number of support vectors was 48, and the regularisation parameter C was kept at 1000. A multiple linear regression method was used to model the relation between output compressive strength and six input parameters, as described in Table 3. The input layer of ANN consists of 6 neurons, the optimum neurons in the hidden layer was determined to be 7, and the log-sigmoid function was used as an activation function and backpropagation algorithm for training. SVM and ANN performed the best, with SVM having a marginal advantage over ANN, and MLR performed moderately fine.

Rizvon and Jayakumar [42] predicted the compressive strength of recycled coarse aggregate concrete and correlated it with the hardened features of the concrete. Five replacement ratios of coarse aggregate (0 to 100%) were determined, and varying water to cement ratios (0.3, 0.4 and 0.48) along with superplasticizer were used to prepare 15 mixes. Multi-linear regression was used to predict the compressive strength with six independent variables. XGB model was also employed to predict the compressive strength of cylinder and cube strength. The XGB model was found to be 0.5% better than MLR in predicting the compressive strength of the cylinder.

Asteris et al. [43] predicted the compressive strength of cement-based mortar using KNN, SVM, RF, DT and AdaBoost. The optimum value of ‘k’ in KNN was determined as 3, and the Minkowski function was used in training. This study used the v-SVM, and a sigmoid kernel was used to build SVM. The number of trees for RF was kept at 10, and the split subset limit was greater than 5. The split subset limit for DT was greater than 6, and the minimum number of instances in DT was kept at 2, with a maximum tree depth of five. The number of estimators for AdaBoost was kept at 30. AdaBoost performed the best with an R2 value of 0.95 for testing, followed by RF, KNN, DT and SVM with testing R2 values of 0.94, 0.87, 0.85 and 0.42, respectively.

Feng et al. [44] compared the performance of AdaBoost with ANN and SVM in the prediction of the compressive strength of concrete. For implementing AdaBoost, the 1030 input and output parameters data was collected, a strong learner was generated, the learner was validated, and finally, the application was made to get the output of compressive strength. Classification and regression tree were used to generate the weak learner, and it was integrated by the median of weak learners. 90% of the data was used as a training set, and 10% was used as a testing set. The data was also trained using ANN and SVM. It was seen that AdaBoost performed better than ANN and SVM, with an R2 value of 0.982, followed by ANN with 0.903 and SVM with 0.855.

Sihag et al. [45] estimated the 28-day compressive strength of high strength concrete using GPR, SVM and ANN. The GPR and SVM had two models each, based on the kernel functions of Pearson kernel function and the Radial basis function. In GPR, the Gaussian noise was kept at 0.1; in SVM, the regularisation parameter C was kept at 10. A sensitivity analysis was performed by eliminating one input in every trial, revealing that cement and silica fumes are the most critical parameters in determining the output. The statistical analysis showed that ANN has better performance than GPR and SVM. Also, it was found that the Pearson kernel performed better than the RBF kernel.

Nyarko et al. [46] predicted the compressive strength of rubberized concrete using KNN, regression trees (RT), RF and ANN. A database of 457 data from the literature was used for training and testing purposes. For KNN, five-fold cross-validation for various values of ‘K’ ranging from 1 to 20 was performed, and the optimum value of ‘K’ was determined as 3. For RT, the minimum parent size was kept between 1 and 40, and five-fold cross-validation determined the optimum minimum parent size of 3. The optimum number of trees to be included in the RF model is 21. For ANN, three hidden layers were used with 9, 3 and 2 hidden neurons, respectively. A tan sigmoidal function was used as an activation function, and the Levenberg Marquardt training algorithm was used to update the weights. The ANN outperformed other methods in the testing phase with an R-value of 0.978, followed by RT, RF and KNN with an R-value of 0.914, 0.9 and 0.88, respectively.

Dutta et al. [47] compared the performance of GPR, MARS, and MPMR in predicting the compressive strength of concrete. A total of 1030 data sets were used in this study; 80% were used for training and 20% for testing. For GPR, the Gaussian noise was kept at 0.1, and the sigma parameter was kept at 0.001. The MARS model was built using 70 basis functions. For MPMR, a radial basis function is used. It is seen that the MARS model performs better with a testing regression value of 0.957, followed by GPR and MPMR with regression values of 0.9485 and 0.9352, respectively.

Chopra et al. [48] compared the performance of DT and ANN in predicting the compressive strength of the concrete at 28, 56 and 91 days. A decision tree model with 33 nodes was built. The number of observations used to build the RF model was 33, and for ANN, an architecture of 6 inputs, ten hidden nodes in the first hidden layer and one output is made. Performance evaluation of the model suggested that ANN performed better than the other two, with an R2 value of 0.95 in the testing phase, followed by 0,84 and 0.69 for RF and DT, respectively.

Mirzahosseini et al. [49] predicted the compressive strength of glass cullet-modified concrete using GEP. For this purpose, 70 numbers of 50mm mortar cubes with glass powder and constant water-cement and sand to cementitious ratios and varying glass powder replacement ratios were cast in the laboratory. Compressive strength was determined at 1, 7, 28 and 91 days. The study was conducted using two GEP models; the first model contained two curing temperatures of 23 and 50 degrees Celcius, for which all test results were included. In the second model, three curing temperatures of 10, 23 and 50 were considered valid on 45 out of 70 tests. Statistical performance evaluation suggested GEP performed exceptionally well with a regression value of 0.95.

Yaseen et al. [50] predicted the compressive strength of lightweight foamed concrete using ELM and compared its performance with MARS, M5 Tree model and SVR. Forty-six sets of input and output data were gathered from the literature study. The MARS model was built using the cubic function for better smoothness; a recursive partitioning regression was used to optimize function approximation. The M5 tree model utilized a split value of 2, a value of smoothening of 15 and 0.05 as the splitting threshold. The SVR model was built with an RBF kernel. The ELM model comprised three layers in the network architecture. Sigmoid was used as an activation function. An optimization study was done to find the optimal number of hidden neurons (between 50 and 150); thus, the model was run 500 times and mean square error was used to evaluate the performance. Performance evaluation of all the models suggested ELM performed the best out of the other models.

Basyigit et al. [51] predicted the compressive strength of concrete by image processing trained using LR, MLR, and NLR. For this, 28 cubes of 100mm were cast with seven different water to cement ratios (0.37 to 0.79), four concrete cubes for each ratio, three cubes were tested for compressive strength, fourth was used to take the image. The images were cropped into 2350x2350 pixels, and a histogram of the pixel value was plotted. The threshold values for aggregate, cement and air gap were defined. The theoretical and obtained values were trained using different methods to obtain the best regression. Multi-linear regression and Non-linear regression performed regression almost with equal efficacy.

The compressive strength prediction of cement by double-layer MEP was proposed by Akkurt et al. [52]; in double-layer MEP, the chromosomes, instead of a single layer, are placed in a double layer. Thus, the chromosomes are capable of holding more information and expressions. Further, crossover, mutation and fitness evaluation are performed. The performance evaluation of all the methods suggested that double-layer MEP performed better than others, with an RMSE value of 1.47.

Baykasoglu et al. [10] predicted the cement compressive strength using GEP and NN and compared it with regression analysis. A four-month (104 days) data was collected from a cement plant in Adıyaman, Turkey. Nineteen input features were used in this study, as described in the table below; the output was considered as 28 days strength. Gene expression programming was executed with an initial population size of 50, and between 3000 and 20000 generations were generated. Subsets of 19 input features are defined with relation, and their performance is checked; the best regression value for GEP is 0.775.

Shallow neural network

This section discusses the shallow neural network models used by researchers to predict the compressive strength of concrete. Shallow neural networks are characterized by a single layer of hidden node layer, while deep neural networks have more than a single hidden node layer [55]. This section discusses neural network models like ANN, MLP, MFNN and similar algorithms. The description of the methods, the features used, the number of data samples and the data source is summarised in Table 5. The summary of performance evaluation is summarised in Table 6.

Table 5 Summary of Shallow Neural Network Algorithms w.r.t features and data points
Table 6 Summary of Shallow Neural Network algorithms with respect to the evaluation metrics

A three-layered feed-forward artificial neural network was used by Güçlüer et al. [35] to predict the compressive strength of concrete. The hidden layer consisted of 3 nodes, and the sigmoid function was used as an activation function. They used a learning rate of 0.3. ANN performed second best after DT with a regression value of 0.856.

Kang et al. [21] used MLP to predict the compressive strength of steel fibre-reinforced concrete. MLP was modelled with eight inputs and two hidden layers. Keras was used to build the MLP model and backpropagation as the training function. ReLu was used as the activation function for the internal hidden layer, and the output layer utilized the Adam function as the activation function. MLP was performed with an RMSE value of 3.41 MPa and a MAE of 2.79 MPa.

Asteris et al. [56] developed a hybrid ensemble model to predict the compressive strength of concrete. The hybrid ensemble model (HENSM) was constructed by training four conventional machine learning algorithms using ANN. The conventional ML models include MARS–linear, MARS non-linear, GPR and MPMR. The University of California dataset repository was used to get 1030 sets of 6 inputs and one output of compressive strength. Around 721 data sets (approx. 70%) were utilized for training and the remaining model for validation. For ANN, only one hidden layer was considered with 20 neurons, achieving the best result. The maximum value of the basis function for MARS – linear and non-linear methods was taken as 29. The non-linear method had a cubic function. For GPR, the width of the radial basis function was determined by trial as 0.4, and the Gaussian noise was taken as 0.07. For MPMR, the radial basis width was determined as 0.7, and the noise was determined as 0.0047. The HENSM model outperformed all other individual ML algorithms to achieve an R2 value of 0.99 and 0.89 in training and testing, respectively. A score analysis was performed to rank the performance of all models, and HENSM scored highest, followed by GPR, MARS – L, ANN, MARS – C and MPMR in decreasing order.

The MLP-ANN and MLP-ANN ensemble were utilized by Kova et al. [39] to predict the compressive strength of self-compacting rubberized concrete. The MLP-ANN algorithm was trained using the LM training algorithm; the input layer consists of 10 nodes, while the output layer has one node of compressive strength. The optimum number of nodes in the hidden layer was found by trial and error as 8, and it utilizes the sigmoid activation function. In contrast, the output layer utilizes the linear activation function. Ensemble models were also created for generalization purposes, with 100 base models with varying nodes up to 16 numbers in the hidden layer.

The ANN model used to predict the compressive strength of Rice husk ash concrete has one input layer with ten neurons, two hidden layers with a maximum of up to ten neurons in each layer, and one output layer with one neuron. ANN performed the best of all the methods used in this study, with an R2 value of 0.98 [28].

Waris et al. [57] predicted the compressive strength of concrete with different cement replacement material (CRM) ratios using image processing and ANN. 18 concrete cylinders were cast for this purpose with different CRM ratios ranging from 0 to 25%. 9 cylinders were tested for their 14-day compressive strength under CTM. The remaining nine cylinders were cut into three slices; a DSLR took images of the surface of the slice. The captured images were converted to greyscale and resized to 256x256 resolution. Statistical features of mean, median and standard deviation were extracted from the pixels of the images; these features were taken as input, and the compressive strength, tensile strength and slump were taken as output. The ANN architecture has ten hidden neurons and two hidden layers. ANN performed well with an R2 value of 0.9865.

Asteris and Mokos [9] predicted the compressive strength of concrete using a backpropagation neural network. The study aims at correlating the experimental readings from the ultrasonic pulse velocity (UPV) test and Schmidt’s rebound hammer test to the compressive strength of concrete. Although these tests can directly indicate concrete's durability or compressive strength, they often exhibit large dispersion in their results and deviation from the actual value of compressive strength. Hence, the BPNN was applied with input as UPV and Schmidt’s rebound hammer value to overcome these limitations. Three BPNN architectures were used. The first architecture (2-25-1) had two neurons in the input layer, 25 in the hidden layer, and one in the output layer. The second architecture (1-26-1) had one neuron in the input layer, 26 neurons in the hidden and one in the output layer. The third architecture (1-28-1) had one neuron in the input layer, 28 in the hidden layer, and 1 in the output layer. A hyperbolic sigmoid transfer function was used as an activation function at the hidden and output layer. The optimum architecture was found to be 2-25-1 when both inputs were used, when only UPV was used as an input (1-28-1) performed better and when only a rebound hammer was used as an input (1-26-1).

Naderpour et al. [58] used a backpropagation neural network to predict the compressive strength of environmentally friendly concrete. The dataset of recycled aggregate concrete from various literature was used in this study. A neural network is built with six inputs, one hidden layer, one output, a backpropagation training algorithm and a log-sigmoid activation function. The optimum number of hidden nodes in the hidden layer was 18. The model’s performance was satisfactory and had a regression of 0.829 in testing.

Dogan et al. [59] predicted the compressive strength of concrete using image processing trained by ANN. One hundred forty-four concrete cylinders of 150mm diameter and 300mm length were prepared in the laboratory. A 50mm portion was cut from the top, and then the surface of 144 concrete samples was photographed using a digital camera in a controlled environment; after this, a compression test was performed on the samples to get the compressive strength as output data. The image was converted to greyscale with a pixel size of 3888 x 2592; thus, a matrix of grayscale values of size 3888 x 2592 was created. This was further simplified by taking mean, median, and standard deviation along the matrix rows and forming a final matrix of 3 x 2592. This matrix was used as an input layer in ANN. The hidden layer consisted of 16 nodes, and the ANN architecture was trained using Levenberg-Marquardt’s training algorithm. The model performed exceptionally well, with an R2 value of 0.98.

Li et al. [54] used image processing with neural networks to predict the compressive strength of the concrete. For this purpose, microstructure images of concrete were taken using micro-computed tomography. The grey-level histogram showed the quantities of different substances. The grey values of pixels are formed as a matrix, namely the grey level matrix. Along with this, a co-occurrence matrix is also formed, showing the surface's texture. A neural network is trained with input as eigenvalues of grey level and co-occurrence matrix, a hidden layer with ten hidden neurons and the output layer with a single neuron of compressive strength. Backpropagation is used as a training algorithm. This method was also compared with traditional ML techniques like MLR and GPR. The performance evaluation showed that ANN had the least MAE of 4.179, followed by 4.46 and 4.95 for GPR and MLR, respectively.

Dogan et al. [60] estimated the compressive strength of concrete using statistical features of concrete images. For this, 60 cubes of 150mm were cast in a laboratory-controlled environment, and the surface images were taken on 7 and 28 days along with NDT tests like ultrasonic pulse velocity. The data of the image was converted into the matrix, with each value of the matrix corresponding to the pixel value of the image. Further, statistical data like mean, standard deviation, the median was extracted from the matrix, which was later used as input in ANN; a Levenberg-Marquardt backpropagation algorithm trained the ANN. Four models of ANN were used for statistical data of mean, standard deviation, and median and when all used together, and hidden layer contained 12, 8, 26 and 14 neurons, respectively. ANN performed exceptionally well, with an R2 value of 0.99.

Onal and Ozturk [61] predicted the compressive strength of cement mortar by establishing a relation between the microstructure and compressive strength using feed-forward backpropagation neural network (ANN). For this purpose,130 samples of 50mm cement mortar cubes were prepared from four different chemical admixture doses and were tested on 1, 2, 7, 28 and 90 days for compressive strength. Before conducting compressive strength tests, microstructural studies were conducted on the polished surface of cubes. Features like the dendrite length, pore length, pore area ratio, average roundness and area of hydrated and un-hydrated parts were used as input in ANN. ANN was performed with an R2 value of 0.9971.

An ANN model was developed by Özcan [32] to predict the compressive strength of silica fume concrete. ANN architecture was built by six input features in the input layer, an optimum number of hidden neurons in the hidden layer was found by trial and error as 11, and the output layer consisted single output of compressive strength. The ANN was trained using the Levenberg-Marquardt training algorithm, and logarithmic sigmoid and pure linear transfer functions were used as activation functions. The ANN performed better with an R2 of 0.977 than its counterpart fuzzy logic method in this study.

The ANN model studied by Sarıdemir et al. [31] for predicting the compressive strength of concrete consisted of 5 input features specified in Table 5, 6 hidden neurons in the first hidden layer, five hidden neurons in the second hidden layer and compressive strength as output neuron. The backpropagation training algorithm and sigmoid activation function were used. Compared to Fuzzy logic, which was also undertaken in this study, ANN performed better with an R2 value of 0.981 in testing.

Topcu and Sarıdemir [33] compared ANN with fuzzy logic to predict the compressive strength of concrete containing fly ash. The ANN architecture had nine input features in the input layer, 11 hidden neurons in the hidden layer, and compressive strength as output in the output layer. Backpropagation and sigmoid function were used as training algorithms and activation functions, respectively. ANN performed marginally less than Fuzzy logic in this study, with an R2 value of 0.997 vs. 0.998 for Fuzzy logic.

Baykasoglu et al. [10] defined a neural network (NN) model to predict the compressive strength of cement. Up to 65 NN models were trained with different numbers of hidden layers (1 to 3) and different numbers of hidden neurons (7, 10, 13); the models were trained using the Delta-Bar-Delta algorithm. The best NN model secured an R2 value of 0.695.

Akkurt et al. [62] predicted the compressive strength of cement mortar using an ensemble of genetic algorithms and artificial neural networks called GA-ANN. A total of 150 data sets were collected from the cement plant for this study. This data set was divided into training and testing sets using GA. The training and testing data sets have almost the same range of data, and the average of both is similar to the average of the entire data set. This proper partitioning of the data ensures optimal learning from ANN. A sensitivity analysis was also performed using different input parameters, and it was found that an increase in tri-calcium silicate, sulfur trioxide and surface area increased the compressive strength of cement mortar.

Guang and Zong [20] predicted the compressive strength of concrete using multi-layer feed-forward neural networks (MFNN). In this study, 11 input features are used, as described in the table below. The neural network consists of 11 input neurons, seven hidden neurons in the hidden layer and one output neuron. Backpropagation was used as a training algorithm. In the first phase of NN training, 65 concrete cubes of 150mm were cast in the laboratory, out of which 50 were used for learning and 15 for testing. In the second phase of NN training, 100 data sets were collected from a cement plant, of which 85 were used for learning and 15 for testing. This method proved to be reasonably accurate in estimating the strength of concrete.

Deep machine learning

This section discusses deep machine learning models like a deep neural network, convolutional neural network, deep residual network, deep belief network, and similar algorithms. The description of the methods, the features used, the number of data samples and the data source is summarised in Table 7. The summary of performance evaluation is summarised in Table 8.

Table 7 Summary of Deep Machine Learning algorithms w.r.t features and data points
Table 8 Summary of Deep Machine Learning algorithms with respect to the evaluation metrics

Ly et al. [65] predicted the compressive strength of rubber concrete using a deep neural network (DNN). 233 datasets from different studies were used for training the deep neural network; this was named dataset 1 and contained 12 inputs, as mentioned in Table 7. Along with this, two supplementary datasets were also gathered from the literature, namely dataset 2 (187 samples and four input parameters which included compressive strength of controlling concrete as a new variable) and dataset 3 (183 samples and seven input parameters which included density of concrete as new variable). The DNN architecture had three hidden layers with neurons ranging from 1 to 20. The DNN was trained using Broyden–Fletcher–Goldfarb–Shanno Quasi-Newton backpropagation algorithm. DNN structure having 12 input nodes, 16 nodes in 1st hidden layer, 14 nodes in 2nd hidden layer, three nodes in 3rd hidden layer and one output node performed the best and showed a regression coefficient of R = 0.98.

Guo et al. [66] developed 3-dimensional microstructure images of concrete to estimate its compressive strength using a deep belief network (DBN). The 3-D microstructure of cement was obtained by micro-computed tomography. From the 3-D images, the grey level histogram, which suggests the hydrating components of cement and the grey level co-occurrence matrix, represents the spatial relationship in the hydrated cement. Statistical features like mean, energy, entropy, kurtosis, variance, and skewness are extracted from the grey-level histogram matrix. Statistical features like IDM (Inverse Different Moment), energy, entropy and correlation are extracted from the grey-level co-occurrence matrix. The extracted features form the input of the deep belief network. The training process of the deep belief network was paralleled on CUDA (compute unified device architecture). DBN performed well with a MAE of 2.81 MPa.

Abuodeh et al. [67] used deep machine learning techniques to predict the compressive strength of ultra-high-performance concrete (UHPC). A backpropagation neural network was implemented with sequential feature selection (SFS), and a neural interpretation diagram (NID) was used to verify selected parameters by SFS. Eight constituents were analysed from 110 UHPC compression data by executing SFS and NID, and the most influential constituents which improved backpropagation neural networks were selected. Abram’s compressive strength model was later modified using the influential features selected. The eight features used for analysis were cement, silica fume, fly ash, sand, steel fibre, quartz powder, water and admixture. Four influential features were cement, fly ash, silica fume and water. BPNN with SFS and NID performed significantly well with an R2 value of 0.801 compared to R2 of 0.215 before SFS and NID.

Huynh et al. [63] applied deep and shallow machine learning to predict the compressive strength of geopolymer concrete blended with fly ash. This research was carried out on the premise that conventional compressive strength prediction of geopolymer concrete requires a large volume of raw material, expensive equipment and extensive time. Deep neural networks (DNN), deep residual networks and ANN were employed for this purpose. DNN architecture comprises seven neurons in the input layer, 128 neurons in the first hidden layer, and 256 in the second. The deep residual network also consisted of 7, 128 and 256 neurons in the input layer, the first hidden layer and the second hidden layer, respectively; however, the deep residual network also had a third hidden layer with 256 neurons for the addition operation. The ANN architecture consists of 7 input neurons and 384 hidden neurons in the first hidden layer. The Adam gradient descent method was used as a training algorithm to update the weights, and ReLU was used as an activation function. It was seen that the deep residual network performed best with an R2 value of 0.934, followed by DNN and ANN with R2 values of 0.912 and 0.893, respectively.

Jang et al. [68] estimated the compressive strength of the concrete using image processing based on deep convolutional neural networks (DCNN). For this purpose, three DCNN algorithms (AlexNet, GoogleNet and ResNet) were proposed. Three concrete cylinders of 100mm diameter were made with OPC for every three water-cement ratios (0.33, 0.5, 0.68) for 3, 7 and 28 days, thus making 27 samples. A portable digital microscope was used to get microscopic images of the top and bottom surfaces of the cylinder. Up to 200 images of a single surface at different sections and illumination levels were taken. Data augmentation, like rotation and flipping, was also applied. The final output layer of AlexNet, GoogleNet and ResNet was modified to Euclidean loss function. The weights were updated using backpropagation, and the activation function used was ReLU. ANN also analysed the image processing, and the performance was compared. It was inferred that DCNN with ResNet with an R2 of 0.764 performed better than the other two DCNN models, with GoogleNet having an R2 value of 0.748 and AlexNet with 0.745, ANN performed poorly with a 0.2 value.

Shin et al. [22] estimated the compressive strength of concrete using the digital vision-based method. For this purpose, a Deep Convolutional Neural Network using three modified algorithms, namely ConcNet_A, ConcNet_G and ConcNet_R, was used. ConcNet_A model is an improved version of AlexNet in which the learning rate is improved using ReLU. ConcNet_G is an improvised version of GoogleNet and uses 22 layers to overcome overfitting problems. ConcNet_R is an improvised version of ResNet, and in this, instead of 152 layers as in ResNet, 50 layers are proposed since the image size for training is considerably smaller than the actual size. For this study, an image size of 84x84 was chosen, randomly cropped from a 112x112 resolution image. Data augmentation, like flip, was also applied. The ConcNet_R performed better prediction with an RMSE of 3.56.

A digital image correlation (DIC) technique was used by Afrazi et al. [70] to identify the surface displacement field and crack initiation along with its propagation in quasi-brittle materials, and a numerical finite element program was also developed to predict the same. The results were consistent with physical measurements. Majedi et al. [71] from their study developed a micromechanical model to overcome the contact problem in finite element modelling. This model improved the results and was comparable with the experimental results.

Deng et al. [53] used CNN to predict the compressive strength of recycled concrete. The concrete samples were prepared with recycled fine as well as coarse aggregates. Mix designs with 16 different water-cement ratios, 16 different fly ash replacement ratios, 21 different coarse aggregate replacement ratios and 21 different fine aggregate replacement ratios were prepared, thus totalling 74 samples. The performance of CNN was then compared with BPNN and SVM. Four inputs were considered, and the architecture of BPNN considered nine hidden nodes in the hidden layer. CNN was trained till 33 epochs. The prediction capability of CNN was comparatively better than BPNN and SVM.

Nguyen et al. [64] used a deep neural network to predict the compressive strength of foamed concrete. The performance was compared with conventional ANN and second-order ANN used by Fan et al. [66]. A higher-order DNN was developed based on the second-order neuron architecture; the higher-order neuron comprises three activation functions, and a second-order neuron could be considered a particular case. Instead of the sigmoid function as used in second-order neurons, here ReLU activation function was used. The evaluation of the performance of higher-order DNN was the best, with a regression coefficient of 0.97 followed by 0.93 and 0.9 for second-order neurons and conventional ANN.

Li et al. [72] estimated the compressive strength of cement using CNN and image processing. For this purpose, microstructure images of concrete were taken using micro-computed tomography. The different grey levels in the image correlate to different substances after hydration. Microstructure images are taken as input layers for CNN, and the convolution layers perform the feature extraction and mapping. The convolution layers contain convolution kernels that abstract features from the concrete images. The CNN architecture in this study consisted of 1 input layer, two convolution layers, two sub-sampling layers, one full connection layer and one output layer. This method performed exceptionally well with a MAE of 3.606 MPa.

Barkhordari et al. [73] predicted the compressive strength of fly ash concrete using several deep neural network model ensembles like a simple averaging ensemble, weighted averaging ensemble, super learner, integrated stacking ensemble, and separate stacking ensemble using various regressors. In this study, 6 DNNs are used for basic learning. The simple averaging ensemble is the average of six DNNs; the weighted averaging ensemble is the weighted average of the six DNNs; stacking ensemble merges the sub-models into a single model; this was performed using various regressors like SVM regressor, AdaBoost regressor, RF, Bagging and Gradient boosting regressor. The integrated stacking ensemble used a neural network in which 6 DNN sub-models were used in the input layer. By trial and error, the hidden neurons were decided in the single hidden layer.

Conclusion and future research

Challenges and concluding remarks

Compressive strength is one of concrete's most important mechanical properties, and determining the same requires expensive equipment and is time-consuming. AI-based techniques showed promising results in predicting the compressive strength of different types of concrete with various cementitious materials and cement mortars. These methods can significantly reduce the cost, labour, time and material involved in the determination. For this purpose, this study reviewed literature from various databases and summarised their findings. The techniques used, dataset, number of data points, evaluation parameters, and performances were also summarised. Following conclusions and challenges were drawn from the studies.

Generalisation

Several types of concretes, like concrete blended with cementitious and waste materials (fly ash, GGBS, silica fumes, rubber, rice husk ash), were modelled using soft computing techniques; however, a unified approach to generalize all concrete types is not perused.

Sensitivity

Soft computing techniques are sensitive to the parameters chosen to build the model and the parameters used to build the dataset. Generating synthetic data or collecting extensive data can help in these scenarios.

Precision

A good prediction for the compressive strength of concrete or cement should have the least deviation from the actual value for different hydration chronologies on various days. Soft computing techniques have been shown to have some errors of different magnitudes at various hydration levels; this needs to be minimized.

Large scale application

Although a lot of research is being carried out in this domain, its practicality and applicability to actual structures on a large scale are minimally carried out. Soft computing techniques like digital vision and image processing can help capture more extensive data and produce quicker results. The chemical formulation of cement and other materials can also be utilized using soft computing techniques to predict compressive strength.

Future research

After analysing the research to date, as per the authors' best knowledge, it was established that the research in the prediction of compressive strength using soft computing techniques is majorly focused on developing models and predicting the compressive strength for hydrated or partially hydrated concrete. Efforts to predict the compressive strength of hardened concrete while it is in its fresh or un-hydrated state were still lacking. Thus, the authors of this paper propose a novel research direction of predicting the compressive strength of hardened concrete while the concrete is still in its fresh state or workable.