Introduction

Concrete (Prasad et al. 2020; Vieira and Figueiredo 2020;Kaufmann 2020;Honglei et al. 2020) is defined as the core building material that deployed all around the world. It is highly noted for its durability, abrasion resistance, resistance to fire, high compressive strength, and impermeability. It is very much essential to compact the concrete for contributing the maximum structured strength. The concrete quality that satisfies the aforesaid need is explained as workability. It is a parameter that defines the concrete property, which identifies the effort needed for compaction, placing, and finishing with reduced homogeneity loss. The overall work required to initialize and maintain flow aids in the determination of the effort needed to place a concrete mixture. On the one hand, this mainly relies on the lubricant’s rheological property (the cement paste) and the internal friction amongst the aggregate particles, in contrast to the friction among the surface and concrete of the model.

The concrete workability (Sokhansefat et al. 2019;Abdalla et al. 2019;Fang et al. 2018;Li et al. 2018) typically relies on numerous factors and one of them is the water-cement ratio, which greatly affects the workability. The workability and water-cement ratio factor are directly proportional to each other. So, the maximization in the water-cement ratio maximizes the concrete workability as well. The test that is used to calculate the parameters nearer to workability is termed the slump test (Domone 1998;Yuan et al. 2020; Tay et al. 2019;Lu et al. 2015) and further offers helpful information on this workability. The concrete consistency is commonly measured by this method and it is deployed mostly on the sites or labs. The drop in the peak of the slumped fresh concrete is measured to deduce the slump from this test. Additional information regarding the concrete workability might be gained by analyzing the slump shape in concrete.

Each construction type needs concrete testing for identifying the slump obtained from the fresh concrete to assure the concrete with needed strength and workability, though no much research works have existed in the literature work with respect to slump prediction of concrete. The purpose of the supply chain in the cement industry is to propose the right supply chain for cement and to prove that the supply chain can produce value for cement companies (Taghipour et al. 2013; Vosooghidizaji et al. 2020; Agudelo and Isabel 2009). The researchers still look at the parameter characterization, which impacts the concrete’s slump value. Furthermore, the technical personnel has tried on various mixture proportions for gaining the concrete of suitable and desired workability, but it results in material wastage and cost and is time-consuming. Therefore, machine learning models (Nilsen et al. 2019; Feng et al. 2020; Bayar and Bilir 2019; Rousseau et al. 2019; Zheng et al. 2019; (Yu et al. 2018) (Yu et al. 2019; Nguyen et al. 2019; Nguyen et al. 2020) (Beno et al. 2014) need to be developed for the sake of minimizing the design cost and saving time and can be the better option for predicting the concrete slump. CNN is more efficient than traditional classifiers, because it reduces the number of parameters to be given as input for learning. It uses pooling and convolution operations thereby perform parameter sharing and so it can perform the learning with a limited number of attributes. This makes the CNN to work on any device. Though it has several advantages than other deep learning algorithms, it is not much used in literary works.

The major contribution of this work is stated below:

  • This work introduces a new fine-tuned CNN design for predicting the slump of concrete, which is done by the logic of optimization.

  • Presents a novel hybrid concept named LU-SLnO, which is the hybrid version of sea lion and dragonfly models.

  • The proposed work in the view of performance is compared and proved over other classical works with respect to error and convergence analysis.

The paper organization is as follows: The “Literature survey” section defines the literature survey on review papers under the topic slump prediction in concrete. The “Designing of optimization-assisted deep learning for slump prediction” section elucidates the designing of optimization-assisted deep learning for slump prediction. The “Proposed hybrid algorithm: objective function and solution encoding” section describes the proposed hybrid algorithm with objective function and solution encoding. The “Results and discussions” section manifests the results along with their discussions. Finally, concludes the paper in the “Conclusions” section.

Literature survey

Related works

Amlashi et al. (2019) have introduced prediction models for predicting the strength of elastic modulus, slump, and cubic samples of BPC. In this, soft computing models like M5Tree, ANN, and MARS were deployed and differentiated. Furthermore, the analysis was made regarding the parameters as well. The experimental analysis thus validated the improved performance of the ANN method over other comparative methods with the precise prediction of these parameters. The slump of BPC gets impacted more by the water variables and less by sand variables.

Shafiq et al. (2018) have presented the investigational outcomes in response to the impact of UEO on hardened concrete and slump properties. For this investigation, the preparation of three concrete groups has been made. A control mix along with the mix of 0.15% UEO dosage was composed within every group. From this, it can say that the small dosage of UEO has increased the reasonable slump amount in concrete by calculating the slump measurement of fresh concrete. Typically, the result has confirmed that there was a substantial minimization in oxygen porosity coefficient and permeability of entire concrete mixes using the small dosage of used engine oil, whereby termed as the pointer of improved long-term durability.

Chandwani et al. (2015) have exposed the work of combining two different nature-inspired computational intelligence approaches like ANN and GA to model the RMC slump. This was exploited using the design mix constituents such as water-binder ratio, cement, coarse aggregates, sand, fly ash, and admixture. Under six diverse statistical parameters, the hybrid model (ANN-GA) was scrutinized in terms of performance over the BPNN model. From this study, the prediction accuracy and convergence speed of ANN seem to get improved using hybridized ANN and GA. The concrete slump was predicted using this trained hybrid approach. This does not need any usage of multiple trials with diverse design mix proportions.

Chen et al. (2014) have processed several parallel subpopulations for avoiding the local optima and also for improving the search diversity at the instance of the GEP optimization process. Through the whole subpopulations, the fascinating solutions were rapidly distributed by the hyper-cubic topology. The case learning thus revealed that the parallel hyper-cubic GEP needs to be more precise than the fundamental GEP and the other two regression methods for evaluating the HPC slump flow. In this, the implemented model was highly preferred since it provides clear formulas with measurable parameters, even though both the implemented model and BPNN provides similar performance.

Meng et al. (2019) have investigated the impact of slump-retaining PCE dispersant, SF, and their mixture, over the hardened and fresh properties of normal Portland cement paste. The properties used for investigation were rheological properties, hydration kinetics, compressive strength, and setting time. The resultant outcomes have formulated, in which the cement hydration rates were accelerated and initial/final setting time was reduced by SF. To the end, there established a feasible correlation among rheological properties, hydration kinetics, and compressive strength evolution. This in turns offered a fundamental to optimize binder formulation and proportion of high/ultra-large performance concrete.

Bondar et al. (2018) have highlighted the investigational findings of AASC for offering an inclusive view on the impact of design variables mixture related to strength, slump, and chloride binding and transport. The result thus verified that AASC has been modeled for diverse grades and diverse workability of concrete. The diffusivity outcomes have formulated that the excess addition of water has not managed the pore connectivity/structure in AASC. Hence, the designing of AASC was made on the basis of the water/binder ratio required for a particular mechanical performance. Furthermore, it was observed that there was an increase in the chloride binding capacity along with the raise in the silica content of the activator and/or the paste content of the concrete.

Agrawal and Sharma (2010) have intended to demonstrate the probable application of NN for slump prediction in HSC. Based on the existing test data having 349 diverse concrete mix designs of HSC collected from a specific RMC batching plant, the NN was trained, tested, and modeled. The prediction of the slump in concrete was made using the most flexible NN model. The input parameters utilized in this dataset were fly ash, cement, coarse aggregate (10 mm), sand, water, water/binder ratio, and super-plasticizer. The experimental outcomes were differentiated based on the performance function or error function.

Ma et al. (2017) have introduced an enhanced mix design approach of SCC in correspondence to the correlation between average diameter of coarse aggregate, slump flow, and coarse aggregate volume. Similarly, in the volume of coarse aggregate condition, the small SF was made suitable by large coarse aggregate average diameter; still, the small coarse aggregate average diameter had matched large SF. The investigational outcomes have thus demonstrated that the SCC properties involving compressive strength and workability have satisfied the requirements. This enhanced mix design model aided in advancing SCC and cuts the cost in real-world engineering projects with better results.

Review

Table 1 shows the features and challenges of the conventional methods regarding concrete slump prediction. ANN is the used method in (Amlashi et al. 2019) that attains minimum values for RMSE, MAD, and MAPE parameters with the reduced slump, compressive strength, and elastic modulus, even though sensitivity analysis is needed to conduct for future use and recommends incorporating the global optimization bio-inspired algorithms. Modified Darcy’s equation (Shafiq et al. 2018) achieves high performance and has reasonable minimization in total porosity. But, there is no concern on UEO effect on the long-term durability of concrete. ANN-GA (Chandwani et al. 2015) has better robustness with fast convergence to the global optimum and poses a better prediction of slump value, even though we plan to use the ELMs to model the material behavior of concrete and the overall performance needs to be improved further. Parallel hyper-cubic GEP (Chen et al. 2014) achieves lower RMSE error and has better accuracy in estimating the slump flow of HPC. But, the prediction ability is restricted because of high non-linear problems. The Bingham fluid model (Meng et al. 2019) has enhanced cement hydration kinetics and has reduced plastic viscosity and yield stress. However, it poses fewer hydration enhancements and the difference in strength between plain and binary pastes is significantly smaller. Fick’s second law of diffusion (Bondar et al. 2018) produces AASC of required strength and poses better workability. Yet, further study is needed to discern the binding capacity. ANN (Agrawal and Sharma 2010) is convenient and easier in predicting any mix proportions and has a better estimation of slump prediction. It still needs further studies for the slump prediction using some soft computing techniques. Improved mix design method (Ma et al. 2017) has excellent workability and low cost and is very useful in developing SCC. For future work, more attention is needed on SCC workability.

Table 1 Features and challenges of conventional concrete slump prediction models

Designing of optimization-assisted deep learning for slump prediction

This paper plans to develop a novel machine learning model for predicting the slump in HSC. The block diagram of the concrete slump prediction modelis depicted in Fig. 1. From the literature work, most of the prediction model relies with the application of the ANN model. This research work makes an attempt to design a new optimized CNN for predicting the slump, which is trained with the input data patterns. They are cement (kg/m3), super-plasticizer (kg/m3), slag (kg/m3), fly ash (kg/m3), water (kg/m3), coarse aggregate (kg/m3), and fine aggregate (kg/m3). Furthermore, the synthetic data are as well generated with the above data. Moreover, the fine-tuning of filter size in the convolutional layer makes the model more accurate and precise in predicting the slump. For this optimization purpose, this work introduces a new “hybrid” algorithm termed LU-SLnO.

Fig. 1
figure 1

Block diagram of the concrete slump prediction model

Design of fine-tuned convolutional neural network

Even after applying the computer vision tasks within the NNs, the prior knowledge integration within the network architecture is a crucial one for superior generalization performance. The spatial information practice is the main intention of CNN O’Shea and Nash (2015) and hence, it relies on discrete convolution.

Convolutional layer

The convolutional layers have to use small filters fs (e.g., 3 × 3 to the maximum as 5 × 5), based on a stride of S = 1 and, significantly, padding the input volume with zeros whereby the convolutional layer has not changed the input’s spatial dimensions. In this work, the filter size fs is optimally tuned using the proposed LU-SLnO algorithm.

Let the convolutional layer is assumed as cl. Consequently, the input of the layer cl consists of \( {p}_1^{\left( cl-1\right)} \) feature maps from the earlier layers, each of size \( {p}_2^{\left( cl-1\right)}\times {p}_3^{\left( cl-1\right)} \). While cl = 1, the input remains as the single data cl that is the composition of one or more channels, which in turn accept the raw data as input of CNN. The layer cl poses the output that contains \( {p}_1^{cl} \) feature maps of size \( {p}_2^{cl}\times {p}_3^{cl} \). \( {\hat{X}}_i^{cl} \) delineates the ith feature map in layer cl and is formulated as per Eq. (1).

$$ {\hat{X}}_i^{cl}={D}_i^{(cl)}+\sum \limits_{j=1}^{p_1^{\left( cl-1\right)}}{P}_{i.j}^{(cl)}\ast {\hat{X}}_j^{\left( cl-1\right)} $$
(1)

In which, bias matrix is expressed as \( {D}_i^{(cl)} \) and the filter of size \( 2{s}_1^{cl}+1\times 2{s}_2^{cl}+1 \) linking the jth feature map in a layer cl − 1 along with the feature map in cl is depicted by \( {P}_{i.j}^{(cl)} \). The output feature map has provided the size as per Eq. (2).

$$ {p}_2^{cl}={p}_2^{\left( cl-1\right)}-2{s}_1^{cl}\kern0.24em and\kern0.24em {p}_3^{cl}={p}_3^{\left( cl-1\right)}-2{s}_2^{cl} $$
(2)

Frequently, the filters are deployed for measuring the similarity of the fixed feature map\( {\hat{X}}_i^{cl} \), i.e., \( {P}_{i.j}^{(cl)}={P}_{i.k}^{(cl)}\; for\;j\ne k \). Every feature map \( {\hat{X}}_i^{cl} \) in the layer cl consists of \( {p}_2^{cl}.{p}_3^{cl} \) units organized in two-dimensional array form. The output is calculated as per the unit at position (g, h) is portrayed in Eqs. (3) and (4).

$$ {\left({\hat{X}}_i^{cl}\right)}_{g,h}={\left({D}_i^{(cl)}\right)}_{g,h}+\sum \limits_{j=1}^{p_1^{\left( cl-1\right)}}{\left({P}_{i.j}^{(cl)}\ast {\hat{X}}_j^{\left( cl-1\right)}\right)}_{g,h} $$
(3)
$$ \kern0.36em ={D}_i^{(cl)}\Big){}_{g,h}+\sum \limits_{j=1}^{p_1^{\left( cl-1\right)}}\sum \limits_{d=-{s}_1^{cl}}^{s_1^{cl}}\sum \limits_{e=-{s}_2^{cl}}^{s_2^{cl}}{\left({P}_{i.j}^{(cl)}\right)}_{d,e}{\left({\hat{X}}_j^{\left( cl-1\right)}\right)}_{g+d,,h+e} $$
(4)

In which, \( {\hat{X}}_j^{\left( cl-1\right)} \) is referred as the trainable weight of the network and the bias matrix is explicated as \( {D}_i^{(cl)} \). The skipping factors \( {u}_1^{cl} \) and \( {u}_2^{cl} \) is evaluated by deploying subsampling. The fundamental notation is to fix the count of pixels in the vertical and horizontal direction, once before the filter application. The size of the output feature maps by employing the skipping factor is evaluated based on Eq. (5).

$$ {p}_2^{cl}=\frac{p_2^{\left( cl-1\right)}-2{s}_1^{cl}}{u_1^{cl}+1} and\;{p}_3^{cl}=\frac{p_3^{\left( cl-1\right)}-2{s}_2^{cl}}{u_2^{cl}+1} $$
(5)

Non-linearity layer

Let the layer cl is considered as a non-linearity layer, wherein the input is \( {p}_1^{cl} \) feature maps and the output included again with \( {p}_1^{cl}={p}_1^{\left( cl-1\right)} \) feature maps offered the size of each as\( {p}_2^{\left( cl-1\right)}\times {p}_3^{\left( cl-1\right)} \), which is stated in Eq. (6).

$$ {\hat{X}}_i^{cl}=f\left({\hat{X}}_i^{\left( cl-1\right)}\right) $$
(6)

In which, the activation function in the layer cl is signified as f and works on pointwise. The supplementary gain coefficient is formulated using Eq. (7).

$$ {\hat{X}}_i^{cl}={ga}_if\left({\hat{X}}_i^{\left( cl-1\right)}\right) $$
(7)

Rectification

Let the rectification layer be considered as cl. Each component has the absolute value with the feature maps and is evaluated based on Eq. (8) with the input comprised of \( {p}_1^{\left( cl-1\right)} \) feature map having size \( {p}_2^{\left( cl-1\right)}\times {p}_3^{\left( cl-1\right)} \).

$$ {\hat{X}}_i^{cl}=\left|{\hat{X}}_i^{cl}\right| $$
(8)

In which, the absolute value is measured as pointwise so that the output contains the\( {p}_1^{cl}={p}_1^{\left( cl-1\right)} \) feature maps with no alteration in size.

Feature pooling and subsampling layer

Considering cl as the pooling layer and their outputs is comprised of \( {p}_1^{cl}={p}_1^{\left( cl-1\right)} \) feature maps having minimized size. Typically, in every feature map, the pooling works by locating the windows at non-overlapping positions and sustains one value for every window, whereby the subsampling of feature maps is exploited. Two kinds of pooling are differentiated in this layer as follows:

  • Average pooling: The operation is exploited as average pooling when the boxcar filter is utilized and is demonstrated as QAg.

  • Max pooling: The maximum value of every window is assumed to be in max pooling and is expressed using QMax.

Fully connected layer

Let us assume the fully connected layer as cl. If the layer cl − 1 is not considered as fully connected, then the layer cl gets the input other than \( {p}_1^{\left( cl-1\right)} \) feature maps having the size \( {p}_2^{\left( cl-1\right)}\times {p}_3^{\left( cl-1\right)} \) and the j layer with the ith unit is measured as per Eq. (9).

$$ {\hat{x}}_i^{cl}=f\left({v}_i^{cl}\right)\kern0.24em with\kern0.24em {v}_i^{cl}=\sum \limits_{j=1}^{p_1^{cl-1}}\sum \limits_{g=1}^{p_2^{cl-1}}\sum \limits_{h=1}^{p_3^{cl-1}}{W}_{i,j,g,h}^{cl}\;\left({\hat{X}}_j^{\left( cl-1\right)}\right) $$
(9)

In which, the weight that associates the unit at the position (g, h) in the jth feature map of layer cl − 1 and the ith unit in cl is elucidated by \( {W}_{i,j,g,h}^{cl} \). The block diagram of the optimized CNN model is depicted in Fig. 2.

Fig. 2
figure 2

Optimized CNN model

Proposed hybrid algorithm: objective function and solution encoding

Solution encoding

As mentioned above, the filter size fs of the convolutional layer is optimally tuned using the proposed hybrid algorithm. The input given to the implemented work is elucidated in Fig. 3. The main objective defined in the implemented prediction model is given in Eq. (10), where err is the error among the predicted and actual value.

$$ obj=\min (err) $$
(10)
Fig. 3
figure 3

Solution encoding

Proposed LU-SLnO algorithm

The proposed LU-SLnO algorithm merges the principle of DA in the sea lion model that obviously deals with the better convergence rate and speed. The proposed SLnO includes three major phases:

  • Tracking and chasing of prey by their whiskers.

  • Pursuing and encircling the prey by calling other members of their subgroups to join them.

  • Attack the prey.

Mathematical modeling of the proposed algorithm is exploited with four stages that are termed as proposed tracking a detection phase, social hierarchy, attacking, and encircling prey.

Detecting and tracking phase

The whiskers facilitate the sea lion to sense the existing prey and for detecting their position when the direction of whiskers is in the opposite of the water wave’s direction, though the vibration of whiskers is less while its orientation is as same as the present orientation.

Sea lion positioned the prey’s location and call for other members to merge its subgroup for pursuing and hunting the prey. The leader is the sea lion which calls others and the updating of the position of target prey is exploited by other members. This algorithm assumes the target prey as the closer one to the optimal solution or recent best solution. Towards the following iteration, to get closer, the sea lion moves over the target prey. In this scenario, the position of the sea lion gets updated using the arithmetical model of Levy update in DA (Jafari and Chaleshtari 2017) given in Eq. (11).

$$ \overrightarrow{S}\left(t+1\right)=\overrightarrow{S}(t)+ Levy(y)\times \overrightarrow{S}(t) $$
(11)
$$ Levy(y)=0.01\times \frac{r_1\times \Phi}{{\left|{r}_2\right|}^{\frac{1}{y}}} $$
(12)
$$ \Phi =\left(\frac{\Gamma \left(1+\xi \right)\times \sin \frac{\pi \xi}{2}}{\Gamma \left(\frac{1+\xi }{2}\right)\times \xi \times {2}^{\left(\frac{\xi -1}{2}\right)}}\right) $$
(13)

Vocalization phase

Sea lions can adapt to stay in both land and water. The sound of sea lion in water can travel four times quicker than in the air. While chasing or hunting the prey, the communication of sea lions is made via numerous vocalizations. In addition, they have the capacity on detecting the sound both on and under the water. On identifying prey, the sea lion calls on other members for encircling and attacking the prey. This is arithmetically given in Eqs. (14), (15), and (16), where the speed of sea lion leader’s sound is portrayed as \( \overrightarrow{S_{leader}} \), the sounds speed in water and air is demonstrated as \( \overrightarrow{P_1} \) and \( \overrightarrow{P_2} \).

$$ \overrightarrow{S_{leader}}=\left|\left(\overrightarrow{P_1}\left(1+\overrightarrow{P_2}\right)\right)/\overrightarrow{P_2}\right| $$
(14)
$$ \overrightarrow{P_1}=\sin \theta $$
(15)
$$ \overrightarrow{P_2}=\sin \varphi $$
(16)

Attacking phase

In the exploration part, the sea lions’ hunting behavior is formulated based on two stages as follows:

  1. a)

    Dwindling encircling approach: This approach is explicated based on Levy value (Nair and Muthuvel 2019) K Levy(y) in Eq. (11) and is decreased steadily via a course of iteration from 2 to 0. This reducing factor directs the sea lion to forward on and to encircle the prey.

  2. b)

    Circle updating position: The bait ball of fishes is the prime target of sea lions and the attack is started from edges which is explicated as per Eq. (17), in which the distance among the search agent (sea lion) and best optimal solution (target prey) is represented as \( \overrightarrow{M}(t)-\overrightarrow{S}(t) \), the absolute value is indicated by ||, and the random number is manifested as l and falls between − 1 and 1.

$$ \overrightarrow{S}\left(t+1\right)=\left|\overrightarrow{M}(t)-\overrightarrow{S}(t).\cos \left(2\pi l\right)\right|+\overrightarrow{M}(t) $$
(17)

Prey searching

From the best search agent in the exploration phase, the position update of the sea lion is formulated. The search agent’s position update in the exploration phase is evaluated with respect to the chosen random sea lion. It can also say that the SLnO algorithm carries out a global search agent and determine the global optimum solution, while \( \overrightarrow{F} \) is larger than 1. This is defined as in Eqs. (18) and (19). The flowchart of the proposed LU-SLnO algorithm is exploited in Fig. 4. The pseudo-code of the implemented LU-SLnO algorithm is given in Algorithm 1.

$$ \overrightarrow{Dis}=\left|2\overrightarrow{B}.\overrightarrow{S_{rnd}}(t)-\overrightarrow{S}(t)\right| $$
(18)
$$ \overrightarrow{S}\left(t+1\right)=\overrightarrow{S_{rnd}}(t)-\overrightarrow{Dis.}\overrightarrow{H} $$
(19)
figure a
Fig. 4
figure 4

Flowchart of the proposed LU-SLnO algorithm

Results and discussions

Simulation setup

The evaluation of this implemented slump prediction model is made using MATLAB. The dataset used for this experiment is downloaded from the link: https://archive.ics.uci.edu/ml/datasets/Concrete+Slump+Test. The data set includes 103 data samples. There are 7 input variables namely cement, slag, fly ash, water, SP, coarse aggregate, fine aggregate, and 3 output variables slump (cm), flow (cm), and 28-day compressive strength (Mpa). Since this paper intents to predict the slump, we have used the slump (cm) as the output variable. Super-plasticizer has an effect on workability and to collect the data; 78 mix proportions were performed at the first phase of the experiment. Furthermore, 35 more proportions are extended to acquire the total samples of 103. More details about the experiment can be referred from Yeh (2008), Yeh and I-Cheng (2009), Yeh and I-Cheng (2008), Yeh and I-Cheng (2007), and Yeh and I-Cheng (2006). The analysis in this work is exploited by varying the learning percentage to a different extent as 50, 60, 70, and 80. For example, with a learning rate of 60%, there are 62 data samples used for training and validation for the remaining 38 data samples were used for the testing set. Furthermore, the 62% of data samples are divided into 70% for training and 30% for testing. Hence, 43 out of the 62 data samples are used for training, while 18 data samples are used for validation The experimental analysis is made by comparing the proposed work with the conventional models like SLnO (Masadeh et al. 2019), DA (Jafari and Chaleshtari 2017; Jadhav and Gomathi 2019), LA (Boothalingam 2018; (Brammya and Deepa 2019), and LA-RE (Shaswat, personal communication) under actual and predicted values and convergence. Furthermore, the overall error analysis is as well as formulated under certain error measures like RMSE, MD, MASE, SMAPE, one norm, MAE, infinity norm, and two norms. The purpose of the slump prediction method is to minimize the error between actual and predicted value. The formula for some errors is given below.

$$ MSE=\sum \limits_{i=1}^N{\frac{\left(\Pr {edicted}_i-{Actual}_i\right)}{N}}^2 $$
$$ RMSE=\sqrt{\frac{\sum \limits_{i=1}^N{\left(\Pr {edicted}_i-{Actual}_i\right)}^2}{N}} $$
$$ MAE=\sqrt{\frac{\sum \limits_{i=1}^N\mid \Pr {edicted}_i- Actual\mid }{N}} $$
$$ SMAPE=\frac{100}{N}\sum \limits_{i=1}^N\frac{\Pr {edicted}_i-{Actual}_i}{\Pr {edicted}_i} $$

Analysis on actual vs. predicted slump

The analysis on actual versus the predicted slump of the implemented model against the classical models is exhibited in Fig. 5. These values are plotted for different learning percentages like 50, 60, 70, and 80 and the results are analyzed. From the graphical representation, it is monitored that the predicted value of the proposed model has highly got closer with the actual slump while comparing over other compared models. The prediction rate is high if the actual and the predicted values are the same. Hence, from this analysis, it is clear that the implemented approach has achieved a superior prediction rate than any other models.

Fig. 5
figure 5

On actual slump vs. predicted slump of the proposed model over state-of-the-art models for different learning percentage a 50, b 60, c 70, and d 80

Convergence analysis

Figure 6 illustrates the convergence analysis of the proposed model against other compared models under diverse learning percentages. The error performance needs to be minimum for attaining the best prediction performance. Particularly, while considering the results under 50% of learning, the error is at the maximum level for both the proposed and conventional models and as the iteration gets increased, the error gets linearly decreased. At the 25th iteration, the proposed model remains with less error than in the range of ~ 8.5. Moreover, almost in all the iterations, the implemented approach is seemed to achieve better performance with minimum error in all the learning percentages.

Fig. 6
figure 6

Convergence analysis of the proposed model over state-of-the-art models for varied learning percentage a 50, b 60, c 70, and d 80

For learning percentage 50, the proposed model in the 5th iteration has achieved the least error than DA and LA by 9.73% and 4.47%, respectively. Similarly, for the 25th iteration, the proposed work is 26.63%, 29.96%, 31.42%, and 14.83% better from SLnO, DA, LA, and LA-RE, respectively, with the minimized error. The analysis is also made for all other learning percentages and the resultants are graphically illustrated, where the proposed model shows its betterment over other conventional models.

Overall error analysis

Table 2 manifests the overall error analysis of the implemented approach over other classical approaches with respect to diverse learning percentages. In learning percentage 50, the proposed work regarding MD measure is achieved better with least value, which is 31.23%, 17.26%, 41.44%, and 17.24% improved than SLnO, DA, LA, and LA-RE, respectively. In the view of the RMSE measure, the proposed work achieves minimum RMSE error than other conventional models like SLnO, DA, LA, and LA-RE by 31.02%, 16.89%, 41.72%, and 17.28%, respectively. Subsequently, on considering learning percentage 80, the proposed model in response to MASE measure has attained betterment with minimized error, which is 32.46%, 16.56%, 14.73%, and 3.6% better from SLnO, DA, LA, and LA-RE, respectively. Similarly, all the other learning percentages are analyzed for all the error measures and are plotted in the table given below. The resultant outcomes thus explicated that the implemented approach has achieved better performance with less error than the existing approaches, thereby validates the betterment of implemented work with increased prediction accuracy.

Table 2 Overall error analysis of the proposed model over conventional models

Conclusions

The main intention of this implemented approach was on finding the reasonable applicability of optimized CNN for slump prediction in HSC. The input that applied for the following process was super-plasticizer, slag, cement, coarse aggregate, water, fine aggregate, and fly ash. The prediction was made precise by incorporating the optimization logic in CNN by fine-tuning the filter size of the convolutional layer. This optimization was exploited by incorporating the two optimization algorithm concept and stated as LU-SLnO. The performance regarding the proposed work was evaluated and the results were compared over the traditional models in the view of convergence analysis and error measure. On considering convergence analysis, for learning percentage 50, the proposed model in the 5th iteration has achieved the least error than DA and LA by 9.73% and 4.47%, respectively. Similarly, for the 25th iteration, the proposed work is 26.63%, 29.96%, 31.42%, and 14.83% better from SLnO, DA, LA, and LA-RE, respectively, with the minimized error. The experimental results are encouraging and they have shown significant performance in predicting the slump. However, to ensure the reliability of the slump prediction, it is essential to experiment under diverse practical constraints and large datasets. Moreover, the real-time environment has to be simulated to the most possible extent. Our future research has been planned to accomplish diverse experimental platform and precise slump prediction model.