1 Introduction

In general, columns are regarded as the most important structural members that resist under applied loads via deformation, strain, and ductility and transmit them to underlying supports. Concrete columns are mainly designed to withstand compressive loads due to the inherent properties of concrete. The failure of such members as subjected by lateral loads is mainly due to the lack of sufficient shear bearing capacity and enough ductility. The strength and ductility characteristics of concrete columns can be enhanced by means of reinforcing with other materials such as steel, particularly, when they are subjected to lateral loads such as seismic and wind loads. Column confinement with transverse reinforcement, commonly in the form of closely spaced steel cross-ties, hoops, or spirals, substantially affect the strength and ductility of reinforced concrete (RC) structural members [1,2,3]. In other words, the most significant design consideration for ductility in plastic hinge regions of RC columns is the provision of sufficient transverse reinforcement [1, 2].

The relationship between the stress and strain for RC columns may be achieved through experimental studies by recording the amount of the column strain at specific intervals of applying stress. The result of such experiments can be demonstrated as a stress–strain curve (see Fig. 1).

Fig. 1
figure 1

Stress–strain curve for axial loading of confined and unconfined concrete specimens

As illustrated in Fig. 1, Ec is the concrete modulus of elasticity, AND εco and εcc are the corresponding strains to \(f_{{{\text{co}}}}^{\prime }\) and \(f_{{{\text{cc}}}}^{\prime }\) that are referred to the peak concrete strengths which represent the maximum load carried by the RC column for unconfined and confined cases, respectively. In pre-peak phase, the form of the stress–strain curve for confined concrete is mainly controlled by three parameters, i.e., \(f_{{{\text{cc}}}}^{\prime }\), εcc, and Ec [1, 2]. Based on the rate of confinement, it can be figured out that load-bearing capacity and corresponding ductility or strain in stress–strain curve grow with increase of the transverse confinement [4]. Besides, it can be seen that the confinement reinforcement increases the compressive strength of the concrete specimens and causes a more ductile response until reaching the ultimate concrete compressive strain of the concrete, εcu, where the first hoop fractures which mainly occur in the post-peak phase [1, 2].

To provide an efficient design, engineers are required to precisely realize the stress–strain behavior of RC columns which are the most important members in RC buildings and structures. In this regard, researchers may evaluate the problem through experimental studies to examine theories and hypotheses through varying and manipulating the variables of the experiment to realize how they affect the system behavior. However, experimental studies in this major are limited due to the high cost, cumbersome procedure of tests and lack of high-capacity testing equipment. Therefore, over the last few decades, several researchers have proposed various analytical-based models for indirect estimation of \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc of confined RC columns considering various variables affecting the system behavior which mainly include compressive strength of unconfined concrete, the amount, placement, and configuration of longitudinal and transverse reinforcement, ratio of the core and gross cross-sectional areas, yield strength of transverse reinforcement, spacing of hoops or lateral ties, cross-sectional shape, e.g., circular or noncircular and dimensions of the structural member and the applied load [1, 2, 5,6,7].

Mander et al. [1] conducted experimental and theoretical studies on circular, rectangular, and square full-scale RC columns at seismic strain rates to explore the effect of transverse reinforcement on confinement behavior and overall performance of concrete compressive strength and strain of RC columns. Thereafter, Mander et al. [2] tried to find an optimal model using their test results. According to their study, their obtained model for confined RC columns with circular cross section is as follows:

$$f_{{{\text{cc}}}}^{\prime }=f_{{{\text{co}}}}^{\prime }\left( - 1.254+2.254\sqrt {1+7.94f_{{\text{l}}}^{\prime }/f_{{{\text{co}}}}^{\prime }} - 2f_{{\text{l}}}^{\prime }/f_{{{\text{co}}}}^{\prime }\right)$$
(1)

where \(f_{{\text{l}}}^{\prime }={k_{\text{e}}}{\rho _{\text{s}}}{f_{{\text{yh}}}}/2\) and \({k_{\text{e}}}={(1 - s/2d)^n}/(1 - {p_{{\text{cc}}}}){\text{ }}\).

in which n must be considered equal to 2 and 1, where circular hoops and spirals are used, respectively:

$${\varepsilon _{{\text{cc}}}}={\varepsilon _{{\text{co}}}}\left[ {1+5(f_{{{\text{cc}}}}^{\prime }/f_{{{\text{co}}}}^{\prime } - 1)} \right]{\text{ }}$$
(2)

where εco should be considered equal to 0.002 if not available.

Saatcioglu and Razavi [4] observed that high-strength concrete columns needed more lateral ties to maintain the ductility than normal-strength concrete columns. It was observed that confined concrete strength and corresponding strain should be expressed in terms of equivalent uniform confinement pressure provided by the transverse reinforcement cage. The equivalent uniform pressure is obtained from average lateral pressure computed from sectional and material properties. Besides, the strain level was expressed in terms of confinement parameters [4]:

$$f_{{{\text{cc}}}}^{\prime }=f_{{{\text{co}}}}^{\prime }+{k_{\text{l}}}({\rho _{\text{s}}}{f_{{\text{yh}}}}/2)$$
(3)
$${\varepsilon _{{\text{cc}}}}={\varepsilon _{{\text{co}}}}\left[ {1+5{k_{\text{l}}}\left[ {{\rho _{\text{s}}}{f_{{\text{yh}}}}/(2f_{{{\text{co}}}}^{\prime })} \right]} \right]$$
(4)

where \({\text{ }}{k_1}=6.7{({\rho _{\text{s}}}{f_{{\text{yh}}}}/2)^{ - 0.17}}\) and εco should be considered equal to 0.002 if not available.

Hoshikuma et al. [6] proposed a stress–strain model of concrete that takes confinement effects into account is developed, based on test results of a series of compression loading tests of reinforced concrete column specimens. The specimens have circular, square, and wall-type cross sections, with various arrangements of hoop reinforcement so as to cover practical bridge column sections designed in Japan. The final obtained model is represented as follows [7]:

$$f_{{{\text{cc}}}}^{\prime }=f_{{{\text{co}}}}^{\prime }+3.83{\text{ }}{\rho _{\text{s}}}{f_{{\text{yh}}}}$$
(5)
$${\varepsilon _{{\text{cc}}}}=0.00218+0.0332{\text{ }}{\rho _{\text{s}}}{f_{{\text{yh}}}}{\text{ }}/f_{{{\text{co}}}}^{\prime }.$$
(6)

Besides, another model is proposed by Sakai [8] which has also been demonstrated by Oreta and Kawashima [7] in which parameters such as tie spacing and column dimension had been considered for indirect estimation of \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc are of circular cross-sectional confined RC columns. The obtained equation is represented as follows:

$$f_{{{\text{cc}}}}^{\prime }=f_{{{\text{co}}}}^{\prime }(0.94+4.7C)$$
(7)
$${\varepsilon _{{\text{cc}}}}=0.0024+0.064C$$
(8)

in which \(C={K_{\text{s}}}\left[ {{\rho _{\text{s}}}{f_{{\text{yh}}}}/(2f_{{{\text{co}}}}^{\prime })} \right]\) and \({K_{\text{s}}}=\left[ {1 - s/(d\tan 30)} \right] \geqslant 0.\)

In addition to several other studies conducted in the same field, different guidelines and specifications recommend different methods and models for computing the appropriate amount of transverse reinforcement for confinement purposes, e.g., ACI 318-08 [9], AASHTO [10], Caltrans [11] and Priestley [12], and each one considers different common and uncommon parameters which characterize the concrete, steel reinforcement, and cross-sectional details. However, the models and formulas suggested by guidelines and codes are often obtained based on analytical methods which are represented before, e.g., widespread used Mander’s model [2]. Besides, although numerous researchers have made great efforts to develop accurate stress–strain models of confined RC columns using analytical methods and assumptions, they often made their conclusions by optimizing the analytical results with their obtained experimental data through empirical methods. On the other hand, such models are designated through simplifying assumptions, and more importantly, they are initiated from few observations and controlling few models through simple statistical regression analyses to find the best model. This is due to the complex parameters affecting the system behavior which cannot thoroughly be considered for producing constitutive models of \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc of confined RC columns. Thus, the complexity of analysis of stress–strain behavior of RC columns accounting for the influences of different parameters implies that there is still the necessity for obtaining more comprehensive models.

With regard to the progress of computer software and hardware systems, many computational techniques have been motivated from the analogy of biological activities and natural evolution, e.g., genetic algorithms, neural networks, support vector machines, and such other algorithms which are branches of data mining, artificial intelligence, and soft-computing techniques. These approaches have been utilized for solving real-world problems, because they can learn from data to find an optimal model relating the output variable to inputs. It means that they do not require prior assumptions about the system of the problem under study. In addition, these approaches can definitely be employed for generating models for engineering problems using experimental results. The outcomes of such modeling techniques will be definitely a great combination of experimental studies and numerical methods. In addition, they have been successfully applied to many civil engineering prediction problems [7, 13,14,15,16,17,18,19,20,21,22,23].

Despite extending promotions in such artificial intelligence (AI) methods, there are few efforts in the literature for using them for prediction of confined compressive strength and strain of circular concrete columns. Recently, Oreta and Kawashima [7] have utilized artificial neural networks (ANNs), Tsai [13] and Tsai and Pan [14] demonstrated the capability of weighted genetic programming (WGP) and genetic programming polynomials (GPP) for modeling \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc of circular cross-sectional confined RC columns. It is worth mentioning that the models obtained by those AI approaches have represented better performance in comparison with the traditional analytical formulas; however, the ANN model is a black-box model which has not been stated as an explicit formula to be used for hand calculation purposes. In addition, the GP-based models developed by Tsai [13] and Tsai and Pan [14] were based on different combinations of input variables and their structures are also complex.

This paper explores the capability of a variant of genetic programming (GP), namely, linear genetic programming (LGP), for indirect estimation of the confined compressive strength or peak stress, \(f_{{{\text{cc}}}}^{\prime }\), and the strain at peak-confined compressive stress, εcc, of confined RC columns with circular cross sections using a database comprised of several costly test results conducted on different specimens previously published in the literature by Oreta and Kawashim [7]. With regard to the literature and the available database, two new models are obtained for predicting each of \(f_{{{\text{cc}}}}^{\prime }\) (MPa) and corresponding εcc (%) considering significant input variables including (1) compressive strength of unconfined concrete cylinder specimen, \(f_{{\text{c}}}^{\prime }\) (MPa). It is worth mentioning that the unconfined concrete cylinder strength \(f_{{\text{c}}}^{\prime }\) is selected as the input variable instead of the unconfined concrete strength of the column, \(f_{{{\text{co}}}}^{\prime }\) (MPa), considering that the associated information with \(f_{{\text{c}}}^{\prime }\) would be obtained easier as obtaining \(f_{{{\text{co}}}}^{\prime }\) requires using the same equipment for testing confined column specimen with the same column geometry which is a high-capacity equipment and testing would be highly costly; (2) core diameter of circular column, d (mm); (3) column height, H (mm); (4) yield strength of lateral or transverse reinforcement, fyh (MPa); (5) ratio of volume of lateral reinforcement to volume of confined concrete core, ρs (%); (6) spacing of lateral reinforcement or spiral pitch, s (mm); and (7) ratio of longitudinal steel to area of core of section, ρcc (%). Finally, \(f_{{{\text{cc}}}}^{\prime }\) and εcc are considered to be a function of the following parameters:

$$f_{{{\text{cc}}}}^{\prime }=f(f_{{\text{c}}}^{\prime },{f_{{\text{yh}}}},{\text{ }}{\rho _{\text{s}}},{\rho _{{\text{cc}}}},d,{\text{ }}s,H)$$
(9)
$${\varepsilon _{{\text{cc}}}}=f(f_{{\text{c}}}^{\prime },{f_{{\text{yh}}}},{\text{ }}{\rho _{\text{s}}},{\rho _{{\text{cc}}}},d,{\text{ }}s,H).$$
(10)

To show the capability of LGP method, out of data range, a new set of data which are not used in training of the model, considered as test data, used after training process. Consequently, to assess the accuracy of the obtained model validation, verification analyses are carried out and the results are compared with those provided by other researchers in the literature.

2 Linear genetic programming

In 1958, Freidberg [24] attempted to solve simple problems by writing computer programs, i.e., the lingua franca for expressing the solutions to a problem. Thereafter, in 1985, Cramer [25] introduced the concept of tree-based genetic programming by applying an evolutionary algorithm (EA) to the programs that were represented as tree-shaped structures. In 1992, Koza [26] extended the work conducted by Cramer [25] and introduced genetic programming (GP) as a paradigm for searching the most fit computer programs that produce some preferred output for specific inputs.

Genetic programming is a subdivision of bio-inspired computational intelligence approaches which is basically created using the synthesis of Darwinian ideas of genetic inheritance to solve complicated problems. The essential goal in GP is to find predictive tree-shaped models relating the output to input variables, where experimental data samples are known based on the concept of survival of the fittest. Linear genetic programming (LGP) is an extended variant of GP with a linear structure similar to the DNA molecule in biological genomes and the name ‘‘linear’’ in LGP refers to the structure of the program representation. LGP evolves developing sequences of instructions from an imperative programming language (C or C++) rather than expressions of a functional programming language such as LISP [27]. Therefore, LGP systems can run faster than comparable tree-based GP interpreting systems and the programs are much simpler to read and to operate and these result in obtaining better models with less customization [28]. Figure 2 represents a comparison between three different structure of a program evolved by tree-based and linear GP.

Fig. 2
figure 2

Comparison of a program structure evolved by tree-based (a) and linear-based (b) GP for the expression of y = f(v[i])=(v [1] − 2) × v [2]/v [3]

In general, in GP and similarly in LGP, a population of programs is randomly initialized and individuals of the population are evaluated using a fitness function. Those members with higher fitness are given a higher chance to become a parent for the next generation, the offspring. Then, selected members are simply copied or transformed, by chance, into new members mainly via reproduction, mutation and recombination or crossover. Finally, comparing members of generation through a fitness function, the process repeat until the convergence conditions are satisfied and the optimal program is selected [15, 21, 29].

During the process of training, there may be some individuals or offspring which have an acceptable error while evaluated by the fitness function. In this regard, reproduction is performed by simply copying a selected member from the current generation to the next generation without modifications. Crossover occurs between two instruction blocks, whereas mutation occurs on a single instruction or chromosome. The crossover operation works by exchanging continuous sequences of instructions between parents. As it is seen in Fig. 3a, a segment of random position and arbitrary length is selected in each of the two parents [f(0) and g(0)] and exchanged. If one of the two children would exceed the maximum length, crossover is aborted and restarted with exchanging equally sized segments. Two commonly used types of the standard LGP mutations are micro- and macro-mutation. The micro-mutation can change an operation type (e.g., change v [1] × v [2] to v [1] − v [2]), or can change the arguments of an operation (e.g., change v [1] + 1 to v [1] + v [1]). The macro-mutation can delete an operation (e.g., change v [1] + v [2] to v [2]), or insert an operation (e.g., change v [1] + v [2] to v [1] + (v [1] × v [1])) [30, 31]. Figure 3 illustrates typical crossover and mutation operation in LGP.

Fig. 3
figure 3

Typical genetic operations in LGP: a crossover and b mutation

The fitness function is a particular type of objective function that is used to assess how close an obtained solution is to achieving the set aims. In other words, how close the obtained results are to real experimental output data. In the present paper, the fitness function used to develop LGP programs which may simply be computed through a difference or error function is represented as follows:

$$f=\sum\limits_{{i=1}}^{N} {(\left| {{E_i} - {P_i}} \right|} )$$
(11)

where Ei is the output experimental or real value, Pi is the corresponding predicted value through the LGP solution case i, and N is the number of output data.

3 Experimental database

A reliable database containing 38 RC column load test results is acquired from Oreta and Kawashima [7] which is a data collection of previous studies conducted by Mander et al. [2], Sakai et al. [32], and Sakai [8]. It is true that more data may exist in the literature, but they may be not readily available or published or in some cases, and the published data are not complete considering the same variables. Thus, the same database is used for the development of the LGP models here which is efficient. The employed database is also brought here for more considerations (Table 1).

Table 1 Employed database for development of LGP models [7]

There are 15 column specimens from Mander et al. [2] and are 500 mm diameter and 1500 mm height. The core diameters of the columns are 438 mm. All columns consist of both lateral and longitudinal bars with varying sizes and spacing. Besides, seven columns are available from Sakai et al. [32] which are of 200 mm diameter and 600 mm height, and provided with both lateral ties and longitudinal bars 10 bars with 6.35 mm diameter. The N series are those provided with only one layer of lateral reinforcement, while the D series consists of two layers of lateral reinforcements. The 16 columns of Sakai [8] are of 300 mm diameter 280 mm core diameter and 900 mm in height. Sixteen longitudinal bars with diameter of 9.53 mm were provided. C1, C2, and C3 correspond to one, two, and three layers of lateral reinforcement, respectively. The descriptive statistics of the variables in the database is given in Table 2 for the aim of considering the statistical consideration of variables.

Table 2 Descriptive statistics of variables in the database

4 Development of LGP-based models

4.1 Data preprocessing

Data preprocessing such as data normalization is often employed in modeling process through machine learning techniques due to different scales of data, distributions of variables data sets and considering the limited range of data [33]. In this regard, a remarkable superiority of LGP and GP subdivisions over several computational and machine learning methods is the fact that input and output variables are directly expressed in terms of computer programs produced by genetic programming consist of functions that are natural for the problem domain and suitable for employed data sets [15, 26]. Therefore, normalization can be eliminated. However, there is another problem in model generalization which is called overfitting. It is a case in which the error on the training set is driven to a very small value, but when new data are presented to the model, the error become large. A commonly used approach to avoid overfitting is to test the model on another group of data which are not used in the training process. To avoid overfitting, it is proposed that the available database should be classified into three sets: (1) training; (2) validation; and (3) test subsets [15, 33, 34]. The training set is utilized to fit the models, the validation set is used to estimate prediction error for model selection, and the test set which is a group of unseen data sets is used for the evaluation of the generalization error of the selected model. Regarding to several suggested values in the literature, the training, validation, and testing data are taken as 50–70, 15–25, and 15–25% of all data, respectively, for the development of models in artificial intelligence (AI) techniques [29, 33]. Herein, among 38 data, approximately 70% data sets are used for learning phase, i.e., training and validation, and about the 30% remaining data sets were used for the testing of the obtained models as the same sets of data were utilized by Oreta and Kawashima [7] for development of ANN-based models for prediction of peak-confined compressive strength and strain of circular RC columns. It is worth mentioning that if the fitness values on training validation and test data are close, it can be concluded that overfitting is avoided [15, 26, 27].

4.2 Initialization of LGP training parameters

Herein, the Discipulus software was used for the development of each LGP-based model. Similar to other evolutionary algorithms (EAs), GP and its variants such as LGP are of many degrees of freedom which should be taken into account to obtain an accurate model. Proper initialization of parameters highly affects finding efficient and accurate models. In this regard, quite a few runs may be required to be done and LGP parameters would be changed for each run which are basically obtained via trial and error. However, these parameters may be applied with the aid of previously suggested values [15, 34]. Herein, the model development process was controlled by considering evolutionary parameters, as represented in Table 3.

Table 3 Different combinations of settings and training parameters for LGP model development

A run in LGP, similar to GP, proceeds through randomly generating initial populations of individuals or programs. Technically, the upper bound for the initial program length is a user-defined parameter and is illustrated in Table 3 as the maximum program size. The lower bound might be equivalent to the minimum length of a program, e.g., one instruction such as one variable, or can be an expression, i.e., more than one instruction. Although the success of finding a high accuracy model relies on the initial and maximum program size parameter, the optimal model is not necessarily the largest. In this regard, a major concern in development of models via GP is the way in which computer programs or individuals grow in size as the search progresses, what is known as bloat. Bloat is excessive code growth in size of the individuals of the evolving population without a notable improvement in fitness. Many methods have been proposed in the literature to control the bloat in modeling process [35]. To cope with this issue, LGP uses limited variable-size chromosomes to a maximum number of instructions, the upper bound, which should be defined in advance by the user [36]. Besides, LGP take advantage of using demes that are semi-isolated subpopulations due to the fact that the evolution process develops faster in them in comparison with a single population of equal size [18, 27]. These remarkable features of LGP in comparison with GP and its other variants make LGP faster which results in obtaining higher accuracy models.

4.3 The proposed LGP-based models

The optimal LGP programs would be achieved after performing the training process or after the convergence criteria have been reached in C++ format. The obtained code can be linked to the optimizer or may be called from the optimization routines [37]. For further considerations, the derived LGP-based codes for both \(f_{{{\text{cc}}}}^{\prime }\) and εcc are represented in “Appendix” at the end of the paper. In the present paper, to facilitate the use of the obtained codes for hand calculation aims and design purposes, they are simplified and converted into a functional representation by successive replacements of variables. The best LGP-based formulas of \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc of circular confined RC columns can be calculated using the following equations:

$$f_{{{\text{cc}}}}^{\prime }=f_{{\text{c}}}^{\prime }+{\rho _s}\sqrt {f_{{\text{c}}}^{\prime }{\rho _{{\text{cc}}}}{\rho _{\text{s}}}\frac{d}{H}\sqrt {{\rho _{\text{s}}}\frac{d}{s}+\frac{{{\rho _{\text{s}}}{f_{{\text{yhc}}}}}}{H}} }$$
(12)
$${\varepsilon _{{\text{cc}}}}=\frac{d}{H}+\frac{{(3.5{\rho _{\text{s}}}{f_{{\text{yh}}}}{\rho _{{\text{cc}}}} - 5.25s{\rho _{\text{s}}}{\rho _{{\text{cc}}}}^{2}+{\rho _{\text{s}}}H+H){\rho _{\text{s}}} - 3s+2d}}{{f_{{\text{c}}}^{\prime }H}}.$$
(13)

It is noteworthy that the statistical regression models, e.g., semi-empirical or empirical models, are mostly obtained after controlling only few equations established in advance. Thus, these models cannot efficiently take into account the interactions between output and independent input variables. In fact, the best model generated by the LGP technique is obtained after controlling numerous linear and non-linear preliminary models, e.g., the proposed design equation for \(f_{{{\text{cc}}}}^{\prime }\) is selected among a total of 628, 531,896 programs evolved and assessed by the LGP method during several runs. Besides, the obtained models have not been established by chance which represent the capability of LGP in finding optimal solutions.

5 Analyses and results

5.1 Prediction performance analysis of the model

To select an appropriate model, supplementary studies must be done and the performance and accuracy of the model should be precisely evaluated. This may be achieved through statistical analyses. According to Smith [38], if a model gives a correlation coefficient (R) > 0.8 (R2 > 0.64), there is a strong correlation between the predicted and observed values which means that the prediction capability of the model is acceptable. However, the negative point of considering R only as a model performance evaluation criterion is the fact that it is insensitive to additive and proportional differences between model predictions and observed values. Therefore, other criteria are required to be considered. The root mean square error (RMSE) and mean absolute error (MAE) may be used as other criteria. These indices are more sensitive to differences between predicted and observed values, particularly occasional large errors. RMSE serves to aggregate the magnitudes of the errors in predictions for various times into a single measure of predictive power and gives disproportionate weight to very large errors. Besides, MAE represents how big of an error can be expected from the forecast on average. The lower the RMSE and MAE values, the better the model performance would be. These parameters are calculated using the following equations:

$$R=\frac{{\sum\nolimits_{{i=1}}^{n} {({o_i} - {{\bar {o}}_i})({p_i} - {{\bar {p}}_i})} }}{{\sqrt {\sum\nolimits_{{i=1}}^{n} {{{({o_i} - {{\bar {o}}_i})}^2}} \sum\nolimits_{{i=1}}^{n} {{{({p_i} - {{\bar {p}}_i})}^2}} } }}$$
(14)
$${\text{RMSE}}={\left[ {\frac{{\sum\limits_{{i=1}}^{n} {{{\left( {{p_i} - {o_i}} \right)}^2}} }}{n}} \right]^{1/2}}$$
(15)
$${\text{MAE}}=\frac{{\sum\limits_{{i=1}}^{n} {\left| {{p_i} - {o_i}} \right|} }}{n}$$
(16)

where oi and pi are the actual and predicted output values for the ith output, respectively, \({\bar {o}_i}\) and \({\bar {p}_i}\) are the average of the actual and predicted outputs, and n is the number of samples.

To find out whether the obtained models overfit, it is suggested that the performance of the model must be checked on different divisions of data. In this regard, the aforementioned performance measure tools, R, RMSE, and MAE, may be used. Close R, MAE, and RMSE values between predicted and actual values on the training and testing divisions of data sets confirm that overfitting in model development is avoided [15]. Values of statistical criteria obtained by LGP-based models based on data classification are summarized in Table 4. According to the close calculated values for different divisions of data in Table 4, it can be concluded that overfitting is avoided. Besides, the performance of the models on the training and testing data demonstrates that they have both good predictive abilities and generalization performance.

Table 4 Calculated values of criteria for proposed LGP models based on data division

Furthermore, it is suggested that the arithmetic values of mean (μ) and standard deviation (σ) of ratios predicted versus actual values (p/a) can be regarded significant indicators of the accuracy and precision of the estimation models. Prediction via a more accurate model results in μ(p/a) closer to 1 and σ(p/a) closer to 0, respectively, in comparison with other models [39].

A comparison between the existing models in the literature and obtained LGP models could indicate the level of accuracy or capability performance of the presented models.

There are several statistical indices proposed in the literature to evaluate the model performance. Herein, with regard to the aforementioned issues, the overall statistical performance results of models of \(f_{{{\text{cc}}}}^{\prime }\) (MPa) and corresponding εcc (%) are summarized in Tables 5 and 6, respectively.

Table 5 Evaluation of performance of different f’cc prediction models considered in this study
Table 6 Evaluation of performance of different εcc prediction models considered in this study

It can be observed in Tables 5 and 6 that the values of R, the correlation coefficient index, for different models are close. Although the ANN model is ranked first in those tables, as was expressed, it is a black-box model which cannot be used as a formula to for hand calculation purposes. This demonstrates the superiority of LGP model over the ANN approach. In case the ANN model can be converted to a tractable equation, the structure of the model will be too complex, considering the number of biases and weights. In this regard, LGP can also generate higher precision models using more instructions in its programs length; albeit, the objective of this study was to propose new simple design equations. It also should be considered that R gives a general viewpoint for the evaluation of the accuracy of the models and only this factor cannot be used to assess the degree of accuracy of models. The differences among models would be more as other indices are considered such as RMSE and MAE which represent an average error between predicted and experimental values. To explain this, another way of observing the systematic mismatches between the predictions and observations can be considered and that is the detection of the residual error (RE) between predicted and actual values for each datum. Figure 4 demonstrates a plot of the RE values obtained by proposed LGP models and those provided by traditional models. The maximum RE values made by each model is also indicated in Tables 5 and 6. The indicated statistical analysis used for validation and verification of the LGP-based models can give the user an idea about the amount of reliability, accuracy or the factor of safety when using the model or selecting the best model among different models.

Fig. 4
figure 4

Comparison of RE values made by different models for all data sets a \(f_{{{\text{cc}}}}^{\prime }\), b εcc

The RE values in Fig. 4 are calculated as predicted minus actual values. An important issue which may be realized from Fig. 4 is the fact that predicted \(f_{{{\text{cc}}}}^{\prime }\) values made by Hoshikuma et al. [6], Mander et al. [1], and Mander et al. [2] models are more than actual \(f_{{{\text{cc}}}}^{\prime }\) values which means those models overestimate the output, whereas that of Sakai [8] underestimate the values in many cases. Similarly, the results of RE plot represent that εcc values are overestimated while using Mander’s model. These consequences could not be realized through statistical indices which demonstrate the necessity of evaluating errors for models before using them.

Nevertheless, it can be also observed from Tables 5 and 6 that LGP-Based models are ranked after the ANN’s model. In this regard, it is noteworthy that although prediction tools such as ANN are capable in predicting target values, they require the structure of the network to be identified in advance. This is usually done through a time-consuming trial and error procedure. Besides, these approaches are black-box tools. That is, such AI tools are not successful in generating practical prediction equations or if they are, the obtained model is highly complex and relies on several factors which often do not have the worth to be used. The main advantage of LGP over AI-based approaches is that it generates a transparent and structured representation of the system being studied via producing a practical model that can be easily utilized for hand calculating domains. Considering the values represented in Tables 5 and 6, and the results differences between predicted and actual values illustrated in Fig. 4, it can be concluded that LGP outperforms other models and methods in terms of accuracy and precision.

5.2 Parametric analysis

It is true that \(f_{{{\text{cc}}}}^{\prime }\) must increase with the increase of \(f_{{\text{c}}}^{\prime }\); in addition, εcc must increase with increase of ρs as expressed by several experimental studies by various researchers. In this regard, the pertinent question is whether the models responses conform to the experimental study results.

In the present study, the results in the aforementioned studies on performance capability of the models represent that the LGP models are able to satisfy the required validation conditions as the models outperformed well-known models in the literature. However, the model response and behavior to different input variables must be examined to assess the obtained model from engineering viewpoints. Herein, to approve the results, they should be compared to the results that experimentally or theoretically provided in the literature. For this purpose, a comparative parametric analysis is performed in this study as recommended by several researchers, e.g., Sakai and Sheikh [5], Saatcioglu and Razavi [4], Mander et al. [1], Mander et al. [2], Hoshikuma et al. [6], and Oreta and Kawashima [7]. The parametric analysis represents the response of the LGP-based \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc models to variations of input variables. The methodology is based on changing only one predictor variable at a time, while the other variables are kept constant at their specified mean values. For this purpose, a set of synthetic data is generated for each variable regarding to its range in the database. These values are presented to the prediction models and \(f_{{{\text{cc}}}}^{\prime }\) and εcc are calculated. This procedure is repeated using another variable until the model response is obtained for all of the predictor variables [18, 19].

Figure 5 represents the responses of the \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc prediction models to the variations of the affecting variables, i.e., \(f_{{\text{c}}}^{\prime }\) (MPa), \(f_{{{\text{co}}}}^{\prime }\) (MPa), d (mm), H (mm), fyh (MPa), ρs (%), s (mm), and ρcc (%). To confirm the accuracy of tendency of LGP models, the procedure of parametric study is comparatively conducted for different models in the literature. As can be seen, \(f_{{{\text{cc}}}}^{\prime }\) values continuously increase with increasing \(f_{{\text{c}}}^{\prime }\) which can be regarded as similar as \(f_{{{\text{co}}}}^{\prime }\), d, fyh, ρs, and ρcc and traditional models tend to respond similarly, whereas they decreases with increase of H and s. Besides, εcc values grow with the increase of d, fyh and ρs, while they decline when \(f_{{\text{c}}}^{\prime }\), \(f_{{{\text{co}}}}^{\prime }\), ρcc, H, and s increase.

Fig. 5
figure 5figure 5figure 5

Comparative parametric studies of various \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc models for circular confined RC columns

These trends highlight that the proposed models are efficient in accurately predicting the \(f_{{{\text{cc}}}}^{\prime }\) and corresponding εcc of RC columns in terms of physical behavior. Particularly, comparing the results with other methods which are basically obtained via experimental studies further verifies the accurate LGP models behavior to each variable from engineering viewpoints.

5.3 Sensitivity analysis

Another important concern involving the analysis of models is to find out how a model output is affected by changes in input values. For this aim, a sensitivity analysis (SA) may be performed. The results of SA can represent how different values of predictors will affect a dependent variable. In other words, it may be realized that parameters with greater value of the sensitivity coefficient have stronger influence than other variables on variation of the model output.

For each input process parameter, the SA percentage is calculated using the following formula [17]. The percent of each obtained output difference for each variable is computed which is often referred to as the sensitivity index (SI). Herein, the SI (%) value can be calculated as follows:

$${N_i}={f_{\hbox{max} }}({x_i}) - {f_{\hbox{min} }}({x_i})$$
(17)
$${\text{S}}{{\text{I}}_i}(\% )=\frac{{N{}_{i}}}{{\sum\nolimits_{{j=1}}^{n} {{N_i}} }} \times 100$$
(18)

where fmax (xi) and fmin (xi) are maximum and minimum of the predicted output by LGP model for the ith input variable, where other input values remain at their mean values and n is the number of involved variables in the model.

The obtained SI value demonstrates the relative importance of each variable in model variation. This procedure is performed for all variables involved in LGP models and the results are represented in Figs. 6 and 7.

Fig. 6
figure 6

Relative importance of the each variable in obtained LGP-based model for prediction of \(f_{{{\text{cc}}}}^{\prime }\) circular confined RC columns

Fig. 7
figure 7

Contributions of the each variable in obtained LGP-based models for prediction of εcc circular confined RC columns

These figures indicate that the LGP model of \(f_{{{\text{cc}}}}^{\prime }\) is more sensitive to \(f_{{\text{c}}}^{\prime }\), ρs, and ρcc as compared to other variables, while that of εcc is basically relies on the variation of ρs, H, ρcc and s.

Moreover, according to Fig. 5 which represents the results of the comparative parametric study, the relative importance of each input variable can also be figured out for each model. In other words, the greater the variation of each model for each variable in its range, the more significant the involved variable affects the model. Besides, the significance of each variable obtained by each model can also be compared to those provided by other models. As a notable example, consider Fig. 5 (l), as ρs increases from 0.28 to 2.5, its range in the employed database, εcc value calculated by the LGP model varies from 0.33 to around 0.92 whereas that of Mander’s model increases from almost 0.44 to around 1.52 which means that the variation of εcc is more in Mander’s model in comparison with LGP model. This also means that the variation of ρs further affects the Mander’s εcc model in comparison with other models. Similarly, the results can be extended for other variables and other models.

Moreover, it should be noted that results of a sensitivity analysis should be taken to account while using a formula. This is a serious issue which should be considered while varying a variable and calculating the amount of a variable via formulas for design purposes. In this regard, each model or formula gives distinct result; however, in the present study, all represented models give correct results and the main distinction is the amount of accuracy in their prediction performance. However, these results may be different for other proposed models. Finally, it is noteworthy that considering the range of the data used for model calibration and also the formulation in addition to the results of parametric study and SA definitely give the user a great comprehension about the design and final conclusion.

Finally, comparing the obtained results of comparative parametric and sensitivity analyses approve that the LGP model can accurately and correctly predict \(f_{{{\text{cc}}}}^{\prime }\) and εcc in terms of physical behavior and engineering viewpoints as the obtained results by LGP models conform to those provided by well-known experimentally obtained models in the literature.

6 Conclusion

The objective of the present paper was to examine potential of LGP as an alternative approach for modeling a considerable structural and civil engineering problem which is estimation of \(f_{{{\text{cc}}}}^{\prime }\) and εcc of RC columns. Reviewing the literature, many mathematical models have been proposed considering different common and uncommon parameters as independent input variables. In this regard, this study essentially resulted in deriving new robust models for prediction of those factors incorporating \(f_{{\text{c}}}^{\prime }\) (MPa), \(f_{{{\text{co}}}}^{\prime }\) (MPa), d (mm), H (mm), fyh (MPa), ρs (%), s (mm), and ρcc (%) as affecting input variables for prediction of \(f_{{{\text{cc}}}}^{\prime }\) and εcc of RC columns with circular cross section. Various analyses were conducted for evaluating the validation and generalization of the proposed models. The results mainly demonstrate that the proposed LGP models notably outperform the well-known models in the literature regarding the prediction capability. To assess the accuracy of the LGP models in terms of physical behavior and from engineering viewpoints, a comparative parametric analysis was carried out and the responses of obtained LGP models to input variables were compared to those obtained by other models. Besides, to find out how the proposed models outputs are affected by varying the inputs, a sensitivity analysis procedure was also performed. The results of comparative parametric and sensitivity analyses approved that LGP models can accurately predict \(f_{{{\text{cc}}}}^{\prime }\) and εcc of circular RC columns and have appropriately captured the input/output non-linear relationships from engineering viewpoints as were conformed to those experimentally or theoretically provided in the literature. It is noteworthy that the main superiority of LGP over other computational and artificial intelligence modeling approaches is the fact that it is able to generate practical models which can be used via hand calculations during design procedures. Besides, the results of this study also represented that LGP can be utilized as a good alternative approach for indirect estimation of \(f_{{{\text{cc}}}}^{\prime }\) and εcc of RC columns with circular cross section as the proposed models are able to predict those parameters with a high degree of accuracy. Furthermore, it should be considered that numerical models highly depend on the data used in their process of model development. The capability of these models is typically limited to the range of the data used for their calibration and generalization, and also the formulation depends on the available variables, in the database. To tackle this, the model can be retrained and improved to make more precise predictive models for a wider range of data and more variables. However, the authors believe that the proposed LGP models can be used for pre-design purposes.