Introduction

In today’s competitive market, composites have become one of the mostly used materials in manufacturing industries due to their certain unique advantageous structural and thermal properties. They provide high strength and specific stiffness, high damping, good resistance against corrosion, high fatigue strength, high volume-to-weight ratio, resistance to chemical and microbiological attacks, low thermal conductivity, low thermal expansion, etc. [1]. A wide variety of composites are used for specific purposes in automotive and aviation industries, marine bodies, storage containers, pipes and industrial floorings. In aerospace industries, glass fiber-reinforced plastics are used in fairings, landing gear doors, storage room doors and passenger compartments. Machining processes are used generally to cut, drill and contour composite for building products. Hole making operations are essential for final assembly. It was reported that over 10,000 holes are made in small single engine aircraft while in large transport aircraft, millions of holes are made, most for fasteners—rivets, bolts and nuts which consume almost 50% of total airframe production cost in airframe assembly [2]. Therefore, precise machining needs to be performed to ensure dimensional and functional stability and to obtain a better productivity.

Composites are constituted of matrix and disperse phase. Based on different combination of phase materials, a wide variety of composites with various properties are developed to meet the different needs of customers. To get both elastic behavior and rigid structure of product, fiber-reinforced plastic composites are often preferred. Glass may be used as fiber component due to its high-temperature corrosion resistance, inertness to chemical and environmental attacks. Besides, extremely fine fibers of glass show high strength and stiffness compared to its own normal brittle behavior. Similarly, thermosetting plastics are preferable for matrix phase due to their exclusive thermal properties (permanently hard during their formation and do not soften even at high temperature opposed to thermoplastics) and high volume to weight ratio. Covalent cross-links anchor the network chains of thermosetting polymers together to resist vibrational and rotational chain motion at high temperatures which gives harder, stronger and dimensional stability than thermoplastics [3]. Epoxy and polyester are two commonly used thermosetting plastics. Less expensive polyester is chosen from economic point of view, and glass fiber-reinforced polyester (GFRP) becomes a widely used material in engineering application.

Though GFRP composites have a lot of advantageous material properties, it leads to too much extent of manufacturing difficulties. Chips formed during machining of GFRP causes a severe irritation to human skin. Drilling of GFRP is substantially different from that of metal and results in many undesirable effects such as rapid tool wear, rough surface finish and defective subsurface layers caused by cracks and delaminations. Thus, improper choice of cutting parameter combination may lead to deleterious effect on product performance.

Delamination is an inter-ply phenomenon associated with drilling operation causing reduction in structural integrity and poor assembly tolerance [4]. Induced delamination occurs both at the entrance and exit surface of workpiece rendering it to be less strong. Cutting force acting in the peripheral direction is the main cause of the peeled-up delamination (entrance surface), while the thrust force is the main source of the push out delamination (exit surface) [5]. At the beginning of drilling operation, thickness of the laminated composite material is stable to withstand the thrust force and as the tool approaches the exit plane, stiffness provided by the remaining plies may not be enough to bear the thrust force, causing the lamina to separate resulting in delamination. However, such delamination at exit side may be minimized by using back-up plates.

The delamination that occurs during drilling severely influences the mechanical characteristics of the material around the hole. A common measure of delamination factor (Fd) at the entrance of drilled hole is considered as

$$ {\text{Fd}} = \frac{{D_{ \rm max } }}{{D_{\text{o}} }} $$
(1)

where Dmax and Do indicate maximum diameter of damaged zone (Fig. 1) and nominal diameter of the hole, respectively. It is obvious that for better performance of product delamination factor should be minimized.

Fig. 1
figure 1

Scheme of measuring delamination factor

There have been significant contributions in experimental study on drilling characteristics of composites [6]. Khashaba [7] studied the influence of drilling and material variables on thrust force, torque and delamination of five different types E-glass fiber-reinforced thermosetting composites such as continuous winding with filler/polyester, cross winding/polyester, chopped/polyester, woven/polyester, woven/epoxy. Effect of tool geometry and tool material on thrust force and delamination in drilling of glass fiber-reinforced epoxy composites was investigated by Abrao et al. [8]. Influence of tool point geometry on thrust force and delamination in drilling of glass/phenolic-woven fabric composite using cemented carbide drills was studied by Velayudham and Krishnamurthy [9]. EI-Sonbaty et al. [2] suggested that higher cutting speed and fiber volume fraction improve surface roughness of drilled-hole wall of glass fiber-reinforced epoxy composites by conventional high-speed twist drill. Delamination also reduces with increase in cutting speed [10]. Davim et al. [11] used different matrix materials to investigate the cutting characteristics of fiber-reinforced plastic composite materials using cemented carbide drills (K10) with appropriate geometry. Behavior of different drill geometries during machining glass fiber-reinforced plastics was also investigated by Davim et al. [12]. Hocheng and Tsao [13] presented a comprehensive analysis of delamination employing various drill bits such as saw drill, candle stick drill, core drill and step drill.

Various techniques have been applied for modeling and optimization of process parameters to yield lower delamination. Mohan et al. [14] conducted delamination analysis in drilling process of glass fiber-reinforced plastic composite materials using Taguchi’s methodology. Palanikumar et al. [15] performed a study on drilling glass fiber-reinforced epoxy composites with two types of cutting tools namely a twist drill and a four flute cutter of high-speed steel using Taguchi’s experimental design method as well as response surface methodology. Multivariable linear regression analysis was applied to investigate the effect of the drilling parameters such as speed and feed on cutting forces, thrust and delamination in drilling chopped glass fiber-reinforced epoxy composites with different fiber volume fraction by Khashaba et al. [16]. Multi-layer feed forward ANN architecture trained using error-back propagation training algorithm was used for modeling of delamination in high-speed drilling of carbon fiber-reinforced composite material by Karnik et al. [17]. Stone and Krishnamurty [5] developed a thrust force controller to minimize delamination, and a relationship between feed rate and desired thrust force was modeled by ANN. Multilayered neural network with Levenberg–Marquardt learning algorithm was also employed for the purpose of modeling of drilling responses [18].

Taguchi method and Grey relation analysis were used for optimization of thrust force, surface roughness and delamination factor for GFRP composite by Palanikumar [19] and for carbon fiber-reinforced plastic by Hsu and Tsao [20]. Gaitonde and Karnik [21] developed ANN model for prediction of burr height and burr thickness during drilling of 25-mm-thick AISI 316L stainless steel. They considered multi-response S/N ratio as fitness function for optimization of the two burr size parameters at different drill diameters by PSO. Though their predictive models were tested with small error, the parameters of PSO should be set more accurately to ensure global optimum. Fixed value of cognitive and social acceleration coefficients, termination criteria set by maximum number of iteration value may not guarantee the simulation to reach global optimum. Combining two responses into a single fitness function with equal weight factors may destroy the individual nature of each response. With suitable parameter settings, PSO also requires less time and memory space (small number of function evaluation) to terminate the simulation process. As such, burr formation at the exit side in drilling can be minimized by using back-up materials.

Multivariable regression analysis, response surface methodology and artificial neural network are thus the three main data-based procedures applied for modeling mechanical drilling process of composites. Compared to artificial neural network, support vector machine (a powerful learning system) is devoid of the four problems of efficiency of training, efficiency of testing, over-fitting and algorithm parameter tuning [22]. Besides, the insensitive zone of SVM captures the small-scale random fluctuations appeared in stochastic-type responses which is beneficial for other researchers to apply the models on different products obtained in different batches. An analytic fuzzy logic classification (AFC)-SVM model for tool wear estimation with high accuracy was built by Brezak et al. [23], and Benkedjouh et al. [24] estimated a SVM learned model for remaining useful life (RUL) of cutting tool. Though the models agree with experimental results with less error, no such clear description of the procedure to choose the internal parameters of SVM was reported. Li et al. [25] used WEKA software for prediction of cell vernier through SVM in manufacturing of TFT-LCD. Their developed model gives better prediction than multiple-regression model, but neither the steps of model building nor the simulation procedure for optimal internal parameter settings was stated. Surface roughness in CNC turning of AISI 304 austenitic stainless steel was modeled with correlation coefficient through three SVM learning systems (LS-SVM, Spider SVM and SVM-KM) and ANN [26]. Parameters of SVM (C and σ) were set by grid search method. Though it is reported that for model development SVM learning systems consume less time than ANN, no such clear explanation about those specific choices of searching region of SVM parameters was stated. Also, the values of SVM internal parameters obtained through grid search method depend on the choice of jumping interval. However, such predictive model of delamination factor (Fd) in drilling of glass fiber-reinforced polyester composite (GFRP) in particular is not found yet in the literature. Therefore, modeling of response through SVM and optimization of the representative model by PSO are proposed in the present work.

In the present study, experimental results on drilling of glass fiber-reinforced polyester composite with different combinations of input parameters, namely material thickness (workpiece property), drill diameter (cutting tool parameter), spindle speed and feed rate, are taken from thesis work submitted by Behra [27]. Model is developed for delamination factor using structural risk minimization principle-based support vector machine (SVM) learning system. For the purpose of model development, optimum combination of three internal factors in SVM, namely regularization parameter (C), radius of loss insensitive hyper-tube (ε) and standard deviation of kernel function (σ), is searched by modified particle swarm optimization technique. Fitted model is tested through follow-up experiments. The representative model is further used for searching optimal setting of input parameters (material thickness, drill diameter, spindle speed and feed rate) for minimum delamination factor in drilling operation.

Support vector machine (SVM) [28]

Support vector machine, a supervised batch learning system, is firmly grounded in the framework of statistical learning theory. Vapnik [29] introduced structural risk minimization (SRM) principle instead of empirical risk minimization (ERM), implemented by most of the traditional artificial intelligence-based modeling technologies. Neural network approaches may have suffered with generalization, producing over fitted models but SRM minimizes upper bound on the expected risk, as opposed to ERM, that minimizes error on the training data. This difference equips SVM with a greater ability to generalize [30].

Ultimate goal in modeling of empirical data is to choose a model from hypothesis space, which is closest to the underlying target function. Suppose, a set of training data {(x1, y1), (x2, y2), …. (xN, yN)} is used for model developing in d-dimensional input space (i.e., xЄ Rd). Key assumption in model developing is that training and testing data set are disjoint, independent and identically distributed according to the unseen but fixed underlying function [22]. The linear target function may be represented in the form [31].

$$ {\text{f}}\left( \varvec{x} \right) \, = \, \left\langle {\varvec{w},\varvec{x}} \right\rangle + b $$
(2)

where < , > indicates dot product in vector space. If the input pattern does not hold any linear relation to output, (nonlinear SVM regression model is shown in Fig. 2), then they are mapped to feature space Φ(x) from high-dimensional input space via kernel functions. So, optimal choice of weight factor w and threshold b (bias term) is prerequisite of accurate modeling. Flatness of the model is controlled by minimizing Euclidean norm ||w||. Besides, empirical risk of training error should also be minimized [32]. So, regularized risk minimization problem for model developing can be written as follows

$$ R_{\text{reg}} \left( f \right) = \frac{{\left\| {\varvec{w}^{2} } \right\|}}{2} + C\mathop \sum \limits_{i = 1}^{N} L\left( {y_{i} , \, f\left( {{\text{x}}_{i} } \right)} \right) $$
(3)

Weight vector w and the bias term b can be estimated by optimizing this function (Eq. (3)), which minimizes not only empirical risks but also reduces generalization error, i.e., over-fitting of model simultaneously. Here, L(y), a loss function is introduced to penalize over-fitting of model with training points. A number of loss functions are already developed for handling different types of problems [30]. ε-insensitive loss function (Fig. 3) is mostly used for process modeling problems. This function may be defined as

$$ \begin{array}{*{20}c} {L\left( {y_{i} , \, f\left( {x_{i} } \right)} \right) = \left| {y_{{i,{\text{experimental}}}} {-} \, f\left( {x_{i} } \right)} \right| \, - \, \varepsilon ,} & {{\text{if}}\; \, \left| {y_{{i,{\text{experimental}}}} {-}f\left( {x_{i} } \right)} \right| \ge \varepsilon } \\ { = \, 0,} & {{\text{if}}\;\left| {y_{{i,{\text{experimental}}}} {-}f\left( {x_{i} } \right)} \right| < \varepsilon } \\ \end{array} $$
(4)

Here, points inside the ε-tube are considered as zero loss; otherwise, a penalization is calculated by introducing C, which is a trade-off between flatness and complexity of the model. Practical significance of this insensitive zone is that the points inside the hyper-tube, i.e., close enough to estimated model are deemed to be well estimated and those outside the tube contribute training error loss. The outsiders of the insensitive zone belong to support vector group. So, size of ε-insensitive zone controls number of support vectors. As radius of insensitive hyper-tube increases, number of support vector reduces and flexibility of the model diminishes. This behavior may be advantageous for eliminating the effect of small random noise in output, but larger value of ε will not completely extract the unseen target function. Besides, higher value of C makes the model more complex with the chance of over-fitting, but too small value may increase training errors. So, optimum choice of this regularization parameter is necessary for better modeling. Two positive slack variables ξi and ξi* are introduced [29, 31] to cope with infeasible constraints of the optimization problem. Hence, the constrained problem can be reformulated as

$$ \begin{aligned} {\text{minimize}}{:} & \quad \frac{{\left\| \varvec{w} \right\|^{2} }}{2} + \, C\;\mathop \sum \limits_{i = 1}^{N} \left( {\xi_{i} + \xi_{i}^{*} } \right) \\ {\text{subject}}\;{\text{to}}{:}& \quad y_{i,\exp } - \left\langle {\varvec{w},\varvec{x}} \right\rangle_{i} - \, b \le \varepsilon + \xi_{i} \\ & \left\langle {\varvec{w},\varvec{x}} \right\rangle_{i} + b \, {-} \, y_{i,\exp } \le \, \varepsilon \, + \, \xi_{i}^{*} \\ & \xi_{i} , \, \xi_{i}^{*} \ge \, 0,\quad \varvec{i} = \, 1\left( 1 \right)N \\ \end{aligned} $$
(5)

This problem can be efficiently solved by standard dualization principle utilizing Lagrange multiplier. A dual set of variables are introduced for developing Lagrange function. It is found that this function has a saddle point with respect to both primal and dual variables at the solution. Lagrange function can be stated as

Fig. 2
figure 2

Nonlinear SVM regression model

Fig. 3
figure 3

ε-Insensitive loss function

$$ \begin{aligned} L & = \frac{{\left\| \varvec{w} \right\|^{2} }}{2} + C\mathop \sum \limits_{i = 1}^{N} \left( {\xi_{i} + \xi_{i}^{*} } \right) - \mathop \sum \limits_{i = 1}^{N} \left( {\eta_{i} \xi_{i} + \eta_{i}^{*} \xi_{i}^{*} } \right) - \mathop \sum \limits_{i = 1}^{N} \alpha_{i} \left( {\varepsilon + \xi_{i} - y_{i} + \left\langle {\varvec{w}, \varvec{x}} \right\rangle_{i} + b\xi_{i}^{*} } \right) \\ & \quad - \mathop \sum \limits_{i = 1}^{N} \alpha_{i}^{*} \left( {\varepsilon + \xi_{i}^{*} + y_{i} - \left\langle {\varvec{w}, \varvec{x}} \right\rangle_{i} - b} \right) \\ \end{aligned} $$
(6)

where L is the Lagrangian and ηi, ηi*, αi, αi* are Lagrange multipliers satisfying ηi, ηi*, αi, αi* ≥ 0. So, partial derivatives of L with respect to w, b, ξi, ξi* will give the estimates of w and b. The present problem is solved by using LibSVM MATLAB Toolbox.

Support vectors can be easily identified from the value of difference between Lagrange multipliers (αi, αi*). Very small values (close to zero) indicate the points inside the insensitive hyper-tube, but nonzero values belong to support vector group [33]. The w can be calculated by [31]

$$ \varvec{w} = \mathop \sum \limits_{i = 1}^{N} \left( {\alpha_{i} - \alpha_{i}^{*} } \right) \varPhi \left( {x_{i} } \right) $$
(7)

The idea of kernel function K(xi, x) gives a way of addressing the curse of dimensionality [30]. It helps to enable the operations to be performed in the feature space rather than potentially high-dimensional input space. A number of kernel functions satisfying Mercer’s condition were suggested by researchers [33, 34]. Each of the functions has its own specialized applicability. Gaussian radial basis function with σ standard deviation [given in Eq. (8)] is commonly used for its better potentiality to handle higher-dimensional input space.

$$ \left. {K\left( {x_{i} ,\varvec{x}} \right) \, = e^{{ - \frac{{\left\| {x_{i} - x^{2} } \right\|}}{{2\sigma^{2} }}}} } \right|_{{\sigma_{\text{optimum}} }} $$
(8)

Thus, the final model with optimum choice of C, ε and σ may be presented as [31]

$$ f\left( x \right) \, = \mathop \sum \limits_{i = 1}^{N} \left( {\alpha_{i} - \alpha_{i}^{*} } \right)K \, \left( {x_{i} ,x} \right) \, + \, b_{\left|{{\begin{array}{*{20}c} {C_{\text{optimum}} } \\ {\varepsilon_{\text{optimum}} } \\ {\sigma_{\text{optimum}} } \\ \end{array} }}\right.} $$
(9)

Particle swarm optimization (PSO) [28]

Particle swarm optimization (PSO) technique is one of the most advanced evolutionary computational intelligence-based optimization methodologies for optimizing real-world multimodal problems. PSO mimics natural behavior found in flock of birds or school of fish seeking their best food sources [35]. In this population based swarm intelligence technique, a set of randomly initialized particles (swarm) are always updated in position and velocity by gathering information from themselves. Effect of each particle as well as the whole swarm’s experience modifies position of the population forwarding to optimum zone. Rate of convergence is purposefully controlled by different factors. Position of global optimum is not affected by the choice of the factors, but convergence is delayed due to improper choice or may lead to entrapping in local optima. For multivariable problem in high-dimensional space, time and memory space needed for reaching optimum solution by PSO are very important.

Number of particles (n) in swarm should be within the range (10, 40) [36]. Lower choice may not gather information from whole space, but higher value of n will take longer time to converge in optimum zone.

Inertia factor (ω) controls the effect of previous velocity of individual particle on current velocity. To modify the rate of convergence another control on simulation was done by introducing constriction factor (Ψ) [37]. This term bounds the velocity effect of particles on their position avoiding clamping of particles to one end of search space [38]. So, higher values of inertia and constriction factor ensure wide searching which is necessary at initial stage but gradual convergence is enhanced at moderately lower value.

Another two important factors are cognitive acceleration coefficient (cog) and social acceleration coefficient (soc) which greatly control the influence of individual’s and whole swarm’s experience, respectively, on particle’s new velocity. Individual best (p ibest ) experience of each particle favors good exploration in the search space but the best position (gbest) of swarm always guides to converge near optimum zone. So, choice of the factors becomes important for converging to global optimum zone quickly avoiding premature entrapping in local optima.

Several researchers use different values of the control factors for their different type of problem definitions. However, in general, for most of the cases, nearly a same range is suggested irrespective of the nature of problem [39]. Shi and Eberhart [40] suggested linearly decreasing inertia factor from 0.9 to 0.4. Cognitive acceleration coefficient should vary linearly with iterations from 2.5 to 0.5 while the variation of social acceleration coefficient would occur just in reverse order [41]. Since constriction factor directly controls the optimization time, it may be considered as linearly time varying from 0.9 to 0.4.

Further, maximum number of iterations is to be set properly. A large value is necessary for adequate convergence. In other words, simulation will be terminated before reaching this limiting value.

Experiment [27]

Experiment was conducted on a radial drilling machine (Make: The American Tool Works Co., USA) [27]. The machine is equipped with a maximum spindle speed of 1500 rpm and variable feed from 0.1 to 0.625 mm/rev with a 5.5 kW drive motor. Maximum radial dimension of workpiece is 1475 mm.

In the present study, material thickness (mt), drill diameter (dd), spindle speed (ss) and feed rate (fr) are considered as input parameters and delamination factor (Fd) measured on entrance surface as response parameter for drilling process. Based on the availability of the machine setting, levels of the input parameters are selected and presented in Table 1.

Table 1 Parameters and their levels

GFRP composite used in the experiment was supplied by VMT Glass Fiber Roofing Industries. Hand lay-up technique is used for producing the composite. Chopped strand mat E-glass fiber is used as reinforcement in polyester resin to prepare the laminate. Mechanical and physical properties of this composite are shown in Table 2. Experiments are carried out on GFRP samples of size 150 mm × 150 mm using wood as backing material.

Table 2 Properties of GFRP composite

HSS taper shank twist drills (Make: Addison & Co. Ltd., India) confirming to IS: 5103-1969/ISO: 235-1980/DIN: 345-1978/BS: 328: Part: 1-1986 specifications were utilized for the drilling experiments. Chemical composition of HSS tool material is listed in Table 3. Three different diameter twist drills (shown in Table 4) of grade M2 were employed.

Table 3 Chemical composition of HSS tool material (wt%)
Table 4 Specifications of twist drills

Damage around the holes at the entrance was measured with a toolmakers’ microscope with 30× magnification (Make: Carl Zeiss Ltd.). Delamination factor (Fd) is calculated using Eq. (1).

Among 81 unique combinations of the levels of different input parameters, randomly 11 sets are kept aside for testing purposes and rest 70 data sets are used for training the learning process.

Analysis and discussion

Model for delamination factor (Fd) is fitted through learning process of SVM based on training data sets. Fitted model is tested through another 11 sets of data. Then the representative model is used for searching optimal combination of input parameters (material thickness, drill diameter, spindle speed and feed rate) for minimum delamination factor (Fd).

Model development

Significance of parameter selection in SVM for model development and thereby for good prediction is evident. Particle swarm optimization (PSO) could be used effectively for this purpose. The most influencing parameters in SVM learning process are C, ε and σ. With optimum choice of C, ε and σ, a set of Lagrange multipliers (αi, αi*) could be found and the model can be estimated using Eq. (9).

At first, search spaces for each of the parameters (C, ε and σ) are to be determined. A wide range of each parameter may be a good search space, but this will take large computation time. Estimation of C based on experimental outputs (target values) was suggested by Levis and Papageorgiou [34], Cherkassky and Ma [42].

$$ C \, = \, max \, \left( {\left| {{\bar{\text{y}}} + \, 3\sigma_{t} } \right|,\;\left| {{\bar{\text{y}}} - 3\sigma_{t} } \right|} \right) $$
(10)

where \( \bar{y} \) and σt represent mean and standard deviation of target values. However, this exact value may lead to erroneous modeling due to the presence of random noises. So, it is better to choose a range for C and searching operation is to be performed within the limits. At first, using Eq. (10) C value is calculated for Fd as CFd = 1.4160. Then, considering normal distribution with mean at this calculated C value and 50% of mean value as standard deviation, two points are chosen at both sides of estimated C value. Thus, range of C for Fd is determined (Table 5).

Table 5 Searching range of SVM parameters

Based on experiments, suitable range of ε and that of σ are also suggested in [34, 42]. Inside the bounds of σ, generalization performance is found as stable. Those limits can be calculated from

$$ \varepsilon \, = \left[ {\frac{{{\bar{\text{y}}}}}{30},\frac{{{\bar{\text{y}}}}}{10}} \right],\;\sigma \, = \left[ {0.1^{{\left( {\frac{1}{\text{z}}} \right)}} ,\;0.5^{{\left( {\frac{1}{\text{z}}} \right)}} } \right], $$
(11)

Here, z indicates number of most influencing attributes in the process. In this study of drilling process, it is four namely material thickness, drill diameter, spindle speed and feed rate. Using Eq. (11), ε and σ values for Fd are calculated (Table 5).

For better implementation of this estimated search range, it was suggested [34, 42] to normalize the training inputs within the range (0, 1). So, the input parameters of drilling process, i.e., material thickness (mt), drill diameter (dd), spindle speed (ss) and feed rate (fr) are normalized using the following formulae.

$$ x_{{1,{\text{norm}}}} = \frac{{{\text{mt}} - 8}}{16 - 8},\quad x_{{2,{\text{norm}}}} = \frac{{{\text{dd}} - 10}}{14 - 10},\quad x_{{3,{\text{norm}}}} = \frac{{{\text{ss}} - 400}}{1100 - 400},\quad x_{{4,{\text{norm}}}} = \frac{{{\text{fr}} - 0.100}}{0.275 - 0.100} $$
(12)

As SVM reduces the chance of generalization error; in the present study, parameters of SVM are selected through minimization of mean absolute percentage error (MAPE) in estimation of training outputs. For the purpose of optimization, MAPE being a function of C, ε and σ, is considered as the objective function. At every step during marching procedure, different combinations of C, ε and σ will develop different regression models. For each of the models developed through the change of C, ε and σ values, MAPE [Eq. (13)] is calculated and searching of minimum MAPE is done among the models using the procedure of particle swarm optimization.

$$ {\text{MAPE}} \left( \% \right) = \frac{100}{\text{N}}\mathop \sum \limits_{{{\text{i}} = 1}}^{\text{N}} \left| {\frac{{{\text{y}}_{{{\text{i}},{\text{experimental}}}} - {\text{y}}_{{{\text{i}},{\text{estimated}}}} }}{{{\text{y}}_{{{\text{i}},{\text{experimental}}}} }}} \right| $$
(13)

In most of the papers, termination criterion is set by a predefined maximum number of iterations. This may not ensure global optima. Rather, a statistical measure of relative dispersion, coefficient of variation (CV) of the position of particles in search space, is expected to help avoiding the premature stopping of simulation procedure. Coefficient of variation is calculated as a ratio of standard deviation to mean and expressed in percentage.

$$ {\text{Coefficient}}\;{\text{of}}\;{\text{variation}} \left( {\text{CV}} \right)\left( \% \right) = 100 \times \frac{{{\text{standard }}\;{\text{deviation}}\;{\text{of}}\;{\text{current}}\;{\text{position}}\;{\text{vector}}}}{{{\text{mean}}\;{\text{of}}\;{\text{current}}\;{\text{position}}\;{\text{vector}}}} $$
(14)

A very small value of this criterion indicates a set of densely compact particles in the swarm, irrespective of their search range in different dimensions. Small coefficient of variation in conjunction with negligible change in the values of MAPE will ensure global optimum. Besides, nature of change of different components of CV values in different dimensions may give a preliminary idea about the underlying relationship among output and inputs. Therefore, above proposed modification is introduced and modified PSO algorithm for searching optimum combination of C, ε and σ for minimum MAPE is described as follows.

  • Step 1 Choose number of particles in swarm (n), maximum number of iterations iter max, accuracy level (acc). Set, inertia factor range (ω initial, ω final), cognitive acceleration coefficient range (cog initial, cog final), social acceleration coefficient range (soc initial, soc final) and constriction factor range (Ψ initial, Ψ final).

  • Step 2 Set iter = 1. Randomly initialize the position of n particles in swarm that is n set of initial combination of C, ε and σ. Also randomly set initial velocity of each particle using following relation.

    $$ v_{\text{iter}}^{{({\text{i}})}} = {\text{range}}^{{({\text{i}})}} \times {\text{ rand}},\quad i = 1\left( 1 \right)3 $$
    (15)

    where “range(i)” and “rand” indicate range of C, ε and σ (shown in Table 5) and a random number within the range (0,1), respectively. So, n set of velocity vectors with components along C, ε and σ coordinates are initialized.

  • Step 3 Set t = 1 and p tbest and gbest as current position vector.

  • Step 4 Check whether t = n or not. If yes go to step 8, otherwise set t = t + 1 and go to step 5.

  • Step 5 Calculate MAPE (objective function value) at tth particle position in swarm using Eq. (13). If iter = 1, then set p tbest as current position vector and go to step 7. Otherwise, go to step 6.

  • Step 6 If this MAPE value is smaller (better) than MAPE at already set p tbest , then update p tbest vector. Otherwise, p tbest is kept unaltered.

  • Step 7 Check whether this current MAPE value is lesser (better) than MAPE at already set gbest. If yes, update gbest vector to current position vector and go to step 4. Otherwise, gbest vector is kept unaltered and go to step 4.

  • Step 8 Calculate coefficient of variation of particles’ position in current swarm using Eq. (14). Each position vector has three components along C, ε and σ. So, three values of coefficient of variation will be found.

  • Step 9 If all the three coefficient of variation values simultaneously become less than accuracy level, then terminate the simulation process and go to step 14, otherwise set k = 1 and go to step 10.

  • Step 10 Considering linear gradient of ω, cog, soc and Ψ, set the value of the control factors for this iteration (iter) using the following relations.

    $$ \begin{aligned} \omega_{\text{iter}} & = \omega_{\text{initial}} + \frac{{\left( {\omega_{\text{final}} - \omega_{\text{initial}} } \right)}}{{\left( {{\text{iter}}_{\text{max} } - 1} \right)}}\left( {{\text{iter}} - 1} \right) \\ {\text{cog}}_{\text{iter}} & = {\text{cog}}_{\text{initial}} + \frac{{\left( {{\text{cog}}_{\text{final}} - {\text{cog}}_{\text{initial}} } \right)}}{{\left( {{\text{iter}}_{\text{max} } - 1} \right)}}\left( {{\text{iter}} - 1} \right) \\ {\text{soc}}_{\text{iter}} & = {\text{soc}}_{\text{initial}} + \frac{{\left( {{\text{soc}}_{\text{final}} - {\text{soc}}_{\text{initial}} } \right)}}{{\left( {{\text{iter}}_{\text{max} } - 1} \right)}}\left( {{\text{iter}} - 1} \right) \\ \psi_{\text{iter}} & = \psi_{\text{initial}} + \frac{{\left( {\psi_{\text{final}} - \psi_{\text{initial}} } \right)}}{{\left( {{\text{iter}}_{\text{max} } - 1} \right)}}\left( {{\text{iter}} - 1} \right) \\ \end{aligned} $$
    (16)
  • Step 11 Update kth particle’s velocity vector and its velocity corrected position vector using the following formulae.

    $$ \begin{aligned} v_{{\left( {\text{m}} \right){\text{iter}} + 1}}^{k} & = \omega_{\text{iter}} \left( {v_{{\left( {\text{m}} \right){\text{iter}}}}^{k} } \right) \, + \, c_{{1,{\text{iter}}}} \left( {\text{rand}} \right) \, \left( {p_{{\left( {\text{m}} \right){\text{best}}}}^{\text{k}} {-}x^{\text{k}}_{{\left( {\text{m}} \right){\text{iter}}}} } \right) + c_{{2,{\text{iter}}}} \left( {\text{rand}} \right) \, \left( {g_{{\left( {\text{m}} \right){\text{best}}}} {-} \, x_{{\left( {\text{m}} \right){\text{iter}}}}^{k} } \right) \\ x_{{\left( {\text{m}} \right){\text{iter}} + 1}}^{k} & = x_{{\left( {\text{m}} \right){\text{iter}}}}^{k} + \, \varPsi \, \left( {v^{k}_{{\left( {\text{m}} \right){\text{iter}} + 1}} } \right),\;m = 1\left( 1 \right)3 \\ \end{aligned} $$
    (17)
  • Step 12 If new positions of particles (C, ε and σ) go outside the specified search range (shown in Table 2), then reset its current position to the violating limit. If k = n, then go to step 13. Otherwise set k = k + 1 and go to step 11.

  • Step 13 If iter = itermax, then go to step 1 and restart the simulation with larger itermax, such that termination criterion will be satisfied before reaching that new limit. Otherwise set iter = iter + 1 and t = 1, go to step 5.

  • Step 14 Set current gbest of the swarm as optimum point, i.e., components of latest gbest vector become the optimum combination of C, ε and σ for minimum MAPE in fitting the regression model on training data.

Therefore, final gbest vector satisfying the termination criterion becomes global optimum in that specified search space. The flow chart for searching optimum parameter combination (C, ε and σ) of SVM through modified PSO by minimizing MAPE of training data is also shown in Fig. 4.

Fig. 4
figure 4

Flow chart for searching optimum combination of C, ε and σ through modified PSO by minimizing MAPE of training data

In the present work, different control factors of modified PSO are set as, number of particles in swarm (n) = 20, ωinitial = 0.9, ωfinal = 0.4, coginitial = 2.5, cogfinal = 0.5, socinitial = 0.5, socfinal = 2.5, Ψinitial = 0.9, Ψfinal = 0.4, itermax = 250, accuracy level (acc) = 1.0 (i.e., simulation will stop when each component of coefficient of variance of particles’ position goes down below 1%).

As PSO is a population-based searching technique; a proper choice of initial positions vectors of the particles within specified search space and their corresponding velocities are to be set randomly. But wide spreading of initial position vectors must be assured by the operator for better searching. The set of values should be memorized by the computing machine for exact repetitiveness of the simulation process.

Different combinations of C, ε and σ change the shape of model, and best fitted model will have minimum fitting error that is minimum MAPE. As such the MAPE of the fitted model should be minimized. In this study, best fitted model with minimum MAPE is constructed on optimum combination of C, ε and σ that is the best position of the particles in swarm.

Thus optimizing MAPE by modified PSO with the initial position and velocity vectors, a set of C, ε and σ for Fd is found (shown in Table 6). Marching steps of the simulation process for minimizing MAPE are shown in Fig. 5. Though MAPE values remain unchanged after some iteration, yet the simulation is continued further. In this work, termination criterion is set by CV of position vectors’ components along each dimension (C, ε and σ), as this will ensure proper convergence of the searching operation to optimum zone. So, iteration process is continued until coefficient of variation of position vector components in each dimension (C, ε and σ) goes down below 1% (refer Fig. 6) although MAPE remains unchanged. Coefficient of variation (expressed in %) of position vectors in different dimensions is calculated by taking the ratio of standard deviation to mean of position vector component along respective dimensions.

Table 6 Optimum SVM parameters and training error measurement
Fig. 5
figure 5

Marching steps for optimization of MAPE for modeling of Fd

Fig. 6
figure 6

Change of CV with iterations for optimization of MAPE for modeling of Fd

Figure 6 indicates that effect of C is lower than ε and σ as CV for C dies out at a slower rate. This may conclude that model of Fd is not too much sensitive to the value of C. But rapid die down nature of ε and σ gives a strong evidence in favor of the presence of random noises in the responses.

Optimum C and ε values are found near the upper and lower bound of search space, respectively. Higher value of C (with respect to specified search space) signifies that the model of Fd is highly complex in nature with large random noises. This gives a good indication in favor of the stochastic nature of drilling composite material. The random noises in the output of this stochastic process are fairly controlled in developing a model by SVM with proper choice of ε. Process outputs within ε-tube do not greatly affect the complexity of model. Here lower value of ε for indicates that the estimated model capture small-scale randomness adequately. Besides, higher value of σ captures the oscillatory nature of training data which are not penalized by ε-insensitive zone. In the present study, optimum σ shifts toward upper end that is there is an oscillatory pattern outside the insensitive tube.

With the estimated optimum combination of C, ε and σ, Lagrange multipliers (αi, αi*) are calculated (shown in Table 9) and the developed model for Fd can be represented as

$$ \begin{aligned} {\text{Fd:}} & \quad f\left( x \right) \, = \mathop \sum \limits_{i = 1}^{N} \left( {\alpha_{i} - \alpha_{i}^{*} } \right)K \, \left( {x_{i} ,x} \right) + b_{\left|{{\begin{array}{*{20}c} {C \, = \, 2.1569} \\ {\varepsilon \, = \, 0.0409} \\ {\sigma \, = \, 0.8409} \\ \end{array} }}\right.} \\ & \quad \left. {{\text{with}}\;K\left( {x_{i} ,x} \right) \, = e^{{ - \frac{{x_{i} - x^{2} }}{{2\sigma^{2} }}}} } \right|_{\sigma \, = \, 0.8409} \\ \end{aligned} $$

To depict the effect of different input parameters (material thickness, drill diameter, spindle speed and feed rate) on delamination factor, surface plots are generated using the above estimated model. The representative surface plots are shown in Figs. 7, 8, 9, 10, 11 and 12. Feed rate is observed a significant factor.

Fig. 7
figure 7

Effect of material thickness & drill diameter on Fd at ss = 750 rpm & fr = 0.1875 mm/rev

Fig. 8
figure 8

Effect of material thickness & spindle speed on Fd at dd = 12 mm and fr = 0.1875 mm/rev

Fig. 9
figure 9

Effect of material thickness & feed rate on Fd at dd = 12 mm and ss = 750 rpm

Fig. 10
figure 10

Effect of drill diameter & spindle speed on Fd at mt = 12 mm and fr = 0.1875 mm/rev

Fig. 11
figure 11

Effect of drill diameter & feed rate on Fd at mt = 12 mm and ss = 750 rpm

Fig. 12
figure 12

Effect of spindle speed & feed rate on Fd at mt = 12 mm & dd = 12 mm

Testing of estimated model

Testing of the estimated model is performed with 11 disjoint data sets obtained from 11 separate follow up experimental runs conducted at the mid-level of feed rate. Testing results are shown in Table 7. The values of mean absolute percentage error (3.1356%) suggest the adequacy of the estimated model for application in practical field of work

Table 7 Testing for Fd

Optimization of Fd

Delamination of work surface always has a deleterious effect on quality of product performance. Thus, lower value of Fd is appreciable to the process engineers. Hence, optimum combination of input parameters (material thickness, drill diameter, spindle speed and feed rate) in drilling process to get minimum Fd in a certain working range is prerequisite in practical field of work. Here, modified PSO (as proposed in subsection 5.1) is applied on the developed SVM learned regression model [Eq. (18)] for searching optimum working condition satisfying minimum Fd within the pre-specified range (shown in Table 1). Control factors for modified PSO (n, ω, cog, soc, Ψ, iter max, acc) are kept same as before. Initial position vectors and corresponding velocity vectors are randomly chosen within the specified search range.

Marching steps of the simulation process for minimization of Fd are shown in Fig. 13. Optimum results are shown in Table 8.

Fig. 13
figure 13

Marching steps for minimization of Fd

Table 8 Optimum result and validation

Coefficient of variation specifically signifies the relative measurement of spreading of the particles in swarm. Though CV value is dependent on random numbers, but the nature of change of this value somehow reflects the physical relationship among parameters and responses. Figure 14 indicates that components of position vector along feed rate, material thickness and spindle speed die down rapidly (that is CV decreases at faster rate), but there is a gradual decaying nature in drill diameter component. Higher spindle speed and feed rate increase thrust force, which weakens the interplay bonding; as a result, separations of laminates occur. Also, thin material causes more push out delamination. So, change of drill diameter does not have too much influence on delamination, but other three parameters—material thickness, spindle speed and feed rate effectively control quality of drilled product.

Fig. 14
figure 14

Change of CV along different dimensions with iteration for minimization of Fd

Conclusions

Prediction of delamination factor of worksurface is very essential to fully diminish or decrease long-term product performance deterioration of glass fiber-reinforced polyester composites. Predictive model of the attribute is, therefore, prerequisite during the product development stage and freezing the design before final production. The glass fiber-reinforced polyester composite being an inhomogeneous and anisotropic material, small-scale variation in response is inevitable during drilling of this material. The developed SVM-based regression model with optimum parameter settings (C, ε and σ) can efficiently capture the small-scale random fluctuations and can predict delamination factor of work surface for the material in a robust way. Validation result of estimated model favors the practical use of the model in the chosen range. Optimum combination of input parameters in drilling for minimum delamination factor is further found by applying modified particle swarm optimization technique on the developed model. Thus, a procedure is proposed that may be used effectively to develop model for this type of materials and determine optimal setting of input parameters.