Introduction

Lead is one of the main strategic metals with significant applications in industries [1,2,3]. The annual lead production has experienced a growing trend recently, while the share of primary resources remained almost constant. In other words, the annual increase in lead production in recent years is due to the greater use of secondary sources. Therefore, finding new secondary sources for lead production and optimizing the recycling processes to attain higher efficiencies would play a key role in sustainable lead production in the future [4, 5].

Non-consumable anodes used in electrowinning and the liberator cells of electrorefining plants are usually made from lead-based alloys. Cold-rolled Pb anode containing 0.07–0.08 wt.% Ca and 1.35 wt.% Sn is the primary choice in the modern copper electrorefining plants [6]. The oxygen generated during the anodic reaction causes these lead anodes to corrode [7, 8]. Despite the significant progress in increasing the quality of these anodes, the corrosion of Pb–Ca–Sn anodes is one of the main problems of nearly all copper plants that use hydrometallurgical processes for copper production. This corrosion causes a large volume of sludge to form at the bottom of the liberator cells and copper electrowinning plant. This sludge contains the corrosion products of the Pb–Ca–Sn anodes. Since about twenty percent of the copper in the world is produced by hydrometallurgical methods [6], it can be imagined that a significant volume of lead-rich sludge is produced annually. Increasing environmental considerations and the feasibility of using this sludge as a secondary source for lead production encouraged factories to recycle this residual.

In Iran, several plants use hydrometallurgical processes to produce cathodic copper. These plants produce tens of tons of lead-rich sludge per year. Accumulation of such lead-containing wastes in the open environment causes environmental severe hazards. Therefore, an effort must be made to recover this spent lead sludge. There are various approaches based on pyrometallurgy or hydrometallurgy principles for the recovery of lead wastes. Unique characteristics, such as simplicity of the process, lack of production of harmful products, and notably better recovery efficiency, caused the pyrometallurgical approaches to be preferred over the hydrometallurgical ones [3]. However, relatively high-energy consumption is one of the disadvantages of the pyrometallurgical routes, which should be addressed by maximizing the lead recovery efficiency [3].

Recently, the support vector regression (SVR), which is a powerful technique in soft computing, was employed as a regression strategy in various categories such as pattern recognition [9], novelty detection [10], and image recognition [11, 12]. Although the SVR has become more widely employed to forecast time-series data and to reconstruct dynamically chaotic systems, a useful model can only be built if the parameters of the SVR are carefully determined. Compared to other nonlinear regression techniques such as neural network [13], the SVR illustrated superior generalization performance. However, the performance of the SVR strongly depends on the selection of kernel (i.e., responsible for creating a linear or near-linear regression procedure surface, estimating the input space to the feature space, or creating a linear or near-linear regression procedure surface in the feature space and penalty (i.e., a criterion for correcting a sample that has been split incorrectly) parameters.

The main objectives of this paper are:

  1. 1.

    recovery of Pb from the sludge of the liberator cells in a copper refinery which uses a hydrometallurgical process;

  2. 2.

    proposing a modified tunicate swarm algorithm (MTSA) strategy by the addition of adaptively updating algorithm parameters to keep the balance between exploitation and exploration searches and to preserve the swarm diversity;

  3. 3.

    strengthening of the MTSA strategy as well as avoiding its premature convergence;

  4. 4.

    using an advanced SVR-MTSA model for optimizing the SVR parameters to construct an efficient SVR model with the optimal generalization performance and regression accuracy;

  5. 5.

    optimization of the process parameters of the recycling method, including temperature, time, coke, Na2CO3, and Fe contents, based on the proposed SVR-MTSA model as a real case study to maximize the efficiency of lead recovery;

  6. 6.

    comparison of the effect of the selected parameters on the efficiency of lead recovery using sensitivity analysis.

Background

Selection of the process for lead waste recovery

Recovery of lead waste is carried out either by pyrometallurgical or hydrometallurgical processes [1, 3, 14,15,16,17]. Compared to the pyrometallurgical processes, the hydrometallurgical ones have some disadvantages, especially in the leaching step. The consumption of a significant amount of water, the complexity of the process, formation of hazardous waste solutions, and the need to decompose lead carbonate at high temperatures to achieve the metallic lead are among these disadvantages. Thus, pyrometallurgical approaches, especially those that allow desulfurization and reduction of various lead-containing phases at high temperatures, are preferred. Considering the available resources regarding the recycling of lead-containing wastes using pyrometallurgical processes, the effect of operational parameters can be summarized as follows:

Effect of temperature and processing time on the lead recovery

According to the literature [18], during the recycling of battery residue and lead sulfates, increasing the temperature up to 1000 °C enhanced the efficiency of lead recovery. Also, the higher temperature promoted the detachment of slag from the metal and also enhanced the matte formation. The processing time was the other parameter that affected the lead recovery. Reducing the processing time less than the optimal value decreased the lead recovery efficiency due to the lack of sufficient time for the reaction between the precursors. Prolonging the processing time increased the likelihood of lead oxidation and its entry to the slag phase, which significantly decreased the efficiency of lead recovery.

Effect of Na2CO3 content

If the waste contains sulfide phases of lead, the amount of sodium carbonate in the precursor affects the efficiency of lead recovery due to its role in the reduction of PbSO4 to Pb in reaction (1) [15]:

$${\text{PbSO}}_{{4({\text{s}})}} + {\text{Na}}_{2} {\text{CO}}_{{3({\text{s}})}} + {\text{C}}_{{({\text{s}})}} = {\text{Pb}}_{{({\text{l}})}} + {\text{Na}}_{2} {\text{SO}}_{{4({\text{l}})}} + {\text{CO}}_{{({\text{g}})}} + {\text{CO}}_{{2({\text{g}})}}$$
(1)

Effect of coke content

High carbon content in the precursor decreases the efficiency of lead recovery due to the increase of sulfur and lead content of the slag. From the thermodynamic point of view, carbon initially reduces the lead oxide according to the following stoichiometric reaction [15]:

$${\text{5PbO}} + 2.5{\text{C}} = 5{\text{Pb}} + 2.5{\text{CO}}_{2}$$
(2)

If the carbon content is higher than the above stoichiometric requirement, reaction (3) will start:

$${\text{PbSO}}_{4} + 2{\text{C}} = {\text{PbS}} + 2{\text{CO}}_{2}$$
(3)

In this case, the potential for the formation of lead sulfide, i.e., a decrease in the efficiency of lead recovery, is enhanced [18].

Effect of iron addition

As shown in reaction (3), the formation of PbS at about 650 °C is inevitable if the amount of carbon in the precursor is too high. In this case, Fe can reduce the lead sulfide to Pb according to reaction (4). This reaction is activated at about 900–1000 °C [2].

$${\text{PbS}} + {\text{Fe}} = {\text{Pb}} + {\text{FeS}}$$
(4)

As shown above, lead could be present in the process as various phases, including PbS, PbSO4, and PbO. There are many parameters, i.e., presence and amount of coke, iron and sodium carbonate as reducing agents, and also processing time and temperature, each of which can affect the final product of the process to some extent. Therefore, it is necessary to use an appropriate optimization algorithm to optimize the process parameters to achieve the highest recovery efficiency.

Support vector regression (SVR)

The support vector machine (SVM) algorithm is a popular machine learning approach that can also be used as a regression method. In other words, the support vector regression (SVR) considers the same features of SVM for classification and has only minor differences. For example, the output of the regression problem is a real number, and so estimating a value that has infinite possibilities is difficult. Nevertheless, the main principles are the same, and it tries to minimize error and individualize the hyperplane, which maximizes the margin. Due to the excellent generalization performance of SVR, it has been widely used in different fields such as face recognition, feature selection, and prediction [9,10,11,12,13].

SVR technique determines the suitable structure with some complexity to the available small size samples during the optimization. It shows a strong generalization ability and dimension independence, which makes it superior to traditional machine learning algorithms. In nonlinear regression, the SVR kernel function is necessary to project the input space into the feature space or create a linear or nearly linear regression hypersurface in the space of the features. The setting of the SVR parameter (i.e., penalty parameter C, the kernel function parameter δ, smoothness response parameter ε) has a great influence on the performance of the SVR [13].

In this paper, a novel approach based on the tunicate swarm algorithm (TSA) for parameter optimization of the SVR is proposed to improve the prediction accuracy of the model.

Tunicate swarm algorithm

Meta-heuristic optimization strategies have outstanding specifications including (i) employing easy concepts with easy implementation; (ii) no need for gradient information; (iii) ability to find a globally optimal, and (iv) adaptable for various problems such as engineering applications. Hence, these strategies are considered as good candidates for optimization of problems. In this paper, a new meta-heuristic optimization algorithm, namely, a tunicate swarm algorithm (TSA) is used since it can find better optimal solutions than other competitive strategies and is suitable for solving the real case engineering design problems [19, 20].

At sea, tunicates use two strategies to find the optimal state in their surroundings to find food, i.e., swarm intelligence and jet propulsion. A tunicate must meet three primary conditions when moving with a jet propulsion behavior: (i) should not conflict with the other tunicates in the search space, (ii) should choose the right path to the best search position, and (iii) should get as close as possible to the best search agent. Simultaneously, swarm behavior updates the position of other tunicates for the best optimal solution. In summary, the mathematical model for satisfying these conditions is explained as follows.

Avoiding conflicts among search agents

To prevent confrontation of tunicates with each other, a new search agent position is defined using \(\vec{A}\), \(\vec{G}\), and \(\vec{M}\) vectors as Eqs. (5), (6) and (7(:

$$\vec{A} = \frac{{\vec{G}}}{{\vec{M}}}$$
(5)
$$\vec{G} = c_{2} + c_{3} - \vec{F}$$
(6)
$$\vec{F} = 2 \cdot c_{1}$$
(7)

in which, \(\vec{G}\) indicates the gravity forces, \(\vec{M}\) is social forces, and \(\vec{F}\) is water flow advection in the deep ocean. c1, c2, and c3 are random values in the range of [0, 1]. \(\vec{M}\) can be estimated using Eq. (8).

$$\vec{M} = [P_{\min } + c_{1} \cdot (P_{\max } - P_{\min } )]$$
(8)

in which, Pmin and Pmax are the initial and subordinate speeds to construct the social interaction. Note that the values of Pmin and Pmax are considered as 1 and 4, respectively, by the authors in [20].

Moving to the best neighbor direction

The tunicates should move towards the best neighbor direction based on Eq. (9):

$$\vec{P}D = \left| {\vec{F}S - r_{and} \cdot \vec{P}_{p}^{{}} (x)} \right|$$
(9)

in which, the distance between the tunicate and the food source is defined by \(\vec{P}D\), x is the current iteration, the optimal food source position is determined by \(\vec{F}S\), the position of the tunicate is depicted by \(\vec{P}_{p}^{{}} (x)\), and rand is a random number in the range of [0, 1].

Converging towards the best search agent

The updated position of the tunicate in respect to the position of the food source, as the best search agent can be determined by Eq. (10):

$$\vec{P}_{p} (x) = \left[ \begin{gathered} \vec{F}S + \vec{A} \cdot \vec{P}D;\quad if:r_{and} \ge 0.5 \hfill \\ \vec{F}S - \vec{A} \cdot \vec{P}D;\quad if:r_{and} \prec 0.5 \hfill \\ \end{gathered} \right].$$
(10)

Swarm behavior

To simulate the behavior of tunicates, initially, the best solutions between the first two optimal are determined. Then, by considering the position of the best tunicate, the positions of other search agents are updated. The swarm behavior of the tunicates can be estimated by Eq. (11).

$$\vec{P}_{p} (\vec{x} + 1) = \frac{{\vec{P}_{p}^{{}} (x) + P_{p} (\vec{x} + 1)}}{{2 + c_{1} }}.$$
(11)

The proposed method

First, a modified tunicate swarm algorithm (MTSA) is introduced to enhance the global and local search ability of the standard TSA. Second, an SVR-MTSA model is developed to optimize the SVR parameters by MTSA approach, since the quality of the model built by the SVR [19] mostly depends on the careful tuning of its parameters. Finally, the optimal values of features from the lead recovery dataset are determined in such a way that the efficiency of lead recovery is maximized based on the SVR-MTSA.

Modified tunicate swarm algorithm (MTSA)

In the original TSA, \(\vec{A}\), \(\vec{G}\), and \(\vec{F}\) are the vectors that enable the search agents to scan the search space randomly without having any conflicts. Variation in these vectors provides the possibility of better exploration and exploitation phases. In respect to other evolutionary optimizers, the original form of TSA, especially when it has a higher dimension and complexity, can get trapped in local optimum solutions. Under this condition, finding the global optimal would face a severe challenge. In this research, the original TSA is improved in two aspects:

  1. (i)

    Self-adaptive parameters: Since the settings of the meta-heuristic algorithm must be adjusted based on each type of problem, the process becomes time-consuming and constitutes a significant burden on the user’s part. Therefore, the proposed algorithm adopts the controlling parameters to the circumstances of the tunicates at a specific moment.

  2. (ii)

    Improving exploration ability: The proposed algorithm changes the direction of the movement for tunicates so that they not only move towards the best search agent but also search for other directions to enhance their exploration phase.

Self-adaptive parameters

In the standard TSA, the value of parameter A (which affects the agents’ searching process) depends on several random variables (i.e., \(c_{1}\), \(c_{2}\), and \(c_{3}\)) and so with selecting inappropriate values for these parameters, the method cannot reach to a global optimum. Therefore, we pursue a way to remove these variables and modify parameter A to depend on a more reliable value. The details of the modification of parameter A is explained in detail as follows:

\(\mathop A\limits^{ \to }\) can be written as

$$\vec{A} = \frac{{c_{2} + c_{3} - \mathop F\limits^{ \to } }}{{P_{\min } + c_{1} (P_{\max } - P_{\min } )}},$$
(12)
$$\vec{A} = \frac{{c_{2} + c_{3} - (2c_{1} )}}{{P_{\min } + c_{1} (P_{\max } - P_{\min } )}} = \frac{{c_{2} + c_{3} - c_{1} }}{{P_{\min } + c_{1} (P_{\max } - P_{\min } )}}.$$
(13)

In the original form of TSA [20], the values of \(P_{\min }\) and \(P_{\max }\) are set to 1 and 4, respectively. \(c_{1} ,c_{2}\), and \(c_{3}\) are constants that are randomly distributed through [0, 1]. Therefore, assuming the interval of these random parameters \(\mathop F\limits^{ \to }\) in Eq. (5) and \(\mathop M\limits^{ \to }\) in Eq. (7) can be considered to change in [− 2, 2] and [1, 4] intervals, respectively. Also, \(\mathop A\limits^{ \to }\) vector affects both exploitation and exploration phases [20]. Hence, the variation of \(\mathop A\limits^{ \to }\) provides the possibility of the step jump of solutions through the search space (as shown in Eq. (5)). In the current study, values of the numerator and the denominator are adjusted high and low for the first iterations. This adjustment enabled the search agents to make long step jumps and explore a wider area of search space. Also, increasing the iteration number, caused \(\mathop A\limits^{ \to }\) to decrease to lower values (i.e., high value for the denominator and low value for the numerator). These modifications enhance the exploitation phase and local searches. The modified version of \(\mathop A\limits^{ \to }\) can be expressed as Eq. (14).

$$\vec{A} = \frac{{2 - 4(t/t_{\max } )}}{{4 - 3(t/t_{\max } )}},$$
(14)

in which, t and \(t_{\max }\) are the current number and the maximum number of iterations, respectively.

Improve exploration ability

As shown in Eq. (10), all tunicates (i.e., search agents) are moved towards the food source (i.e., best search agent). In other words, the tunicates are concentrated on the local search and so balancing between exploration and exploitation is necessary. To enhance the exploration phase in the current study, Eq. (14) was used to estimate \(\mathop A\limits^{ \to }\) value. If \(\mathop A\limits^{ \to }\) is higher than 1.25, then finding of the food source (i.e., exploration) is restricted to the neighborhood of the solution selected by the random strategy. Otherwise (i.e., \(\mathop A\limits^{ \to }\) is lower than 1.25), the neighborhood of the best solution so far is exploited. This process is continued to a preset maximum iteration. Figure 1 illustrates the flowchart of the proposed MTSA.

Fig. 1
figure 1

Flowchart of the proposed MTSA

Parameter optimization of SVR by MTSA

The performance of the SVR strongly depends on the adequate calibration of its parameters (i.e., penalty, C, kernel, δ, and smoothness response, ε). In the original SVR, these parameters are adjusted either by trial and error or by empirical experiments. However, the computational cost is increased exponentially according to the number of parameters and the number of sampling points for each parameter. Recently, meta-heuristic techniques (e.g., genetic algorithm (GA) and particle swarm optimization (PSO)) have been used to optimize the SVR parameters for better global search abilities against numerical optimization strategies. In this work, MTSA was applied for the selection of the SVM parameters. The population of the solutions (i.e., solutions for the SVR parameters) is generated by the MTSA and SVRs are built with each solution. Then, they are trained, and their parameters are evaluated in the test stage. In this research, we mainly concentrated on the parameter optimization of the SVR to minimize the prediction error and proposed an SVR-modified tunicate swarm algorithm (SVR-MTSA).

For generating the population of solutions, the search space dimension must be calculated. The search space of the problem for optimization of the SVR parameters has three dimensions (i.e., ε, C, and δ). For the ith tunicate, yi is the corresponding predicted value by \(x_{i} = (x_{i\varepsilon } ,x_{iC} ,x_{i\delta } )\) in the best position of xi. The solutions are coded with real values. For testing the created SVRs, a cross-validation technique is used, which is a popular technique for the estimation of the generalization error. The K-fold cross-validation is a kind of cross-validation that is a popular technique for the evaluation of model. The main idea of this method is that all observations participate in the testing process. In K-fold cross-validation, the dataset is divided into k parts and one part is applied for validation. Then, the remaining parts are considered for training. Figure 2 indicates the tenfold cross-validation that is used in this paper.

Fig. 2
figure 2

10-fold cross-validation process

In each iteration, SVR-MTSA tries to fit the training folds. Eventually, the final fitness is calculated based on the average of the values obtained by the validation sets. One of the main advantages of this technique over the usual division of the data set into two parts (i.e., test and train) is that the pessimistic bias is reduced by considering more training data in contrast to setting aside a relatively large part of the dataset as the test set. Then, the generalization error is determined as the mean of the test errors in the k experiments. In the current study, the dataset is divided into three folds, i.e., testing, validation (calculating the performance fitness of the model), and training. The correlation coefficient (R2, Eq. (15)), mean absolute percentage error (MAPE, Eq. (16)), mean bias error (MBE, Eq. (17)), and root mean square error (RMSE, Eq. (18)) are used as criteria for the evaluation of the investigated models. Generally, the closer to zero the MAPE, MBE, and RMSE values, the higher the accuracy of the models. A closer R2 to unity is more favorable if R2 is considered the criterion for the accuracy of the models [21].

$$R^{2} = 1 - \frac{{\sum\nolimits_{i = 1}^{n} {(y_{k} - f(x_{k} ,w))^{2} } }}{{\sum {(y_{k} - \overline{f}(x_{k} ,w))^{2} } }},$$
(15)
$${\text{MAPE}} = \frac{1}{n}\left[ {\sum\limits_{i = 1}^{n} {\frac{{(y_{k} - f(x_{k} ,w))}}{{f(x_{k} ,w)}}} } \right] \times 100,$$
(16)
$${\text{MBE}} = \frac{1}{n}\left[ {\sum\limits_{i = 1}^{n} {(y_{k} - f(x_{k} ,w))} } \right],$$
(17)
$${\text{RMSE}} = \left[ {\sum\limits_{i = 1}^{n} \frac{1}{n} (y_{k} - f(x_{k} ,w))^{2} } \right]^{\frac{1}{2}} .$$
(18)

The RMSE metric (Eq. (18)) can be used to define the performance index (i.e., fitness) of the models.

The fitness for the ith agent is calculated by Eq. (19), where \({\text{RMSE}}_{\max }\) indicates the highest RMSE value for the agents in the population with n agents.

$${\text{Fitness}}_{i} = \frac{{{\text{RMSE}}_{\max } - {\text{RMSE}}(I_{i} )}}{{\sum\nolimits_{j = 1}^{n} {({\text{RMSE}}_{\max } - {\text{RMSE}}(I_{j} ))} }}.$$
(19)

Figure 3 illustrates the flowchart of the SVR-MTSA.

Fig. 3
figure 3

Flowchart of the SVR-MTSA

At the end of the iterations, the optimized SVR, which achieved the lowest error (the best fitness), is considered as the SVR-MTSA.

Determination of the optimized values for the selected practical features

The recovery of lead from a secondary source is chosen as an engineering example for finding the optimal value for the lead recovery, based on the SVR-MTSA approach. The main steps of the optimization process with the SVR-MTSA approach is shown in Fig. 4.

Fig. 4
figure 4

General steps of the SVR-MTSA strategy

The optimization process is carried out in two main steps:

  1. (1)

    The best values for the SVR parameters (i.e., \(\varepsilon ,\,\,C\)) and the radial basis function kernel (i.e., \(\delta\)) is determined using the MTSA strategy. In this step, the empirical values are selected as input for temperature, processing time, and the amount of reducing agents (coke, Fe, and Na2CO3). The efficiency of Pb recovery, which should be maximized, is calculated as output;

  2. (2)

    The optimum values of the process parameters to attain the highest recovery efficiency is achieved by the obtained SVR-MTSA.

As was mentioned earlier, the MTSA is the enhanced version of the original form of the TSA strategy with two additional improvements (i.e., adoption of a self-adaptive method and an improved exploration phase). Each column of tunicates is analog to one parameter of the experimentally collected dataset, and hence the number of columns is equal to the number of recovery process parameters. A random number between the minimum and maximum values of the corresponding recovery process parameters in the selected dataset is used for the initialization of the tunicate. Then the evaluation of fitness for solutions is carried out by the proposed SVR-MTSA model. Figure 5 shows the steps carried out for the optimization of the parameters based on MTSA and SVR-MTSA.

Fig. 5
figure 5

Optimization of the parameters with MTSA and SVR-MTSA

The ability of the SVR-MTSA for the optimization of the efficiency of Pb recovery is also compared to those of back-propagation neural networks (BPNN), original SVR-cross-validation, SVR with basic TSA (SVR-TSA), SVR with grey wolf optimization algorithm (SVR-GWO), SVR with particle swarm optimization (SVR-PSO) and SVR with multi-verse optimizer (SVR-MVO) models.

Experimental method

Lead-rich waste of the liberator cells in the electrorefining plant of Khatoonabad copper refinery in Shahrebabak, Iran, was selected as the secondary source for lead recovery. The identity of the phases in this waste was determined using XRD (Philips, X’pert-MPD system by Cu-Kα) and XRF (Phillips, 1404) techniques. Recycling was carried out based on a pyrometallurgical process. In this process, the mixture of coke (provided by Zarand Iranian Steel Company), steel turning chips, and Na2CO3 (produced by Kaveh Soda Chemical Industries Company) was heated to different temperatures for different periods. The levels selected for these parameters are summarized in Table 1.

Table 1 The selected levels for each process parameter

A mixture of precursors was placed in a cast-iron crucible in an electric furnace pre-heated to a preset temperature. After the processing time, the crucible was removed from the furnace, and the melt was poured into a steel mold after the slag was removed. To determine the lead recovery efficiency, the solidified ingot was weighted using a digital scale with 0.0001 g accuracy, and its chemical composition was measured using a mass spectrometry technique.

In this study, D-optimal array, i.e., one of the most common design of experiments (DOE) techniques [22,23,24,25,26,27], was used to determine the schedule of experiments. This technique suggested the performance of 102 trials (Table 2), instead of 36,864 (based on the full factorial design(, without affecting the accuracy of the collected dataset [21].

Table 2 Details of the experiments carried out in this study, Fe, coke, and Na2CO3 are in g per 100 g precursor

Results and discussion

Table 3 shows the efficiency of lead recoveries in various experiments carried out in this research with various process parameters denoted in Table 2.

Table 3 Lead recovery, in % with a measurement accuracy ± 1%, obtained in the experiments denoted in Table 2

Table 3 shows that the process parameters greatly influenced the lead recovery and caused it to vary between 2 and 98%. The bivariate (Pearson) correlation analysis is used to find the correlation between the selected process parameters. It is necessary to note that if the process parameters have significant correlation coefficients due to the incorrect exaggeration of each parameter and the evolution of multicollinearity, then the proposed model would not have a high accuracy [26]. Figure 6 shows the correlation matrix of the process parameters. As shown, the correlation coefficient between any two process parameters is less than 0.125, and therefore, there is no significant correlation between any of the selected parameters.

Fig. 6
figure 6

The correlation matrix of selected process parameters in the current study

Phase analysis of Pb-rich waste

The concentration of phases present in the lead-rich waste of the liberator cells, measured by XRF, is shown in Table 4. The XRD spectrum obtained from this waste is also shown in Fig. 7. These results confirmed the main phases present in the waste to be As2O3, Sb2O3, PbO, and PbSO4.

Table 4 Concentration of phases present in the lead-rich waste of the liberator cells, determined by XRF
Fig. 7
figure 7

XRD spectrum obtained from the lead-rich waste of the liberator cells

According to the literature [2, 23], reduction of PbO, PbS, and PbSO4 were carried out by coke, Fe, and Na2CO3, respectively, and hence, determination of the amount of these phases for the start of the optimization process was essential. During the selection of the minimum and maximum levels of the reducing agents, all the Pb was supposed to be in the form of PbO and PbSO4.

SVR-MTSA modeling

Figure 8 shows the change of fitness versus iteration for different meta-heuristic strategies. Based on the criteria for the evaluation, i.e., RMSE, MAPE, R2, and MBE, the figure shows that the SVR-MTSA outperformed the BPNN and other variations of the SVR model.

Fig. 8
figure 8

Changes of fitness versus iteration for different meta-heuristic strategies

All the curves had a descending trend, which is typical for optimization algorithms. At the start of the iteration process, this decreasing trend was more significant for the SVR-TSA and SVR-MTSA strategies. Due to the exploitation process, the fitness curve of SVR-MTSA decreased more smoothly. It can be concluded that the SVR-MTSA strategy can effectively be balanced between exploitation and exploration compared to SVR-TSA using a new updating strategy. Finally, the results indicated that the fitness of the SVR-MTSA model could achieve 6% better than SVR-TSA and 11% better than SVR-GWO. One of the reasons for this superiority is because MTSA balances between exploration and exploitation by changing the corresponding variable and the number of iteration while GWO sets this variable by a random number.

Figure 9a represents the RMSE (Eq. (18)) versus iteration number for various models. As mentioned, a model with higher accuracy has a smaller RMSE value. It can be seen that the SVR-MTSA had a fast convergence speed and could reach to the lowest RMSE due to its ability to search the most promising areas. Moreover, the SVR had a superior generalization performance compared to the neural network in a regression problem [27]. As shown in Fig. 8b, the efficiency of Pb recovery is enhanced after 200 iterations due to the effective scanning of search spaces by the proposed model.

Fig. 9
figure 9

Comparison of a RMSE and b the efficiency of the meta-heuristic strategies on predicting the efficiency of lead recovery as a function of iteration number

Figure 10 shows the change in the correlation coefficient, R2, versus iteration number for different optimization algorithms. The closer to unity the R2 value, the higher the accuracy of the optimization method. Figure 10 revealed that the SVR-MTSA had the highest R2 value. The high accuracy of the MTSA algorithm can be attributed to the fact that it is not limited to the local optimal values and explores the entire search space. In other words, the MTSA generates the best optimal solution and continuously improves the solution by changing the direction of the movement of the tunicates, which enhances its exploration ability. Figure 11 summarizes the various statistical criteria for evaluating the accuracy of the three variations of the proposed SVR model, named MTSA-SRV1, MTSA-SRV2, and SVR-MTSA3 with three different kernels, i.e., Gaussian, linear and polynomial, respectively. Based on all the accuracy criteria, this figure determined the SVR-MTSA1 model to have the most accurate kernel.

Fig. 10
figure 10

Correlation coefficient, R2, versus iteration number for different optimization algorithms

Fig. 11
figure 11

Statistical criteria for evaluating the accuracy of MTSA-SRV1, MTSA-SRV2 and SVR-MTSA3 models with Gaussian, linear and polynomial kernels, respectively

Validation test

A test with the optimal process parameters suggested by the SVR-MTSA model (shown in Table 5) to achieve the highest Pb recovery is carried out to evaluate the accuracy of the proposed model. The model suggested this experiment to gain a 99.96% recovery. The result of the test showed the recovery to be about 99%, a good agreement with the predicted value.

Table 5 The optimal process parameters suggested by the SVR-MTSA model to obtain the highest Pb recovery, and the recovery obtained by the model and experimentally

As shown in Table 6, For 100 iterations, SVR-GWO has better performance in terms of fitness whereas SVR-MTSA is the second best optimizer. With the increasing number of iterations (i.e., 200), the modified parameter A helps to the proposed SVR-MTSA and can escape from local optima and reach to a better fitness compared to when the number of iteration is set to 100.

Table 6 Comparison of the performance of the SVR-MTSA algorithm with six other optimization algorithms

The output does not always indicate better convergence and diversity because sometimes the result obtained may be different from the optimal solution. The Wilcoxon signed-rank test [28] is performed for the average value of the fitness values and R-squared. In each metric, the difference between the mean results is calculated for each pair. Then, these differences are arranged in an ascending order and if the proposed technique is better than the competitor method, the positive rank is obtained. Table 6 presents the Wilcoxon test, where + , − , and = represent that the output of SVR-MTSA is superior, inferior, and equal to competitor techniques, respectively. From Table 7, it is observed that SVR-MTSA outperforms all the competitor algorithms except SVR-GWO, which finds superior for fitness and R-squared when iteration is set to 100. For 200 iterations, the proposed method (i.e., SVR-MTSA) reaches better fitness value in average compared to other techniques.

Table 7 Wilcoxon signed-rank test

We used two more criteria namely variance accounted for (VAF in Eq. (20)) and Median of Absolute errors (MedAE in Eq. (21)) to better show the performance of competitive algorithms.

$${\text{VAF}} = \left[ {1 - \frac{{{\text{var}} (y - \hat{y})}}{{{\text{var}} (y)}}} \right] \times 100,$$
(20)
$${\text{MedAE = median(}}y - \hat{y}{),}$$
(21)

where y is the actual output and \(\hat{y}\) is the output predicted by the suggested model. In addition, \(\overline{y}\) indicates the average of the actual outputs.

From Fig. 12, it can be seen that the proposed algorithm (i.e., SVR-MTSA) improves VAF by 7.2% and 6.5% compared to SVR-GWO and SVR-TSA, respectively. A model with a high value for VAF parameter means that it has been able to predict the output with high accuracy. The SVR where the parameters are optimized with GWO has better performance until the 170th iteration compared to SVR-MTSA and then it falls into the local optimum due to the lack of local search operator. While the proposed MTSA helps SVR to achieve an appropriate value for Gaussian kernel and reduce error.

Fig. 12
figure 12

The comparison of different methods in terms of VAF

Figure 13 illustrates the median of absolute errors for various methods over 200 iterations. This parameter indicates the oscillation of errors. According to this parameter, a model with higher accurate prediction obtains lower error oscillation. The SVR-MTSA improves MedAE compared to Basic-SVR and BPNN-GD by 27.6% and 32.8%, respectively. Back propagation neural network and basic support vector regression use gradient descent to optimize neural network weights and SVR parameters. The gradient descent easily traps on local optimal [29] and suffers from the weak exploration operations. While by training the SVR with MTSA, the better result is achieved since it has better exploration activities in SVR-MTSA than these two algorithms.

Fig. 13
figure 13

The median of absolute errors for different methods over 200 iterations

The efficiency of the Pb recovery obtained experimentally was compared to those predicted by the SVR-MTSA model for the variation in processing time, temperature, coke, Fe, and Na2CO3 contents. The results, shown in Fig. 14a–e, respectively, showed a good agreement between these two values. Therefore, it can be concluded that the proposed model can predict lead recovery with reasonable accuracy for any combination of process parameters.

Fig. 14
figure 14

Comparison of the efficiency of the Pb recovery obtained experimentally with those predicted by the SVR-MTSA for the variation in a processing time, b temperature, c coke, d Fe, and e Na2CO3 contents

Sensitivity analysis using SVR-MTSA strategy

The sensitivity analysis is used to rank the effect of process parameters on the lead recovery efficiency using the proposed SVR-MTSA strategy. In this analysis, each parameter was changed between its minimum and maximum levels, while the other parameters were kept constant at their mean value. The results, shown in Fig. 15, revealed that temperature, coke content, processing time, Na2CO3 content, and Fe content had the most significant effect on the lead recovery efficiency, respectively.

Fig. 15
figure 15

The sensitivity analysis of selected features on the efficiency of Pb recovery, note: more fluctuation in each parameter means its higher impact

Comparison of the effect of experimental parameters on the lead recovery

Determining the simultaneous effect of process parameters on the efficiency of the Pb recovery process is essential. This knowledge can help the industry to choose the best combination of parameter values in any specific circumstance. In this regard, the interactions of binary parameters are evaluated using the SVR-MTSA, and the results are shown graphically in Fig. 16. It is worth noting that the values of the other parameters in each graph are kept constant at their proposed optimal values. The sensitivity analysis showed that temperature and coke content had the most significant effect on the recovery. A large portion of the figures that illustrate the binary influence of these process parameters (i.e., Fig. 16a–g)) include purple counters. While, the portion of the purple counters significantly decreased in Fig. 16c, i, and j with Fe and Na2CO3 at their axis. Determination of temperature and coke as the most effective process parameters implied that the amount of sulfide and oxide phases of Pb in the Pb-rich waste had a determinative effect on the efficiency of the recovery.

Fig. 16
figure 16

Comparison of the effect of two parameters on the efficiency of Pb recovery. The value of other parameters was kept constant at their proposed optimal values (T = 990 °C, t = 110 min, Coke = 11 g/100 g and Na2CO3 = 14.5 g/100 g)

Conclusion

In this study, a hybrid of support vector regression and a modified tunicate swarm algorithm (SVR-MTSA) strategy is developed to optimize the process parameters for recovery of Pb from the residual of the liberator cells.

  1. 1.

    The proposed SVR-MTSA strategy could keep a good balance between the exploration and exploitation abilities during the optimization;

  2. 2.

    the high ability of the proposed MTSA strategy for optimization of the SVR parameters is demonstrated;

  3. 3.

    the experimental results showed that the SVR-MTSR strategy could be effectively employed to predict the efficiency of the lead recovery process in any selected process parameters;

  4. 4.

    the sensitivity analysis of the proposed model revealed that temperature and coke content were the process parameters that had the most significant influence on the recovery of lead in the selected recovery process;

  5. 5.

    a recovery efficiency of higher than 99% could be achieved when the optimal process parameters for Pb recovery from the liberator cells waste were chosen using the proposed SVR-MTSA model. This recovery efficiency was higher than those obtained from the neural network model.