1 Introduction

The problem of finding the position of passive target based on noisy TOA measurements is an essential task that has received extensive attention in various applications, such as radar, wireless communications, navigation, environmental monitoring, biomedical health monitoring (Wu et al. 2016; Deak et al. 2012; Sachs 2013), etc. In general, depending on the nature of the target to be determined, localization systems can be classified into active and passive, based on the target’s participation in the localization process (Deak et al. 2012). In the active localization, the target actively participates in the localization process. Unlike active localization systems, in passive localization systems, the target is not involved in the localization process. The passive localization aims to determine the position of the target without any device or tag; in fact, the target can only reflect or scatter the signals from transmitter to receivers (Wu et al. 2016).

Generally, the global positioning system (GPS) and other similar global navigation satellite systems provide access for obtaining the exact position of an object anywhere on the Earth surface (Yan et al. 2018). However, their positioning accuracy is not satisfactory due to the restriction of signal arrival in special environments, such as indoor, urban and underwater acoustics, and therefore, other solutions are required for target localization (Choi et al. 2014; Noroozi and Sebt 2015). In this context, the passive localization system is one of the widely used systems to estimate the unknown position of the target, due to its ability to achieve very high localization accuracy. Hence, this paper considers the passive target localization problem using noisy TOA measurements, as the sum of the signal propagation time from the known position of the transmitter to the target with unknown position, and that from the target to each receiver with known position. In this case, the signal propagation time can be measured and used to estimate ranges (transmitter–target–receiver distances). In fact, the obtained noisy distance measurements are nonlinear with respect to the unknown position of the target. Hence, the sum of transmitter-target and target-receiver ranges defines an ellipse. In this way, the estimated position of the passive target can be found inside the region bounded by the corresponding ellipses. In this context, the estimation of the target position is one of the most challenging optimization problems, which in recent years has become of practical significance (Chalise et al. 2014). In the line-of-sight (LOS) environment, the TOA measurement errors can be modeled as Gaussian distributed random variables. Therefore, the ML estimator which is based on an a priori statistical information of the measurement error by maximizing the likelihood function, can be employed in order to obtain the unknown target position (Wang et al. 2014). The corresponding ML objective function for the given localization problem is a highly nonlinear and non-convex function, where conventional optimization methods are not suitable for solving such a problem. Consequently, this has motivated the development of different efficient optimization methods, which can accurately solve the ML estimation problem. Among them, the SDP has gained considerable attention as an efficient method that can be used to solve convex relaxations for the given localization problem (Chalise et al. 2014). The basic idea of the SDP method is to transform the non-convex problem to a convex one that monotonically converges to the global optimal solution. The main advantage of the SDP method is that it does not require an initial guess at the beginning of the optimization process, and in this way, the global optimal solution of the ML objective function can be efficiently found (Chan et al. 2013). In fact, the SDP method has received significant attention due to its numerical reliability and robustness in solving a variety of localization problems. The SDP problem can be efficiently solved using the MATLAB package CVX, where the solver is SeDuMi (Grant and Boyd 2014). However, the SDP method has some disadvantages in the presence of significant measurement noise, which is reflected in the accuracy of the global optimal solution, and thus, it cannot provide high estimation accuracy (Wang et al. 2013). In view of this, the problem of finding the global optimal solution with high accuracy for the multimodal objective function becomes a significant challenge.

Motivated by these facts, this paper proposes evolutionary algorithms (EAs) to overcome these drawbacks in finding the global optimal solution for a given multimodal optimization problem (Zhang et al. 2016). Generally, the evolution process of EAs consists of two stages, namely global exploration and local exploitation, which have the crucial role for the performance of the algorithm (Mohamed and Almazyad 2017). In such a context, the exploration stage refers to search for the global solution, while the exploitation stage concentrates on the promising area to refine the solution around the current global best solution. Hence, finding an appropriate balance between exploration and exploitation of the search space is an important challenge for the optimization algorithm.

Various EAs have been successfully applied to solve complex optimization problems, such as genetic algorithm (GA) (Kuila et al. 2013), DE (Storn and Price 1997), FA (Wu et al. 2019), cuckoo search algorithm (CSA) (Goyal and Patterh 2014), particle swarm optimization (PSO) (Zhang et al. 2015b), etc. Among the various existing algorithms, the DE and FA have been widely applied for different multimodal optimization problems, due to their compact structure, easy implementation, reliability and robustness (Wu et al. 2019, 2018).

The FA is nature-inspired metaheuristic algorithm, which is based on the flashing behavior of fireflies. The search process of FA depends on the attraction of fireflies, whereby a less bright firefly moves toward a brighter firefly (Yang 2010b). Hence, the characteristics of the flashing light of fireflies provide effective diversity in searching for the global optimal solution. In this regard, the objective function for a given optimization problem is associated with the light intensity. The FA has been successfully employed to solve different complex optimization problems, and it has shown its superiority in comparison with the optimization performance of well-known algorithms, such as GA and PSO (Yang 2009). Similar to other EAs, the FA has some drawbacks, such as premature convergence to a local optimum, slow convergence rate, especially in the later stage of the evolution process, and sensitivity to the choice of control parameters (Wang et al. 2017).

The hybridization of FA with other metaheuristic algorithms has been performed, in order to overcome drawbacks of conventional algorithms, which results in a powerful and highly effective optimization algorithm. For this purpose, the simulated annealing (SA) has been hybridized with FA into hybrid SFA, with the aim to improve the accuracy of the global optimal solution during the evolution process (Alweshah and Abdullah 2015). Moreover, to control the randomization movement of the FA, the crossover and mutation operators of GA have been integrated into the FA, which has increased convergence speed (Luthra and Pal 2011). There are also different hybrid variants of the FA with other metaheuristic methods, such as hybrid FA with DE (AHEFA) (Lieu et al. 2018), FA with PSO (HFPSO) (Aydilek 2018) , FA with harmony search (HS/FA) (Guo et al. 2013) and FA with ant colony optimization algorithm (HAFA) (Goel and Maini 2018), which are very robust in finding global optimal solutions. Therefore, hybridization of these algorithms has been successfully applied to solve various complex optimization problems, due to their effectiveness, improvement of convergence speed and solution accuracy, and the ability to avoid being trapped in local minima.

The DE is another population-based stochastic algorithm, developed by Storn and Price (Storn and Price 1997), that can be used for solving numerous optimization problems in different fields of science and engineering (Baraldi et al. 2016; Awad et al. 2018). The performance of the DE algorithm is significantly influenced by its control parameters, such as scale factor (F), crossover rate (CR) and population size (NP), and trial vector generation operators, i.e., mutation and crossover operators (Cui et al. 2016). In the DE algorithm, the values of the control parameters are kept fixed throughout the evolution process, which may cause the algorithm to be easily trapped in local optima of a multimodal objective function. Therefore, in order to improve the search performance of DE algorithm, many different mechanisms have been introduced to adjust control parameters in an adaptive or self-adaptive manner (Das et al. 2016). Among them, the JADE is a well-known DE variant with an effective parameter adaptation technique, where control parameters F and CR, associated with each individual in the population, are randomly and independently generated according to Cauchy and Gaussian distributions, respectively (Zhang and Sanderson 2009). Their expected mean values are adaptively updated using the successful values of F and CR during the evolution process to maintain diversity of the population and improve the robustness of the algorithm. Building on the success of JADE, the SHADE is proposed as the improved success history-based parameter adaptation technique, where the values of F and CR are adapted based on learning from successes and failures of past generations to accurately locate the global optimum (Tanabe and Fukunaga 2013).

Besides the control parameters, the mutation operator plays an important role in the evolution process of the DE algorithm. Various mutation operators have been suggested in the literature, such as DE/rand/1, DE/rand/2, DE/best/1, DE/best/2, DE/current-to-best/1, DE/current-to-rand/1 and other variants of them (Mohamed and Suganthan 2018). In addition, in the JADE algorithm, an effective mutation operator DE/current-to-pbest/1 is developed by improving the classic DE/current-to-best/1 operator (Zhang and Sanderson 2009). More recently, in composite DE (CoDE), a new mutation operator is proposed, which combines three trial vector generation strategies, such as DE/rand/1/bin/, DE/rand/2/bin, DE/current-to-rand/1, and three control parameter settings to generate trial vectors (Wang et al. 2011). In this regard, in order to achieve a proper balance between exploration and exploitation abilities, the mutation operator cannot be the same at different stages of the evolution process (Qian et al. 2018).

Motivated by the above considerations, in this paper, an improved adaptive hybrid algorithm, based on hybridization of the DE and FA, named AHFADE, is presented in order to efficiently solve the passive target localization problem. This paper aims to maintain population diversity and improve the convergence speed by incorporating evolutionary operators of the DE into the FA. To further improve the optimization performance, the control parameters of DE and FA are automatically and adaptively updated during the evolution process, where parameters F and CR are generated using Cauchy and Gaussian distributions, respectively. In addition, the proposed AHFADE algorithm incorporates two mutation operators DE/rand/1 and DE/current-to-pbest/1 into the FA to achieve an effective trade-off between exploration and exploitation abilities. Therefore, the purpose of this paper is to develop a robust optimization algorithm to solve the passive target localization problem for a wide range of measurement noise.

The CRLB is an important and widely applied in the localization for the analysis of the accuracy of the estimated target position, which provides a lower bound for the root mean square error (RMSE) of any unbiased estimator (Shen et al. 2012). Hence, the CRLB for the passive target localization problem is compared with the RMSE performance of the proposed AHFADE algorithm and the existing SDP, DE and FA.

The main contributions of this paper are summarized as follows:

  • The passive target localization problem is defined in two dimensions based on the noisy TOA measurements obtained from a single transmitter and multiple receivers in LOS environment. Since the ML objective function of this localization problem is a highly nonlinear and non-convex function, it is necessary to provide sophisticated optimization algorithms, to estimate the position of the target.

  • In order to efficiently solve the above ML estimation problem, the well-known SDP method can be employed to transform the non-convex optimization problem into a convex problem.

  • The improved AHFADE algorithm, as a hybridization of the DE and FA, is established to provide desirable accuracy in LOS environment. The proposed AHFADE algorithm aims to improve the accuracy of the global optimal solution by incorporating evolutionary operators of the DE into the FA. Furthermore, the control parameters of the proposed AHFADE algorithm have been dynamically updated at each iteration to generate potential solutions with higher diversity.

  • The statistical analysis using the Wilcoxon signed-rank test and Friedman test has been performed, in order to compare the performance of the proposed AHFADE algorithm to several well-known algorithms on CEC2014 benchmark test problems. The obtained results further confirmed that the hybridization proposed in this paper can improve the overall optimization performance of the proposed AHFADE algorithm.

  • The obtained numerical simulation results have shown that the AHFADE algorithm achieves the CRLB and has significantly better performance than existing algorithms. Furthermore, the simulation results indicate that the AHFADE is robust against large measurement noise and changes in the geometric configuration of the transmitter, target and receivers compared to SDP, FA and DE algorithms.

The rest of the paper is organized as follows. Section 2 gives a review of background and related work. The passive target localization system based on the noisy TOA measurements is presented in Sect. 3. In Sect. 4, the ML objective function is derived. The SDP method for solving the passive target localization problem is presented in Sect. 5. Section 6 introduces the DE algorithm, followed by the corresponding improvements. FA with an adaptive attractiveness function is described in Sect. 7. Section 8 introduces an improved AHFADE algorithm. In Sect. 9, the CRLB for passive target localization problem is derived. Numerical simulation results are provided in Sect. 10. The conclusions and some possible directions for future work are given in Sect. 11.

2 Background and related work

The accuracy of passive target position has always been an important and challenging issue in many applications that has received significant interest among the research community (Shen et al. 2012). It is well known that the reliability and accuracy of the localization depend on the geometric configuration of the transmitter, receivers and target, the measurement accuracy and the environment (e.g., ambient noise) (Bishop et al. 2010). Consequently, numerous localization algorithms have been proposed in the literature for different applications, which can be divided into two categories: range-based and range-free algorithms (Chowdhury et al. 2016). The range-based localization algorithms utilize range measurements, such as distance or angle information between the unknown position of the target and the known position of each receiver in order to estimate the position of the target. These algorithms use various measurement techniques, such as TOA (Shen et al. 2012), time difference of arrival (TDOA) (Huang et al. 2015), received signal strength (RSS) (Li et al. 2017), angle of arrival (AOA) (Xu and Doğançay 2015) or combination of them (Coluccia and Fascista 2018). The localization accuracy is dependent on the distance or angle measurements to estimate the position of the target for a given localization problem. Generally, the unknown position of the target can be determined using geometric methods such as triangulation, trilateration and multilateration or by applying an estimation method (Asmaa et al. 2014). Among them, the TOA is a promising method for target localization based on a set of receivers with known positions, which has attracted considerable research focus, due to its low cost and high localization accuracy (Sadowski and Spachos 2018). Conversely, the range-free localization algorithms do not require range or angle information between sensor nodes, and they utilize only connectivity information for localization. Basically, wireless sensor networks (WSNs) have tremendous ability to collect and transfer information of the deployed sensors, which can be used to determine the possible position of the target in their radio range (Halder and Ghosal 2016). The well-known range-free localization algorithms are: the centroid or weighted centroid algorithm, distance vector hop (DV-hop) and approximate point-in-triangulation test (APIT) (Singh and Sharma 2015). Compared with range-based algorithms, the range-free algorithms do not need additional hardware to determine the actual range; hence, they are cost-effective, easy to implement and have a low power consumption; however, this leads to less accurate localization performance (Halder and Ghosal 2016).

One of the most commonly used estimation methods to estimate the unknown position of the target is the nonlinear least squares (NLS) estimator (Zekavat and Buehrer 2011). Hence, the estimated target position can be obtained by minimizing the squared sum of the measurement errors with respect to the unknown target position. Thus, the given localization problem can be formulated as nonlinear minimization problem, which does not have closed-form solution, and it is difficult to solve.

Local search optimization algorithms have been widely applied for solving various localization problems (Kaur et al. 2016). These optimization algorithms can be classified into gradient-based and direct search methods, which only perform local search around the initial solution (Rao and Rao 2009). Different gradient-based algorithms such as steepest decent, Gauss–Newton and conjugate gradient have been proposed in the literature to find the unknown position of the target for the given NLS problem (Arora 2004). However, local search algorithms require feasible starting point to find the global optimum (Rao and Rao 2009). Therefore, these algorithms cannot guarantee convergence to the global optimum of a multimodal optimization problem and thus, cannot be used to solve passive target localization problem. In order to obtain a closed-form solution, the nonlinear optimization problem has been linearized using linear least squares (LLS) and weighed least squares (WLS) algorithms under the small Gaussian noise assumption (Wang 2015; Einemo and So 2015). To further improve the localization performance, especially in the presence of high-noise measurements, the constrained weighted least squares (CWLS) algorithm is developed, to obtain the estimated position of the target (Qu and Xie 2016). The CWLS rearranges the nonlinear equations into a set of linear equations, by explicitly incorporating the relationship between the target position and the auxiliary variable as a second-order equality constraint. In fact, the CWLS problem can be formulated as a quadratically constrained quadratic program (QCQP), which is converted to an unconstrained optimization problem. In this regard, the simulation results have shown that the RMSE of the CWLS algorithm can achieve the CRLB accuracy under small measurement errors (Lin et al. 2013). Another widely employed estimation method in the literature is the ML estimator, which maximizes the likelihood function of the unknown target position (Destino and Abreu 2011). However, the corresponding ML objective function is a highly nonlinear and non-convex function, where conventional optimization methods are not suitable for solving this kind of optimization problem. In order to overcome the non-convexity of the ML estimation problem, it can be transformed into a convex optimization problem using SDP or second-order cone programming (SOCP) (Kocuk et al. 2016). In this respect, both the SDP and SOCP methods can be effectively employed to approximately solve non-convex optimization problems, and find a global or near-global optimal solution. Simulation results have shown that the SDP method provides better localization accuracy compared to SOCP method. However, these methods cannot provide a high estimation accuracy in the presence of significant measurement noise (Biswas et al. 2006). To overcome this problem, there has been great interest to improve and develop efficient metaheuristic optimization algorithms, which can successfully find the global optimal solution for a given localization problem, without using gradient information during the search process.

Numerous metaheuristic algorithms like PSO, CSA, artificial bee colony (ABC), etc., have been widely applied in order to solve the localization problem in terms of determining the unknown position of the target (Kulkarni et al. 2019; Jiang et al. 2018). Therefore, there is still a need to choose an appropriate optimization algorithm as the straightforward way to minimize the localization error in all environments for a given localization problem. In this regard, the PSO algorithm and its improved variants have been employed to improve the localization accuracy for the passive target localization problem based on TDOA measurements (Cakir et al. 2014). The considered localization algorithm estimates the target position based on the nonlinear least squares residuals. The comparative analysis has shown that the improved variants of PSO algorithm can provide higher localization accuracy in comparison with the conventional PSO and well-known LLS and WLS algorithms for the given problem. Furthermore, the TDOA passive localization algorithm based on the ML objective function and CSA has been proposed to estimate the possible position of the target (Jiang et al. 2018). In the proposed algorithm, the Lévy flight mechanism has been adopted to effectively avoid being trapped in local optimum of a multimodal objective function by improving population diversity. Additionally, the simulation results show that the CSA has more efficient global exploration ability of the search space than existing conventional PSO algorithm. The DE algorithm and its hybrid variants such as PSO and DE (PSODE) (Zhang et al. 2015a) and DE and FA (HFDLA) (Harikrishnan et al. 2016) have been also proposed to estimate unknown position of the target in order to improve the localization performance and overcome the drawbacks of getting trapped in a local minimum during the evolution process.

Therefore, this paper proposes an improved robust AHFADE algorithm, based on the hybridization of DE and FA, to solve the multimodal passive target localization problem with high accuracy even in highly noisy environments.

3 Localization problem

In this section, the passive target localization problem based on TOA measurements is presented to estimate the unknown position of the target in LOS environment. The proposed localization system consists of N receivers, \(\left\{ {{R_{i}}} \right\} _{i = 1}^{N}\), whose known positions are \({\mathbf {x}_{i}^{r}} = \begin{bmatrix}{x_{i}^{r}}&{y_{i}^{r}} \end{bmatrix}^T \in {\mathbb {R}}^2 , \forall i \in \left\{ 1,2,\ldots ,N \right\} \) such that \(N\ge 3\), where \(T_x\) is the transmitter located at the origin of the coordinate system, i.e., \({\mathbf {x}}_{T} = \begin{bmatrix} 0 &{} 0\\ \end{bmatrix}^T \in {\mathbb {R}}^2\), as illustrated in Fig. 1. Assume that the true coordinates of the target are \({\mathbf {x}} = \begin{bmatrix} x &{} y\\ \end{bmatrix}^T \in {\mathbb {R}}^2\).

Fig. 1
figure 1

Passive target localization system using noisy TOA measurements

The passive target localization problem uses noisy TOA measurements, where the transmitter emits a signal and the receivers subsequently receive the signal reflected from the target. Thus, each receiver captures two signals in a row: one direct signal from the transmitter and the reflected signal from the target. It is assumed that the transmitter and receivers are perfectly synchronized, and the target reflects the TOA signal in all directions without any interaction with the transmitter and receivers in the network (Shen et al. 2012). The additive white Gaussian noise is widely used in localization algorithms, which is a reasonable assumption in LOS conditions (Wu et al. 2016). In order to find the estimated position of the target, there must be at least three TOA measurements. In fact, the noisy TOA measurements are composed of time from the transmitter to the unknown target position, and the time from the target to each receiver, which are functions of the unknown target position, i.e.,

$$\begin{aligned} {t_{i}} = \frac{1}{c}\left( {{{\left\| {\mathbf {x}} \right\| }_2} + {{\left\| {{\mathbf {x}} - \mathbf {x}_i^r} \right\| }_{2}}} \right) + {\bar{n}_i}, \quad \forall i \in \left\{ 1,2,\ldots ,N \right\} , \end{aligned}$$
(1)

where \( \left\| \cdot \right\| _2 \) denotes the Euclidean distance in two-dimensional space, c is the speed of light, and \({{\bar{n}}_i}\) is the zero-mean Gaussian measurement noise with variance of \(\sigma _i^2 \). Then, multiplying both sides of Eq. (1) by the speed of light c, yields range measurements (transmitter–target–receiver distances) denoted by \(\left\{ {{\tilde{d}_{i}}} \right\} _{i = 1}^{N}\), which can be expressed as follows:

$$\begin{aligned} {\tilde{d}_{i}}&= c \cdot t_i \nonumber \\&= d_{i} + n_{i}, \quad \forall i \in \left\{ 1,2,\ldots ,N \right\} , \end{aligned}$$
(2)

where the true range measurement \({d_{i}} = {{{\left\| {\mathbf {x}} \right\| }_2} + {{\left\| {{\mathbf {x}} - \mathbf {x}_i^r} \right\| }_{2}}} \) is the sum of transmitter–target and target–receiver distances, \({n_i} = c{{\bar{n}}_i}\) follows Gaussian distribution \({\mathcal {N}}\left( 0,\,\sigma _{ni}^2 \right) \) with zero-mean and variance of \(\sigma _{ni}^2 = {c^2} \sigma _i^2\). Hence, the TOA range measurements for every transmitter and receiver \(\left( {{T_x} - {R_i}} \right) \) pair have positive errors, which imply that the range measurements are greater than the true ones, as shown in Fig. 1, (Zhang et al. 2014).

In the absence of measurement errors, from the geometric interpretation, the sum of transmitter–target and target–receiver ranges \(d_{i}\) at time \({t_{i}}\), defines the solid-line ellipse \( E_i\), whose foci are placed at transmitter and receiver positions \({\mathbf {x}_T}\) and \(\mathbf {x}_i^r\), respectively, as shown in Fig. 1. Therefore, the true range \({d_{i}}\) of each transmitter and receiver \(\left( {{T_x} - {R_i}} \right) \) pair is constant for every point on an ellipse \( E_i\) , which can be defined as follows:

$$\begin{aligned} {E_i} = \left\{ {\left( {x,y} \right) \in {\mathbb {R}}^2 : {d_i} = \sqrt{{x^{2}} + {y^{2}}} + \sqrt{{{\left( {x - x_i^{r}} \right) }^2} + {{\left( {y - y_i^{r}} \right) }^2}} } \right\} .\nonumber \\ \end{aligned}$$
(3)

Hence, the vector of true range measurements can be expressed as follows:

$$\begin{aligned} \mathbf {d\left( x\right) } = \begin{bmatrix} {{{\left\| {\mathbf {x}} \right\| }_2} + {{\left\| {{\mathbf {x}} - \mathbf {x}_1^r} \right\| }_{2}}}\\ {{{\left\| {\mathbf {x}} \right\| }_2} + {{\left\| {{\mathbf {x}} - \mathbf {x}_2^r} \right\| }_{2}}}\\ \vdots \\ {{{\left\| {\mathbf {x}} \right\| }_2} + {{\left\| {{\mathbf {x}} - \mathbf {x}_{N}^{r}} \right\| }_{2}}} \end{bmatrix}. \end{aligned}$$
(4)

Therefore, \(\mathbf {d\left( x\right) }\) defines a set of ellipses \(\left\{ {{E_{i}}} \right\} _{i = 1}^{N} \), with respect to different transmitter and receiver pairs \({T_x}-{R_i}, \) \(\forall i \in \left\{ 1,2,\ldots ,N \right\} \), with two foci placed at positions \(\mathbf {x}_T\) and \(\mathbf {x}_i^r,\) respectively. Thus, the true position of the passive target \(\mathbf {x}\) is obtained by the intersection of three or more solid line ellipses \(\left\{ {{E_i}} \right\} _{i = 1}^{N \ge 3}\), as shown in Fig. 1. Therefore, the target position is obtained as the unique solution in the absence of measurement errors.

Additionally, in the presence of measurement errors, the vector form of Eq. (2) can be expressed as follows:

$$\begin{aligned} {\tilde{\mathbf{d}}} = \mathbf {d\left( x\right) } + \mathbf {n}, \end{aligned}$$
(5)

where \({\mathbf {n}} = \begin{bmatrix} n_1 &{} n_2 &{} \ldots &{} n_N \\ \end{bmatrix}^T\) is the vector of zero-mean noise. In this case, three or more dashed-line ellipses derived from noisy TOA measurements do not have a unique intersection point. Thus, the estimated position of the target can be found within the shaded region caused by error in the TOA measurements, as depicted in Fig. 1. Consequently, the localization accuracy is related to the size of this region which includes the target, i.e., the obtained small region can result in high localization accuracy.

Hence, in this paper, the main goal is to estimate the actual position of the target based on noisy TOA measurements by solving a highly nonlinear and multimodal ML estimation problem, as explained in the next section.

4 Maximum likelihood method

The estimation of the unknown position of the passive target can be obtained through maximizing the likelihood function. Under the assumptions of Gaussian noisy TOA measurements, which are independent and identically distributed, with a single transmitter, it is possible to express the likelihood function \( L\left( {\mathbf {x}} \right) \) for the target position (Shen et al. 2012) as follows:

$$\begin{aligned} L\left( \mathbf {x} \right)&= f\left( {{\tilde{\mathbf{d}}} \left| \mathbf {x} \right. } \right) = \frac{1}{{{{\left( {2\pi } \right) }^{N/2}}\det {{\left( \mathbf {C} \right) }^{1/2}}}} \nonumber \\&\quad \times \exp \left( { -\frac{1}{2}{{\left( {{\tilde{\mathbf{d}}} - \mathbf {d} \left( \mathbf {x} \right) } \right) }^T}{\mathbf {C}^{-1}}\left( {{\tilde{\mathbf{d}}} - \mathbf {d}\left( \mathbf {x} \right) } \right) } \right) , \end{aligned}$$
(6)

where \( f\left( {{\tilde{\mathbf{d}}} \left| \mathbf {x} \right. } \right) \) denotes the probability density function of the measurements, \( \det \left( \cdot \right) \) is the determinant of matrix, and \(\mathbf {C} \) is the diagonal covariance matrix given as:

$$\begin{aligned} \mathbf {C} = {\text {diag}} \lbrace {{\sigma _{n1}^2, \ldots ,\sigma _{nN}^2}} \rbrace . \end{aligned}$$
(7)

The logarithm of the likelihood function is used more often in practice as it turns the product of the terms into summation. Thus, the logarithm of Eq. (6) can be expressed as:

$$\begin{aligned} \ln L\left( {\mathbf {x}} \right) = k -\frac{1}{2}{{\left( {{\tilde{\mathbf{d}}} - \mathbf {d} \left( \mathbf {x} \right) } \right) }^T}{\mathbf {C}^{-1}}\left( {{\tilde{\mathbf{d}}} - \mathbf {d}\left( \mathbf {x} \right) } \right) , \end{aligned}$$
(8)

where \(k = \ln \left( {1}/{{{{\left( {(2\pi } \right) }^{N/2}}\det {{\left( \mathbf {C} \right) }^{1/2}}}}) \right) \) is a constant independent of \(\mathbf {x}\).

Consequently, the ML estimator requires the maximization of the log-likelihood function, which is equivalent to minimizing the negative logarithm of the likelihood function. Hence, the ML estimation problem can be formulated as:

$$\begin{aligned} \mathop {\min }\limits _{{\mathbf {x}} \in {\mathbb {R}}^2} {J_{\mathrm{ML}}}\left( \mathbf {x} \right) , \end{aligned}$$
(9)

where the corresponding ML objective function \( J_{\mathrm{ML}}\left( \mathbf {x} \right) \) can be written as:

$$\begin{aligned} {J_{\mathrm{ML}}}\left( \mathbf {x} \right) = {{\left( {{\tilde{\mathbf{d}}} - \mathbf {d} \left( \mathbf {x} \right) } \right) }^T}{\mathbf {C}^{-1}}\left( {{\tilde{\mathbf{d}}} - \mathbf {d}\left( \mathbf {x} \right) } \right) . \end{aligned}$$
(10)

Therefore, the estimated target position \(\hat{\mathbf {x}}\) can be obtained by minimizing the objective function \( J_{\mathrm{ML}}\left( \mathbf {x} \right) \) with respect to \(\mathbf {x}\), as follows:

$$\begin{aligned} \hat{\mathbf {x}} = \mathop {\arg \min }\limits _{{\mathbf {x}} \in {\mathbb {R}}^2} \left( {{J_{\mathrm{ML}}}\left( \mathbf {x} \right) } \right) . \end{aligned}$$
(11)

Figure 2 shows the corresponding contour plot for a given optimization problem and the gradient vectors, which indicate the direction of minimum of the objective function \(J_{\mathrm{ML}}\left( \mathbf {x} \right) \).

Fig. 2
figure 2

Illustration of the objective function \(J_{\mathrm{ML}}\left( \mathbf {x} \right) \) with the corresponding contour plot

From the contour plot in Fig. 2, it can be noticed that the objective function \(J_{\mathrm{ML}}\left( \mathbf {x} \right) \) is a highly nonlinear and non-convex function, whose global minimum \({{\hat{{\mathbf {x}}}}}\) corresponds to the unknown position of the target. In respect of this, the global optimal solution of the ML estimation problem cannot be obtained in closed form. Thus, in order to solve this kind of complex optimization problem, it is necessary to use sophisticated optimization algorithms, which are described in the next sections.

5 Semidefinite programming method

In this section, the SDP method is applied to transform the non-convex ML estimation problem into a convex optimization problem. According to Eqs. (9) - (11), the ML estimation problem with respect to \(\mathbf {x}\) can be formulated as:

$$\begin{aligned} {{\hat{{\mathbf {x}}}}} = \mathop {\arg \min }\limits _{\mathbf{{x}} \in {R^2}} {{\left( {{\tilde{\mathbf{d}}} - \mathbf {d} \left( \mathbf {x} \right) } \right) }^T}{\mathbf {C}^{-1}}\left( {{\tilde{\mathbf{d}}} - \mathbf {d}\left( \mathbf {x} \right) } \right) . \end{aligned}$$
(12)

Based on the approach (Wang and Wu 2016), the vector \(\mathbf {d}\left( \mathbf {x} \right) \) can be introduced as follows:

$$\begin{aligned} \mathbf {d}\left( \mathbf {x} \right) = \mathbf {H}{} \mathbf{{g}}, \end{aligned}$$
(13)

in which the \(\left( {N + 1} \right) \times 1\) vector \(\mathbf {g}\) is defined as:

$$\begin{aligned} \mathbf{{g}} = {\left[ {{{\left\| \mathbf{{x}} \right\| }_2}\begin{array}{*{20}{c}} {} \end{array}{{\left\| {\mathbf{{x}} - \mathbf{{x}}_1^r} \right\| }_2}\begin{array}{*{20}{c}} {} \end{array}{{\left\| {\mathbf{{x}} - \mathbf{{x}}_2^r} \right\| }_2} \ldots {{\left\| {\mathbf{{x}} - \mathbf{{x}}_N^r} \right\| }_2}} \right] ^T}, \end{aligned}$$
(14)

and

$$\begin{aligned} {\mathbf {H}} = \begin{bmatrix} \mathbf {1}_N &{} \mathbf {I}_N \\ \end{bmatrix} = \begin{bmatrix} 1&{}1&{} \cdots &{}0&{}0\\ 1&{}0&{} \cdots &{}0&{}0\\ \vdots &{} \vdots &{} \ddots &{} \vdots &{} \vdots \\ 1&{}0&{} \cdots &{}0&{}1 \end{bmatrix}. \end{aligned}$$
(15)

here \({\mathbf {1}_N}\) and \({\mathbf {I}_N}\) are column vector of length N with all entries equal to one and the \(N \times N\) identity matrix, respectively. The equality constraints are introduced between the transmitter \({\mathbf {x}_T}\) and the unknown position of the target \(\mathbf {x}\), and that target and ith receiver located at \(\mathbf {x}_i^r\), denoted by \(\left[ \mathbf{{g}} \right] _{1,1}^2 = \left\| \mathbf{{x}} \right\| _2^2 \) and \(\left[ \mathbf{{g}} \right] _{i,1}^2 = \left\| {\mathbf{{x}} - \mathbf{{x}}_i^r} \right\| _2^2,\;\begin{array}{*{20}{c}} {} \end{array}\forall i \in \left\{ {1,2,\ldots ,N} \right\} \), respectively. Then, the optimization problem in Eq. (12) can be rewritten as the following constrained optimization problem:

$$\begin{aligned} \mathop {\min }\limits _{\mathbf {x,g}}\quad&{\left( {{{\tilde{{\mathbf {d}}}}} - \mathbf{{Hg}}} \right) ^T}{\mathbf{{C}}^{ - 1}}\left( {{{\tilde{{\mathbf {d}}}}} - \mathbf{{Hg}}} \right) \nonumber \\ \text {s.t.} \quad&\left[ \mathbf{{g}} \right] _{1,1}^2 = {\left\| {\mathbf {x}} \right\| _2^2},\nonumber \\&\left[ \mathbf{{g}} \right] _{i,1}^2 = {\left\| {\mathbf{{x}} - \mathbf{{x}}_i^r} \right\| _2^2}, \nonumber \\&\forall i \in \left\{ {1,2, \ldots ,N} \right\} . \end{aligned}$$
(16)

The objective function for the above optimization problem, by corresponding algebraic manipulation, can be expressed as follows:

$$\begin{aligned} \begin{array}{l} {\left( {{{\tilde{{\mathbf {d}}}}}- \mathbf{{Hg}}} \right) ^T}{\mathbf{{C}}^{ - 1}}\left( {{{\tilde{{\mathbf {d}}}}} - \mathbf{{Hg}}} \right) \\ = {\left[ {\begin{array}{*{20}{c}} \mathbf{{g}}\\ 1 \end{array}} \right] ^T}\left[ {\begin{array}{*{20}{c}} {{\mathbf{{H}}^T}{\mathbf{{C}}^{ - 1}}{} \mathbf{{H}}}&{}{ - {\mathbf{{H}}^T}{\mathbf{{C}}^{ - 1}}{{{\tilde{\mathbf{d}}}}}}\\ { - {{{{{\tilde{\mathbf{d}}}}}}^T}{\mathbf{{C}}^{ - 1}}{} \mathbf{{H}}}&{}{{{{{{\tilde{\mathbf{d}}}}}}^T}{\mathbf{{C}}^{ - 1}}{{{\tilde{\mathbf{d}}}}}} \end{array}} \right] \left[ {\begin{array}{*{20}{c}} \mathbf{{g}}\\ 1 \end{array}} \right] \\ \end{array}. \end{aligned}$$
(17)

Based on the matrix property \({\mathbf{{x}}^T}{} \mathbf{{Ax}} = {{\mathrm{Tr}}}\left( {\mathbf{{x}}{\mathbf{{x}}^T}{} \mathbf{{A}}} \right) \) and introducing the matrix notation \(\mathbf{{G}} = \mathbf{{g}}{\mathbf{{g}}^T}\), then Eq. (17) is given as follows:

$$\begin{aligned} \begin{array}{l} \left[ {\begin{array}{*{20}{c}} {{\mathbf{{H}}^T}{\mathbf{{C}}^{ - 1}}{} \mathbf{{H}}}&{}{ - {\mathbf{{H}}^T}{\mathbf{{C}}^{ - 1}}{{{\tilde{\mathbf{d}}}}}}\\ { - {{{{{\tilde{\mathbf{d}}}}}}^T}{\mathbf{{C}}^{ - 1}}{} \mathbf{{H}}}&{}{{{{{{\tilde{\mathbf{d}}}}}}^T}{\mathbf{{C}}^{ - 1}}{{{\tilde{\mathbf{d}}}}}} \end{array}} \right] \left[ {\begin{array}{*{20}{c}} \mathbf{{g}}\\ 1 \end{array}} \right] \\ = {{\mathrm{Tr}}}\left( {\left[ {\begin{array}{*{20}{c}} \mathbf{{g}}\\ 1 \end{array}} \right] \left[ {\begin{array}{*{20}{c}} {{\mathbf{{g}}^T}}&{}1 \end{array}} \right] \mathbf {P}} \right) = {{\mathrm{Tr}}}\left( {\left[ {\begin{array}{*{20}{c}} \mathbf{{G}}&{}\mathbf{{g}}\\ {{\mathbf{{g}}^T}}&{}1 \end{array}} \right] \mathbf {P}} \right) \end{array}, \end{aligned}$$
(18)

in which

$$\begin{aligned} \mathbf {P} = \left[ {\begin{array}{*{20}{c}} {{\mathbf{{H}}^T}{\mathbf{{C}}^{ - 1}}{} \mathbf{{H}}}&{}{ - {\mathbf{{H}}^T}{\mathbf{{C}}^{ - 1}}{{{\tilde{\mathbf{d}}}}}}\\ { - {{{{{\tilde{\mathbf{d}}}}}}^T}{\mathbf{{C}}^{ - 1}}{} \mathbf{{H}}}&{}{{{{{{\tilde{\mathbf{d}}}}}}^T}{\mathbf{{C}}^{ - 1}}{{{\tilde{\mathbf{d}}}}}} \end{array}} \right] , \end{aligned}$$
(19)

where \({{\mathrm{Tr}}}\left( \cdot \right) \) denote the trace of a square matrix.

Introducing the \(z = {\mathbf{{x}}^T}{} \mathbf{{x}}\) , the constraints can be expressed as:

$$\begin{aligned}&{\left[ \mathbf{{G}} \right] _{1,1}} = \left\| \mathbf{{x}} \right\| _2^2= {\mathbf{{x}}^T}{} \mathbf{{x}}=z,\nonumber \\&{\left[ \mathbf{{G}} \right] _{i,i}} = \left\| {\mathbf{{x}} - \mathbf{{x}}_i^r} \right\| _2^2 = {\mathbf{{x}}^T}{} \mathbf{{x}} - 2{\mathbf{{x}}^T}{} \mathbf{{x}}_{i - 1}^r + {\left( {\mathbf{{x}}_{i - 1}^r} \right) ^T}{} \mathbf{{x}}_{i - 1}^r \nonumber \\ {}&= z - 2{{\mathbf {x}}^T}{\mathbf {x}}_{i - 1}^r + {\left( {{\mathbf {x}}_{i - 1}^r} \right) ^T}{\mathbf {x}}_{i - 1}^r, \nonumber \\&\forall i \in \left\{ {2, \ldots ,N+1} \right\} . \end{aligned}$$
(20)

Then, according to Eqs. (18) and (20), the optimization problem in Eq. (16) can be reformulated as:

$$\begin{aligned} \mathop {\min }\limits _{{\mathbf {x}}, {\mathbf {g}},{\mathbf {G}},{\mathop { z}\nolimits } } \quad&{{\mathrm{Tr}}} \left( \begin{bmatrix} {\mathbf {G}}&{}{\mathbf {g}}\\ {{{\mathbf {g}}^T}}&{}1 \end{bmatrix} \mathbf {P} \right) \nonumber \\ \text {s.t.} \quad&{\left[ \mathbf{{G}} \right] _{1,1}} = z,\nonumber \\&{\left[ \mathbf{{G}} \right] _{i,i}} = z - 2{{\mathbf {x}}^T}{\mathbf {x}}_{i - 1}^r + {\left( {{\mathbf {x}}_{i - 1}^r} \right) ^T}{\mathbf {x}}_{i - 1}^r, \nonumber \\&\forall i \in \left\{ {2, \ldots ,N+1} \right\} , \nonumber \\&{\mathbf {G}} = {\mathbf {g}}{{\mathbf {g}}^T},\nonumber \\&z = {{\mathbf {x}}^T}{\mathbf {x}}. \end{aligned}$$
(21)

Furthermore, the Cauchy–Schwartz inequality for the off-diagonal elements of the matrix \(\mathbf {G}\) can be written as:

$$\begin{aligned} {\left[ \mathbf{{G}} \right] _{i,j}}&= {\left\| {{\mathbf {x}} - {\mathbf {x}}_{i - 1}^r} \right\| _2}{\left\| {{\mathbf {x}} - {\mathbf {x}}_{j - 1}^r} \right\| _2} \nonumber \\&\ge \left| {{{\left( {{\mathbf {x}} - {\mathbf {x}}_{i - 1}^r} \right) }^T}\left( {{\mathbf {x}} - {\mathbf {x}}_{j - 1}^r} \right) } \right| \nonumber \\&= \left| {{{\mathbf {x}}^T}{\mathbf {x}} - {{\left( {{\mathbf {x}}_{i - 1}^r + {\mathbf {x}}_{j - 1}^r} \right) }^T}{\mathbf {x}} + {{\left( {{\mathbf {x}}_{i - 1}^r} \right) }^T}{\mathbf {x}}_{j - 1}^r} \right| , \nonumber \\ \forall i,j&\in \left\{ {1,2, \ldots ,N+1} \right\} , \quad i \ne j, \end{aligned}$$
(22)

with \({\mathbf {x}_0^r}={\mathbf {x}_T} = \begin{bmatrix} 0 &{} 0\\ \end{bmatrix}^T\).

Due to the nonlinear equality constraints \({\mathbf {G}} = {\mathbf {g}}{{\mathbf {g}}^T}\) and \(z = {{\mathbf {x}}^T}{\mathbf {x}}\), the optimization problem in Eq. (21) is a non-convex, where the global optimal solution is very difficult to obtain. Therefore, the non-convex equality constraints \({\mathbf {G}} = {\mathbf {g}}{{\mathbf {g}}^T}\) and \(z = {{\mathbf {x}}^T}{\mathbf {x}}\), can be relaxed into the convex inequality constraints, as follows:

$$\begin{aligned} \begin{array}{l} \mathbf{{G}} - \mathbf{{g}}{\mathbf{{g}}^T}\succeq \mathbf{{0}},\\ {z} - \mathbf{{x}}{\mathbf{{x}}^T}\succeq \mathbf{{0}}. \end{array} \end{aligned}$$
(23)

After applying the Schur complement (Boyd and Vandenberghe 2004), these constraints can be equivalently expressed as:

$$\begin{aligned}&\begin{bmatrix} {\mathbf {G}}&{}{\mathbf {g}}\\ {{{\mathbf {g}}^T}}&{}1 \end{bmatrix} \succeq \mathbf{{0}}, \nonumber \\&\begin{bmatrix} {z}&{}{\mathbf {x}}\\ {{{\mathbf {x}}^T}}&{}{{\mathbf {I}}_2} \end{bmatrix} \succeq \mathbf{{0}}. \end{aligned}$$
(24)

It should be noted that the matrices given in Eq. (24) are symmetric and positive semidefinite matrices, and therefore, the inequality constraints in Eq. (24) should be convex functions.

Finally, based on the above relaxation, the optimization problem in Eq. (21) becomes a convex problem, which can be written as follows:

$$\begin{aligned} \mathop {\min }\limits _{{\mathbf {x}},{\mathbf {g}},{\mathbf {G}},{\mathop { z}\nolimits } } \quad&{{\mathrm{Tr}}} \left( \begin{bmatrix} {\mathbf {G}}&{}{\mathbf {g}}\\ {{{\mathbf {g}}^T}}&{}1 \end{bmatrix} \mathbf {P} \right) \nonumber \\ \text {s.t.} \quad&{\left[ \mathbf{{G}} \right] _{1,1}} = z,\nonumber \\&{\left[ \mathbf{{G}} \right] _{i,i}} = z - 2{{\mathbf {x}}^T}{\mathbf {x}}_{i - 1}^r + {\left( {{\mathbf {x}}_{i - 1}^r} \right) ^T}{\mathbf {x}}_{i - 1}^r, \nonumber \\&\forall i \in \left\{ {2, \ldots ,N+1} \right\} , \nonumber \\&{\left[ \mathbf{{G}} \right] _{i,j}} \ge \left| {z - {{\left( {{\mathbf {x}}_{i - 1}^r + {\mathbf {x}}_{j - 1}^r} \right) }^T}{\mathbf {x}} + {{\left( {{\mathbf {x}}_{i - 1}^r} \right) }^T}{\mathbf {x}}_{j - 1}^r} \right| , \nonumber \\&\forall i,j \in \left\{ {1,2, \ldots ,N+1} \right\} , \quad i \ne j, \nonumber \\&\begin{bmatrix} {\mathbf {G}}&{}{\mathbf {g}}\\ {{{\mathbf {g}}^T}}&{}1 \end{bmatrix} \succeq \mathbf{{0}}, \nonumber \\&\begin{bmatrix} {z}&{}{\mathbf {x}}\\ {{{\mathbf {x}}^T}}&{}{{\mathbf {I}}_2} \end{bmatrix} \succeq \mathbf{{0}}. \end{aligned}$$
(25)

Therefore, the SDP problem given in Eq. (25) is a convex one, which can be efficiently implemented and solved by the CVX toolbox (Grant and Boyd 2014) using SeDuMi as the solver (Grant and Boyd 2008).

6 Differential evolution algorithm and the proposed improved versions

In this section, the conventional DE algorithm is first introduced. Then, a brief description of the improved adaptive versions of the DE such as SHADE and JADE algorithms with the mutation operator DE/current-to-pbest/1 is presented and implemented in this work.

6.1 Conventional DE algorithm

The DE is a population-based algorithm, which is based on four phases: initialization, mutation, crossover and selection, which are briefly summarized as follows.

6.1.1 Initialization

Generally, the population of NP random individuals at the current generation G, where each individual is expressed by n-dimensional vector as a potential solution, can be represented as follows:

$$\begin{aligned}&\left\{ {{{\mathbf {x}}_i^{\left( G \right) }}:x_{i,j}^{\left( G \right) } \in \left[ {x_{i,j}^L,x_{i,j}^U} \right] } \right\} ,\nonumber \\&\forall i \in \left\{ {1,2,\ldots ,\mathrm{NP}} \right\} ,\forall j \in \left\{ {1,2,\ldots ,n} \right\} \end{aligned}$$
(26)

where \(x_{_{i,j}}^L\) and \(x_{_{i,j}}^U\) denote the lower and upper bounds of the jth component of the ith individual \(x_{{i,j}}\), respectively. Therefore, the initial jth component of the ith individual is randomly generated according to a uniform distribution within the corresponding boundary constraints \(x_{i,j}^L \le x_{i,j}^{\left( 0 \right) } \le x_{i,j}^U\), which can be defined as follows:

$$\begin{aligned} x_{i,j}^{\left( 0 \right) } = {x_{i,j}^L} + {\mathrm{rand}}_j\left( {{x_{i,j}^U} - {x_{i,j}^L}} \right) , \end{aligned}$$
(27)

where \({\mathrm{rand}}_j\) is a uniformly distributed random number in the range \(\left[ {0,1} \right] \).

6.1.2 Mutation

The mutation operator plays a vital role during the evolution process, where a mutant vector for each target vector \({\mathbf {x}}_i^{\left( G \right) }\) can be expressed as follows:

$$\begin{aligned} {\mathbf {v}_{i}^{\left( G+1 \right) }} = \begin{bmatrix} v_{i,1}^{\left( G +1 \right) } &{} v_{i,2}^{\left( G +1 \right) } &{} \ldots &{} v_{i,j}^{\left( G +1 \right) } &{} \ldots &{} v_{i,n}^{\left( G +1 \right) } \\ \end{bmatrix}^T. \end{aligned}$$
(28)

In recent years, a number of different mutation operators for exploration of the search space, have been developed and proposed in the literature (Mohamed and Suganthan 2018), which are described as follows: DE/rand/1

$$\begin{aligned} {\mathbf {v}}_i^{\left( G +1 \right) } = {\mathbf {x}}_{r_1}^{\left( G \right) } + F \left( {{\mathbf {x}}_{r_2}^{\left( G \right) } - {\mathbf {x}}_{r_3}^{\left( G \right) }} \right) , \end{aligned}$$
(29)

DE/best/1

$$\begin{aligned} {\mathbf {v}}_i^{\left( G +1 \right) }= {\mathbf {x}}_{\mathrm{best}}^{\left( G \right) } + F \left( {{\mathbf {x}}_{r_1}^{\left( G \right) } - {\mathbf {x}}_{r_2}^{\left( G \right) }} \right) , \end{aligned}$$
(30)

DE/current-to-best/1

$$\begin{aligned} {\mathbf {v}}_i^{\left( G +1 \right) } \!=\! {\mathbf {x}}_i^{\left( G \right) } + F \left( {{\mathbf {x}}_{\mathrm{best}}^{\left( G \right) } \!-\! {\mathbf {x}}_i^{\left( G \right) }} \right) \!+\! F \left( {{\mathbf {x}}_{r_1}^{\left( G \right) } \!-\! {\mathbf {x}}_{r_2}^{\left( G \right) }} \right) , \end{aligned}$$
(31)

where \({r_1}\), \({r_2}\) and \(r_3\) are distinct integers randomly generated from the set \(\left\{ 1,2,\ldots ,\mathrm{NP} \right\} \backslash \left\{ i \right\} \), the \({\mathbf {x}}_{\mathrm{best}}^{\left( G \right) }\) denotes the individual with the best objective function value and \(F \in \left[ {0,1} \right] \) is the scale factor.

6.1.3 Crossover

After the mutation phase, the evolution process involves the crossover operator to generate a new vector, which is called the trial vector, in order to increase the diversity of the population. The DE algorithm uses the binomial crossover operator, which combines the components of the target vector \(\mathbf {x}_i^{\left( G \right) }\) and the corresponding mutant vector \(\mathbf {v}_i^{\left( G +1 \right) }\) to generate the trial vector \(\mathbf {u}_i^{\left( G +1 \right) }\) according to

$$\begin{aligned} {u}_{i,j}^{\left( G+1 \right) }&= \left\{ \begin{array}{ ll} {v}_{i,j}^{\left( G+1 \right) } &{}\quad \text {if} \, {rand{_{i,j}} \leqslant \mathrm{CR} \vee j = j_{\mathrm{rand}}} \\ {x}_{i,j}^{\left( G \right) } &{}\quad otherwise \end{array} \right. \end{aligned}$$
(32)

where the crossover rate \(\mathrm{CR} \in \left[ {0,1} \right] \) is a pre-defined rate, and \({j_{\mathrm{rand}}}\) is an integer randomly generated from set \(\left\{ {1,2, \ldots ,n} \right\} \) . Thus, if and only if \(rand_{i,j} \leqslant \mathrm{CR}\) or \(j = j_{\mathrm{rand}}\), then the binomial crossover operator copies the jth variable of mutant vector \({\mathbf {v}}_i^{\left( {G+1} \right) }\) to its corresponding element in the trial vector \({\mathbf {u}}_i^{\left( {G+1} \right) }\). Otherwise, the parameter is inherited from the corresponding target vector \({\mathbf {x}}_i^{\left( {G} \right) }\).

6.1.4 Selection

The next phase is selection, in which the DE compares the objective function value of the trial vector \(f\left( {{\mathbf {u}}_i^{\left( G+1 \right) }} \right) \), with corresponding objective function value of the target vector \(f\left( {{\mathbf {x}}_i^{\left( G \right) }} \right) \). If and only if the objective function value of the trial vector \(f\left( {{\mathbf {u}}_i^{\left( G+1 \right) }} \right) \) is less than or equal to the objective function value of the target vector \(f\left( {{\mathbf {x}}_i^{\left( G \right) }} \right) \), then \({\mathbf {u}}_i^{\left( {G + 1} \right) }\) is set to \({\mathbf {x}}_i^{\left( {G + 1} \right) }\) . Otherwise, the old target vector \({\mathbf {x}}_i^{\left( {G} \right) }\) is retained. Therefore, the selection operator can be defined as follows:

$$\begin{aligned} \mathbf {x}_{i}^{\left( G+1 \right) }= \left\{ \begin{array}{ ll} \mathbf {u}_{i}^{\left( G +1 \right) } &{}\quad \text {if} \, {f\left( {{\mathbf {u}}_i^{\left( G+1 \right) }} \right) \leqslant f\left( {{\mathbf {x}}_i^{\left( G \right) }} \right) } \\ \mathbf {x}_{i}^{\left( G \right) } &{} \quad \mathrm{otherwise} \end{array} \right. , \end{aligned}$$
(33)

where \(\mathbf {x}_{i}^{\left( G+1 \right) }\) is the target vector selected for the next generation.

The algorithm is repeated until the termination criterion (the maximum number of generations \(G_{max}\)) is satisfied.

6.2 Improved DE algorithm

In this paper, the powerful efficient DE variants named JADE and SHADE have been proposed with the aim to provide the dynamically adapted control parameters for each individual in each generation of evolution according to Cauchy distribution for \(F_i^{\left( G \right) }\) and normal distribution for \(\mathrm{CR}_i^{\left( G \right) }\) (Zhang and Sanderson 2009; Tanabe and Fukunaga 2013). Therefore, the control parameters \(F_i^{\left( G \right) }\) and \(\mathrm{CR}_i^{\left( G \right) }\) for each individual \(\mathbf{{x}}_i^{\left( G \right) }\) in each generation of evolution are independently generated, as follows:

$$\begin{aligned} F_i^{\left( G \right) } = rand{c_i}\left( {\mu _F^{\left( G \right) },0.1} \right) , \end{aligned}$$
(34)
$$\begin{aligned} \mathrm{CR}_i^{\left( G \right) } = rand{n_i}\left( {\mu _{\mathrm{CR}}^{\left( G \right) },0.1} \right) , \end{aligned}$$
(35)

where \(rand{c_i}\left( {\mu _F^{\left( G \right) },0.1} \right) \) is a random variable distributed according to the Cauchy distribution with median \(\mu _F^{\left( G \right) }\) and scale parameter 0.1, and \(rand{n_i}\left( {\mu _{\mathrm{CR}}^{\left( G \right) },0.1} \right) \) represents a random number generated from normal distribution with mean \(\mu _{\mathrm{CR}}^{\left( G \right) }\) and standard deviation 0.1.

The parameters \(\mu _F^{\left( 0 \right) }\) and \(\mu _{\mathrm{CR}}^{\left( 0 \right) }\) are initialized to 0.5. If \(F_i^{\left( G \right) } > 1\) , then \(F_i^{\left( G \right) }\) is set to 1, while when \(F_i^{\left( G \right) } < 0\), \(F_i^{\left( G \right) } \) can be regenerated using Eq. (34). Similarly, if \(\mathrm{CR}_i^{\left( G \right) } > 1,\) then \(\mathrm{CR}_i^{\left( G \right) }\) is truncated to 1; otherwise if \(\mathrm{CR}_i^{\left( G \right) }\) is less than 0, then \(\mathrm{CR}_i^{\left( G \right) }\) is set to 0.

In each generation, all successful values of \(F_i^{\left( G \right) }\) and \(\mathrm{CR}_i^{\left( G \right) }\) are saved in the sets \({S_F}\) and \({S_{\mathrm{CR}}}\), respectively. Hence, at the end of each generation, the parameters \(\mu _F^{\left( G \right) }\) and \(\mu _{\mathrm{CR}}^{\left( G \right) }\) can be updated as below:

$$\begin{aligned} \mu _F^{\left( G +1 \right) }= & {} \left\{ {\begin{array}{*{20}{l}} {\left( {1 - c} \right) \mu _F^{\left( G \right) } + c \cdot \mathrm{mean}_{{\mathrm{WL}}}\left( {{S_F}} \right) }&{}\quad {if\, {S_F} \ne \emptyset }\\ {\left( {1 - c} \right) \mu _F^{\left( G \right) } + c \cdot rand\left( {0,1} \right) }&{}\quad {otherwise} \end{array}} \right. ,\nonumber \\ \end{aligned}$$
(36)
$$\begin{aligned} \mu _{\mathrm{CR}}^{\left( G +1\right) }= & {} \left\{ {\begin{array}{*{20}{l}} {\left( {1 - c} \right) \mu _{\mathrm{CR}}^{\left( G \right) } + c \cdot \mathrm{mean}_{{\mathrm{WA}}}\left( {{S_{\mathrm{CR}}}} \right) }&{}\quad {if\, {S_{\mathrm{CR}}} \ne \emptyset }\\ {\left( {1 - c} \right) \mu _{\mathrm{CR}}^{\left( G \right) } + c \cdot rand\left( {0,1} \right) }&{}\quad {otherwise} \end{array}} \right. ,\nonumber \\ \end{aligned}$$
(37)

where c is the learning rate, which is set to \(c=0.1\) (Zhang and Sanderson 2009). Here, the \(\mathrm{mean}_{{\mathrm{WL}}}\left( \cdot \right) \) and \(\mathrm{mean}_{{\mathrm{WA}}}\left( \cdot \right) \) denote the weighted Lehmer mean and weighted arithmetic mean, respectively, which can be expressed as follows:

$$\begin{aligned} \mathrm{mean}_{{\mathrm{WL}}}\left( {{S_F}} \right)= & {} \frac{{\sum \nolimits _{k = 1}^L {{w_k} \cdot S_{F,k}^2} }}{{\sum \nolimits _{k = 1}^L {{w_k} \cdot {S_{F,k}}} }}, \end{aligned}$$
(38)
$$\begin{aligned} \mathrm{mean}_{{\mathrm{WA}}}\left( {{S_{\mathrm{CR}}}} \right)= & {} \frac{{\sum \nolimits _{k = 1}^L {{w_k} \cdot S_{\mathrm{CR},k}^2} }}{{\sum \nolimits _{k = 1}^L {{w_k} \cdot {S_{\mathrm{CR},k}}} }}, \end{aligned}$$
(39)

in which \(L = \left| {{S_F}} \right| = \left| {{S_{\mathrm{CR}}}} \right| \) denotes the size of the sets \({S_F}\) and \({S_{\mathrm{CR}}}\) , respectively. Then, the weight \({w_k}\) can be calculated as follows:

$$\begin{aligned} {w_k} = \frac{{\Delta {f_i}}}{{\sum \nolimits _{k = 1}^L {\Delta {f_i}} }}, \end{aligned}$$
(40)

where \(\Delta {f_i} = \left| {f\left( {\mathbf{{u}}_i^{\left( G \right) }} \right) - f\left( {\mathbf{{x}}_i^{\left( G \right) }} \right) } \right| \) denotes the absolute difference between the objective function values of trial vector and target vector.

6.3 The DE/current-to-pbest/1 mutation operator

One of the improved DE variants, named JADE, introduces the effective mutation operator DE/current-to-pbest/1 that accelerates the convergence speed of the algorithm toward the global optimal solution (Zhang and Sanderson 2009). In this regard, the DE/current-to-pbest/1 is used in this work due to its simplicity of implementation, in order to provide more effective search direction during the evolution process. Then, the corresponding mutation vector based on the proposed DE/current-to-pbest/1 mutation operator can be generated, as follows:

$$\begin{aligned} \mathbf {v}_i^{\left( G \right) } = \mathbf{{x}}_i^{\left( G \right) } + F_i^{\left( G \right) }\left( {\mathbf{{x}}_{\mathrm{pbest}}^{\left( G \right) } - \mathbf{{x}}_i^{\left( G \right) }} \right) + F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r1}^{\left( G \right) } - \mathbf{{x}}_{r2}^{\left( G \right) }} \right) ,\nonumber \\ \end{aligned}$$
(41)

where \(F_i^{\left( G \right) }\) is the scale factor for each individual generated from the Cauchy distribution according to Eq. (34) and \(\mathbf{{x}}_{\mathrm{pbest}}^{\left( G \right) }\) is randomly selected from the top \(\left( p \times 100\% \times \mathrm{NP} \right) \) individuals in the current population with \(p \in \left( {0,1} \right] \).

Obviously, from Eq. (41), it can be observed that the mutation vector exploits the nearby region of each \(\mathbf{{x}}_i^{\left( G \right) }\), in the direction of the solution \(\mathbf{{x}}_{\mathrm{pbest}}^{\left( G \right) }\), which further improves the search ability of the algorithm, especially in the later stage of evolution process. Consequently, the global optimal solution can be easily reached for the given optimization problem.

7 Firefly algorithm and the proposed improved version

7.1 Firefly algorithm

The FA is one of the most powerful swarm-based metaheuristic algorithms used for solving multimodal optimization problems, which has been developed by Xin-She Yang (Yang 2010b). In essence, the FA is based on the following three assumptions:

  • All fireflies are unisex.

  • Attractiveness is proportional to their brightness, implying that the less bright firefly will move to the brighter firefly. The firefly will continue to move randomly if no brighter firefly is found in a swarm.

  • The brightness of a firefly represents the objective function value for a given optimization problem. Thus, for a minimization problem, the brightness of each firefly is inversely proportional to the value of the objective function.

The first step in the application of the FA is to randomly generate an initial swarm of NP fireflies over the n-dimensional search space. Generally, in order to properly create the FA, there are two important issues that need to be defined, such as the variation of light intensity and the formulation of attractiveness between fireflies in the swarm.

The light intensity \(I\left( r \right) \) varies with distance, which can be described as the following monotonically decreasing exponential function of the distance

$$\begin{aligned} I\left( r \right) = {I_0}{e^{ - \gamma {r^2}}}, \end{aligned}$$
(42)

where \(I_0\) is the original light intensity and \( \gamma \) is a fixed light absorption coefficient. The parameter \( \gamma \) has a crucial impact on the convergence speed of the algorithm. Theoretically, the parameter \( \gamma \) should take any value from the interval \(\gamma \in \left[ {0,\infty } \right) \). In most applications, it typically varies from 0.1 to 10, and could be usually set to 1.

The attractiveness function \(\beta \left( {{r_{i,j}}} \right) \) between two fireflies i and j is proportional to the light intensity \(I\left( r \right) \), and can be defined as follows:

$$\begin{aligned} \beta \left( {{r_{i,j}}} \right) = {\beta _0}{e^{ - \gamma r_{i,j}^2}}, \end{aligned}$$
(43)

where \({\beta _0} \in \left[ {0,1} \right] \) is the maximum attractiveness at \({r_{i,j}=0}\), i.e., when two fireflies are found at the same point of search space.

The Euclidean distance \({r_{i,j}}\), between any pair of fireflies i and j at positions \({\mathbf{{x}}_i}\) and \({\mathbf{{x}}_j}\) , respectively, can be obtained as follows:

$$\begin{aligned} {r_{i,j}} = {\left\| {{\mathbf{{x}}_i} - {\mathbf{{x}}_j}} \right\| _2} = \sqrt{\sum \limits _{k = 1}^n {{{\left( {{x_{i,k}} - {x_{j,k}}} \right) }^2}} } . \end{aligned}$$
(44)

Here, \({x_{i,k}}\) and \({x_{j,k}}\) are the kth element of the ith and jth firefly positions within the search space, respectively.

The search process of FA depends on the attraction between fireflies in the swarm. In this regard, the movement of the less bright firefly i toward another brighter firefly j determines the new position \(\mathbf{{x}}_i^{\left( {G + 1} \right) }\) at \(\left( G + 1 \right) \)th iteration, which can be generated as follows:

$$\begin{aligned} \mathbf{{x}}_i^{\left( {G + 1} \right) } = \mathbf{{x}}_i^{\left( G \right) } + \beta \left( {{r_{i,j}}} \right) \left( {\mathbf{{x}}_j^{\left( G \right) } - \mathbf{{x}}_i^{\left( G \right) }} \right) + \alpha {\varvec{\varepsilon }}_i^{\left( G \right) }. \end{aligned}$$
(45)

Hence, Eq. (45) consists of three terms. The first term gives the current position of the ith firefly, the second term is related to the attraction, while the third term represents the random motion of ith firefly. According to Eq. (45), the randomization parameter \(\alpha \) is normally selected within range \(\left[ {0,1} \right] \), where the \( {\varvec{\varepsilon }}_i^{\left( G \right) }\) is a vector of random numbers defined by the Gaussian distribution, i.e., \( {\varvec{\varepsilon }}_i^{\left( G \right) } = \left( {rand_{i,j} - 0.5} \right) \), and \({rand_{i,j}}\) is a random number uniformly distributed in \(\left[ {0,1} \right] \).

7.2 Improved version of firefly algorithm

The conventional FA suffers from the problem of being trapped in local optimum of a multimodal objective function, which can lead to premature convergence. In this regard, to solve the highly nonlinear and multimodal optimization problems, there is a need to modify the FA (Yang 2009; Wang et al. 2017).

According to Eq. (43), it is obvious that at the early stage of the evolution process, the distance between fireflies and the global optimal solution is extremely large; thus, the attractiveness tends to be zero and becomes nearly non-existent. Then, the movement of fireflies is based on the random walk through the search space, so the convergence of the algorithm is not guaranteed. On the other hand, with the increase in generations, the distance between fireflies and the global optimal solution is extremely small; thus, the attractiveness remains constant. Therefore, the modification in the search process of the FA is still needed, in order to enhance the optimization performance in aspect of balance between the global exploration and local exploitation abilities.

In order to provide an appropriate balance between exploration and exploitation of the search space of the conventional FA, this paper proposed the improved attractiveness function \({\tilde{\beta }} \left( {{r_{i,j}}} \right) \) , which can be created, as follows:

$$\begin{aligned} {\tilde{\beta }} \left( {{r_{i,j}}} \right) = {\beta _0}{e^{ - \gamma r_{i,j}^2\frac{G}{{{G_{\max }}}}}}\frac{{{f_{\mathrm{best}}}}}{{{f_{\mathrm{mean}}}}}, \end{aligned}$$
(46)

where \(f_{\mathrm{mean}}\) and \(f_{\mathrm{best}}\) are the mean and best values of the objective function in the previous generation, respectively.

Based on Eq. (46), the updated firefly position can be rewritten as:

$$\begin{aligned} \mathbf{{x}}_i^{\left( {G + 1} \right) } = \mathbf{{x}}_i^{\left( G \right) } + {\tilde{\beta }} \left( {{r_{i,j}}} \right) \left( {\mathbf{{x}}_j^{\left( G \right) } - \mathbf{{x}}_i^{\left( G \right) }} \right) + \alpha {\varvec{\varepsilon }}_i^{\left( G \right) }. \end{aligned}$$
(47)

Simulation results obtained for attractiveness functions \(\beta \left( {{r_{i,j}}} \right) \) and \({\tilde{\beta }} \left( {{r_{i,j}}} \right) \) according to Eq. (43) (dashed line) and Eq. (46) (solid line), respectively, during the evolution process are presented in Fig. 3.

Fig. 3
figure 3

Comparison between the attractiveness functions \(\beta \left( {{r_{i,j}}} \right) \) and \({\tilde{\beta }} \left( {{r_{i,j}}} \right) \) during the evolution process

As presented in Fig. 3, in the early stage of the evolution process, the attractiveness function \( \beta \left( {{r_{i,j}}} \right) \) defined in Eq. (43) tends to be zero, then, it oscillates between 0 and 1.0, which can slow down the search process, and therefore, the convergence of the algorithm is not guaranteed.

On the contrary, the proposed attractiveness \({\tilde{\beta }} \left( {{r_{i,j}}} \right) \) defined in Eq. (46) gradually increases with growth of generations, in order to enhance the ability of finding the global optimal solution, as shown in Fig. 3. This modification helps to maintain the diversity within the swarm and improves the convergence speed of the algorithm in order to create balance between the global exploration and local exploitation ability of the proposed algorithm.

Therefore, in the evolution process, the improved attractiveness \({\tilde{\beta }} \left( {{r_{i,j}}} \right) \) is adaptively increased and significantly outperforms the original attractiveness \(\beta \left( {{r_{i,j}}} \right) \) in search ability.

8 Adaptive hybrid firefly differential evolution algorithm

Based on the above considerations, this section introduces the AHFADE algorithm, based on the hybridization of the DE with the FA. The main goal of the proposed algorithm is to create an efficient search strategy by incorporating mutation operators of the DE into the FA, to achieve the global optimal solution with high accuracy and fast convergence. The mutation operator DE/rand/1 has strong global exploration with slow convergence speed, while the DE/current-to-pbest/1, which has been proposed in JADE (Zhang and Sanderson 2009), has more suitable local exploitation ability that leads to faster convergence speed. Therefore, in order to enhance the diversity and convergence of the FA, and to prevent the algorithm from falling into local optimum of a multimodal objective function, the following two modifications are proposed.

The first modification is performed to improve the exploration of FA using the mutation operator DE/rand/1 by replacing the term \(\alpha {{\varvec{\varepsilon }}_i}\) in Eq. (47) with \(F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r2}^{\left( G \right) } - \mathbf{{x}}_{r3}^{\left( G \right) }} \right) .\) For implementing the first modification, the \({F_{i }^{\left( G \right) }}\) from Eq. (34) is substituted in Eq. (47) and \({\mathbf{{x}}_j^{(G)}}\) in Eq. (47) is replaced with \(\mathbf{{x}}_{r_1}^{(G)}\) from Eq. (29). Then, the expression for updating the new position of the ith firefly is given as follows:

$$\begin{aligned} \mathbf{{x}}_i^{\left( {G + 1} \right) }&= \mathbf{{x}}_i^{\left( G \right) } + {\tilde{\beta }} \left( {{r_{i,j}}} \right) \left( {\mathbf{{x}}_{{r_1}}^{\left( G \right) } - \mathbf{{x}}_i^{\left( G \right) }} \right) \nonumber \\&+ F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r2}^{\left( G \right) } - \mathbf{{x}}_{r3}^{\left( G \right) }} \right) . \end{aligned}$$
(48)

After mathematical transformations, Eq. (48) can be rewritten as follows:

$$\begin{aligned} \mathbf{{x}}_i^{\left( {G + 1} \right) }&= \left( {1 - {\tilde{\beta }} \left( {{r_{i,r_1}}} \right) } \right) \mathbf{{x}}_i^{\left( G \right) } + {\tilde{\beta }} \left( {{r_{i,r_1}}} \right) \mathbf{{x}}_{{r_1}}^{\left( G \right) } \nonumber \\&\quad + F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r2}^{\left( G \right) } - \mathbf{{x}}_{r3}^{\left( G \right) }} \right) , \end{aligned}$$
(49)

where \({r_{i,{r_1}}}\) is the distance between fireflies i and \({r_1}\) at the positions \(\mathbf{{x}}_i^{\left( G \right) }\) and \(\mathbf{{x}}_{{r_1}}^{\left( G \right) },\) respectively.

The second modification is aimed to improve the exploitation of the FA using the mutation operator DE/current-to-pbest/1. To achieve this, the third term in Eq. (49) is replaced with \(F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r1}^{\left( G \right) } - \mathbf{{x}}_{r2}^{\left( G \right) }} \right) \) and \(\mathbf{{x}}_{{r_1}}^{\left( G \right) }\) is substituted with \(\left( {\mathbf{{x}}_{\mathrm{pbest}}^{\left( G \right) } - \mathbf{{x}}_i^{\left( G \right) }} \right) \), which is given in Eq. (41). As a result, the new expression for updating the new position of the ith firefly can be expressed as:

$$\begin{aligned} \mathbf{{x}}_i^{\left( {G + 1} \right) }&= \left( {1 - {\tilde{\beta }} \left( {{r_{r_{i,\mathrm{pbest}}}}} \right) } \right) \mathbf{{x}}_i^{\left( G \right) } + {\tilde{\beta }} \left( {{r_{r_{i,\mathrm{pbest}}}}} \right) \left( {\mathbf{{x}}_{\mathrm{pbest}}^{\left( G \right) } - \mathbf{{x}}_i^{\left( G \right) }} \right) \nonumber \\&\quad + F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r1}^{\left( G \right) } - \mathbf{{x}}_{r2}^{\left( G \right) }} \right) , \end{aligned}$$
(50)

where \({r_{i,\mathrm{pbest}}}\) is the distance between fireflies i and pbest at the positions \(\mathbf{{x}}_i^{\left( G \right) }\) and \(\mathbf{{x}}_{\mathrm{pbest}}^{\left( G \right) },\) respectively.

Based on the above considerations, an adaptive parameter is proposed in this work that relies on the normalized objective function values to automatically switch between global exploration and local exploitation, which can be expressed as:

$$\begin{aligned} \Delta = 1 - \left| {\frac{{f_{\mathrm{mean}} - f_{\mathrm{best}}}}{{f_{\mathrm{worst}}^{} - f_{\mathrm{best}}^{}}}} \right| , \end{aligned}$$
(51)

where \(f_{\mathrm{mean}}\) , \(f_{\mathrm{best}}\) and \(f_{\mathrm{worst}}\) are the mean, best and worst objective function values of individuals in the previous generation, respectively.

Obviously, the balance between global exploration and local exploitation can be controlled by changing the value of \(\Delta \). Therefore, according to Eq. (51), when the value of \(\Delta \) is close to 0, the population diversity is low, which corresponds to global exploration of the search space. Hence, Eq. (49) can be applied in order to explore the search space and locate the region of the global optimum. On the other hand, when the value of \(\Delta \) is close to 1, the population diversity is high, which corresponds to local exploitation of the search space. In this process, Eq. (50) can be employed to improve the local search ability and enhance convergence speed. Therefore, in the evolution process, the new position of ith firefly can be updated based on adaptive parameter \(\Delta \), as shown below:

$$\begin{aligned} \mathbf{{v}}_i^{\left( {G + 1} \right) }={\left\{ \begin{array}{ll} \left( {1 - {\tilde{\beta }} \left( {{r_{i,{r_1}}}} \right) } \right) \mathbf{{x}}_i^{\left( G \right) } + {\tilde{\beta }} \left( {{r_{i,{r_1}}}} \right) \mathbf{{x}}_{r1}^{\left( G \right) } \\ + F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r2}^{\left( G \right) } - \mathbf{{x}}_{r3}^{\left( G \right) }} \right) , \quad \Delta \le \omega \\ \left( {1 - {\tilde{\beta }} \left( {{r_{i,\mathrm{pbest}}}} \right) } \right) \mathbf{{x}}_i^{\left( G \right) } + {\tilde{\beta }} \left( {{r_{i,\mathrm{pbest}}}} \right) \left( {\mathbf{{x}}_{\mathrm{pbest}}^{\left( G \right) } - \mathbf{{x}}_i^{\left( G \right) }} \right) \\ + F_i^{\left( G \right) }\left( {\mathbf{{x}}_{r1}^{\left( G \right) } - \mathbf{{x}}_{r2}^{\left( G \right) }} \right) , \quad otherwise \end{array}\right. }\nonumber \\ \end{aligned}$$
(52)

where \(\omega = 0.5\) is the predefined threshold value.

These modifications have significant influence on the convergence speed and population diversity of the proposed algorithm. In this manner, the exploration and exploitation abilities of the proposed algorithm are strengthened. Additionally, the crossover and selection operators are the same as in the conventional DE algorithm and can be easily implemented into the structure of the proposed AHFADE algorithm.

The flowchart of the proposed AHFADE algorithm for the passive target localization problem is illustrated in Fig.4.

Fig. 4
figure 4

The flowchart of the proposed AHFADE algorithm

The methodology of the proposed AHFADE algorithm, described above, can be divided into following steps:

  1. Step 1

    Initialize parameters

    Initialize parameters of the proposed AHFADE algorithm, including the population size, the maximum number of generations, the lower and upper bounds.

  2. Step 2

    Generate the initial population of fireflies

    The position of fireflies in the swarm is initialized uniformly and randomly in the n-dimensional search space.

  3. Step 3

    Compute light intensity

    Compute light intensity of each firefly in current generation.

  4. Step 4

    The following process is repeated until the stopping criterion is reached

  5. Step 5

    Calculate the distance between fireflies

    Determine the distance \({r_{ij}}\) between any two fireflies \({\mathbf{{x}}_i}\) and \({\mathbf{{x}}_j}\) in the swarm, according to Eq. (44).

  6. Step 6

    Sort the population

    The fireflies are sorted according to objective function values.

  7. Step 7

    Update the proposed attractiveness \({\tilde{\beta }} \left( {{r_{i,j}}} \right) \)

    The proposed attractiveness \({\tilde{\beta }} \left( {{r_{i,j}}} \right) \) is updated according the normalized objective function values, as defined in Eq. (46).

  8. Step 8

    Compute the adaptive scale factor \(F_i^{\left( G \right) }\)

    Using Eq. (34) update the value of adaptive scale factor \(F_i^{\left( G \right) }.\)

  9. Step 9

    Determine the adaptive crossover rate \(\mathrm{CR}_i^{\left( G \right) }\)

    Compute the adaptive crossover rate \(\mathrm{CR}_i^{\left( G \right) }\) according to Eq. (35).

  10. Step 10

    Calculate the value of the adaptive parameter \(\Delta \)

    Calculate the value of the adaptive parameter \(\Delta \) using Eq. (51).

  11. Step 11

    Check parameter \(\Delta \) and move the fireflies

    If \(\Delta \le \omega \), then Eq. (49) is applied; otherwise, when \(\Delta > \omega \), Eq. (50) is used.

  12. Step 12

    Crossover operator

    The crossover operator is applied to the population according to Eq. (32).

  13. Step 13

    Selection operator

    The selection operator defined in Eq. (33) is applied.

  14. Step 14

    Stopping criterion

    Stop the AHFADE algorithm if a termination criterion is satisfied and continue to Step 15; otherwise, go to Step 4 and repeat the search for the global optimal solution.

  15. Step 15

    Display of the results

    Display the best solution found and its corresponding objective function value.

Finally, the AHFADE algorithm can be implemented for solving the passive localization problem. Consequently, Fig. 5 shows the position of fireflies during evolution process with 60 fireflies in the generation.

Fig. 5
figure 5

The position of fireflies during the evolution process

Based on the initially created swarm, the proposed algorithm guides the swarm of fireflies in each generation toward better solution, as shown in Fig. 5. It is clearly seen that at the end of the evolution process, the swarm of fireflies will gather around the global optimal solution in the search space, and as a result, the estimated position of the target is found, which is shown in the same figure. In this way, the proposed AHFADE algorithm uses 60 generations to satisfy the stopping condition and obtain the global optimal solution of the considered ML objective function.

9 Cramér–Rao lower bound

In the passive target localization problem, the CRLB establishes a lower bound on the covariance matrix and provides a useful tool for evaluating the performance of any unbiased estimator (Shen et al. 2012). The CRLB is obtained from the diagonal elements of the inverse of the Fisher information matrix (FIM), denoted by \(\mathbf {I\left( x \right) }\). Based on the probability density function \(f\left( {{\tilde{\mathbf{d}}} \left| \mathbf {x} \right. } \right) \) given in Eq. (6), the FIM of the proposed target localization system is given as follows:

$$\begin{aligned} \mathbf {I\left( x \right) }&= E\left[ {\left( {\frac{{\partial \ln \left( {f\left( {\left. {\tilde{\mathbf{d}}} \right| \mathbf {x}} \right) } \right) }}{{\partial \mathbf {x}}}} \right) {{\left( {\frac{{\partial \ln \left( {f\left( {\left. {\tilde{\mathbf{d}}} \right| \mathbf {x}} \right) } \right) }}{{\partial \mathbf {x}}}} \right) }^T}} \right] \nonumber \\&= - E\left[ {\frac{{{\partial ^2}\ln \left( {f\left( {\left. {\tilde{\mathbf{d}}} \right| \mathbf {x}} \right) } \right) }}{{\partial \mathbf {x}\partial {\mathbf {x}^T}}}} \right] . \end{aligned}$$
(53)

Then, the FIM can be written as:

$$\begin{aligned} \mathbf {I\left( x \right) }= \begin{bmatrix} {{I_{xx}}}&{}{{I_{xy}}} \\ {{I_{yx}}}&{}{{I_{yy}}} \end{bmatrix}, \end{aligned}$$
(54)

where the corresponding elements are defined as follows:

$$\begin{aligned}&{I_{xx}}= \sum \limits _{i = 1}^N {\frac{1}{{\sigma _{ni}^2}}{{\left( {\frac{{x - {x_i^r}}}{{{{\left\| {\mathbf {x} - \mathbf {x}_i^r} \right\| }_{\text {2}}}}} + \frac{x}{{{{\left\| \mathbf {x} \right\| }_{\text {2}}}}}} \right) }^2}}, \end{aligned}$$
(55)
$$\begin{aligned}&{I_{xy}} = {I_{yx}} = \sum \limits _{i = 1}^N {\frac{1}{{\sigma _{ni}^2}}\left( {\frac{{x - {x_i^r}}}{{{{\left\| {\mathbf {x} - \mathbf {x}_i^r} \right\| }_{\text {2}}}}} + \frac{x}{{{{\left\| \mathbf {x} \right\| }_{\text {2}}}}}} \right) } \nonumber \\&\quad \times \left( {\frac{{y - {y_i^r}}}{{{{\left\| {\mathbf {x} - \mathbf {x}_i^r} \right\| }_{\text {2}}}}} + \frac{y}{{{{\left\| \mathbf {x} \right\| }_{\text {2}}}}}} \right) , \end{aligned}$$
(56)
$$\begin{aligned}&{I_{yy}} = \sum \limits _{i = 1}^N {\frac{1}{{\sigma _{ni}^2}}{{\left( {\frac{{y - {y_i^r}}}{{{{\left\| {\mathbf {x} - \mathbf {x}_i^r} \right\| }_{\text {2}}}}} + \frac{y}{{{{\left\| \mathbf {x} \right\| }_{\text {2}}}}}} \right) }^2}}. \end{aligned}$$
(57)

Therefore, the inverse of the FIM is expressed as follows:

$$\begin{aligned} \mathbf {I}^{-1}= \frac{1}{{\det \left( \mathbf {I} \right) }} \begin{bmatrix} {{I_{yy}}}&{} -{{I_{xy}}} \\ {- {I_{yx}}}&{}{{I_{xx}}} \end{bmatrix}. \end{aligned}$$
(58)

Derivations of Eqs. (55)–(57) are given in Appendix A. The CRLB of passive target location vector components is given by the rth diagonal element of the inverse of the FIM, as follows:

$$\begin{aligned} \text {CRLB}\left( {{{\left[ \mathbf {x} \right] }_r}} \right) = {\left[ {{\mathbf {I}^{ - 1}}} \right] _{r,r}}, \end{aligned}$$
(59)

where \({\left[ {{\mathbf {I}^{ - 1}}} \right] _{r,r}}\) is the rth row and the rth column element of the corresponding matrix \(\mathbf {I}^{-1}\) and \(r=1,2\). Hence, the CRLB on the variance of an unbiased estimator \(\hat{\mathbf {x}}\) is expressed as:

$$\begin{aligned} E\left[ {\left( {\hat{\mathbf {x}}-\mathbf {x}} \right) \,{{\left( {\hat{\mathbf {x}}-\mathbf {x}} \right) }^T}} \right] \geqslant {{\mathrm{Tr}}}\left( {{\mathbf {I}^{ - 1}}\left( \mathbf {x} \right) } \right) , \end{aligned}$$
(60)

where \(\hat{\mathbf {x}}\) is the estimated value of \({\mathbf {x}}\).

10 Experimental study

In this section, experiments are conducted to evaluate the localization performance and to perform the statistical comparison of the AHFADE algorithm with other well-known algorithms on CEC2014 benchmark problems. In this regard, the obtained numerical results are outlined into two subsections, which are described below.

10.1 Statistical comparisons of the proposed AHFADE algorithm on CEC2014 benchmark

In this section, experiments are carried out to compare the proposed AHFADE with well-known state-of-the-art algorithms, such as SHADE (Tanabe and Fukunaga 2013), jDE (Brest et al. 2006), FA (Yang 2010b) and Lévy-flight FA (LFA) (Yang 2010a) algorithms on CEC2014 benchmark. The formulation of test functions and optimization problems used in this analysis has been described in (Liang et al. 2013). The CEC2014 benchmark considers the single objective real-parameter numerical optimization and consists of 30 test functions, where the dimensions of the search space are \(D = 10, 30, 50 \) and 100. Based on the characteristics of the objective functions defined in CEC2014 benchmark problem, the 30 test functions can be categorized into the following groups:

  • \({f_1} {-} {f_3}\) represent optimization problems with unimodal objective functions,

  • \({f_4} {-} {f_{16}}\) denote simple multimodal objective functions,

  • \({f_{17}} {-} {f_{22}}\) are hybrid functions, where the variables are divided into subcomponents, in which each component is defined using a different basic function,

  • \({f_{23}} {-} {f_{30}}\) denote composition functions, which combine the properties of the sub-functions while maintaining the continuity in the region of the global optimum.

In order to perform statistical analysis and compare the performance of the proposed AHFADE algorithm, the solution error measure \(\left( {f\left( {{{{\hat{\mathbf{x}}}}}} \right) - f\left( {\mathbf{{x}^*}} \right) } \right) \) has been employed, where \({{{\hat{\mathbf{x}}}}}\) denotes the best solution of the algorithm obtained in one run and \(\mathbf{{x}^*}\) represents the well-known global optimal solution of the corresponding CEC 2014 test function. Experiments for each algorithm on each test function have been independently run for 50 times, and statistical results are provided. According to the recommendations, the termination criteria are set to 10000D (Liang et al. 2013), and the population size is set to \({N_P} = 100\) (Tanabe and Fukunaga 2013). For all considered problems, the search space is bounded by \({\left[ {\begin{array}{*{20}{c}} { - 100}&{100} \end{array}} \right] ^D}.\) To perform evaluation, from a statistical point of view, the obtained solutions have been analyzed and compared using two nonparametric statistical hypothesis tests, e.g., Wilcoxon signed-rank test and Friedman test.

Firstly, the Wilcoxon signed-rank test has been employed to determine whether the first algorithm obtained statistically better solution compared to the second algorithm. In the analysis, the Wilcoxon signed-rank test has been applied with significance level \(\alpha = 0.05\). Therefore, in the results, the \({R^ + }\) denotes the sum of ranks where the first algorithm outperformed the second, and \({R^ - }\) represents the sum of ranks for which the first algorithm performed worse than the second algorithm. Furthermore, in the corresponding columns designated with signs \(+\), \(\approx \), − (better, equal and worse), we have denoted the number of test functions where the first algorithm is better than, equal to, or worse than the second algorithm, respectively. Based on (Derrac et al. 2011), for the null hypothesis of the Wilcoxon signed-rank test, it is assumed that ”there is no difference between the mean results of the two samples”. On the other hand, the alternative hypothesis states that ”there is a difference in the mean results of the two samples”. Therefore, p value has been employed in the statistical analysis and compared to the significance level \(\alpha \). Thus, the null hypothesis can be rejected, when the p value is less than or equal to the \(\alpha = 0.05.\) The compared results have been indicated with the following three signs \((+,-,\approx )\) according to the results of the statistical test. In this regard, the plus \((+)\) sign designates that the first algorithm is significantly better than the second, minus \((-)\) sign means that the first algorithm is significantly worse than the second, and approximation \((\approx )\) sign denotes that there is no significant difference between two algorithms.

Secondly, the Friedman test has been employed to obtain the ranks of all algorithms over every test function on all considered dimensions D of the search space, in order to find the significant differences between the performance of two or more algorithms. In this respect, the algorithm with the lowest rank denotes the best performing algorithm, while the algorithm which has the highest rank is considered as the worst. According to (Derrac et al. 2011), for the null hypothesis of the Friedman test, it is assumed that ”there is no difference among the performance of all algorithms”, whereas the alternative hypothesis states that ”there is a difference among the performance of all algorithms”.

The computational results of the proposed AHFADE and other considered algorithms on CEC2014 benchmark problems are presented in Table 1, where ”Mean” represents the mean value of the objective function of the best individual, ”STD” represents the standard deviation, and ”Sign” denotes whether the considered algorithm has better (\(+\)), similar (\(\approx \)) or worse (−) performance compared to the proposed AHFADE algorithm. From the numerical results in Table 1, it can be observed that the proposed AHFADE algorithm obtains the best ”Mean” and ”STD” values on most test functions compared to other considered algorithms. However, on hybrid functions \(f_{17}\) and \(f_{21}\), and composite function \(f_{28}\) the AHFADE algorithm has the similar performance to jDE algorithm over the dimensions \(D=30,50\) and 100. Compared to the jDE and SHADE algorithms, the AHFADE algorithm achieves the similar performance on functions \(f_{19}\) and \(f_{4}\) on dimensions \(D=50\) and \(D=30\), respectively, as well as on composite functions \(f_{26}\) and \(f_{28}\) on \(D=50\). Furthermore, the AHFADE algorithm has worse performance than jDE and SHADE algorithms on unimodal function \(f_{2}\), multimodal function \(f_{7}\) and composite functions \(f_{23}\) and \(f_{24}\) on dimensions of the problem \(D=30, 50\) and 100, respectively.

Table 1 Experimental results on 30 CEC2014 test functions in terms of mean and standard deviation of function value

In comparison with the FA and LFA algorithms, it can be observed that AHFADE significantly outperforms these algorithms on all considered dimensions. However, on multimodal problems \(f_5\), \(f_{12}\) and \(f_{16}\) over the dimensions \(D=50\) and 100, the proposed algorithm achieves worse performance compared to the FA and LFA algorithms. Therefore, it can be concluded that the proposed AHFADE algorithm has clearly the best performance among the SHADE, FA and LFA algorithms, and for higher dimensions of the problem, e.g., \(D=30, 50\) and 100, it achieves similar performance to jDE algorithm, especially on composite functions. In this way, the proposed algorithm successfully finds promising solutions on all problems and all dimensions.

In order to perform the statistical comparison of the performance of considered algorithms, the results have been analyzed using Wilcoxon signed-rank test. In this regard, in Table 2 the statistical results of applying Wilcoxon’s signed-rand test between AHFADE and jDE, SHADE, FA and LFA algorithms using CEC2014 test functions on over all dimensions have been shown.

Table 2 Results of the Wilcoxon test between AHFADE and other considered algorithms using CEC2014 for \(D =10, 30, 50\) and 100

From the results in Table 2, it can be observed that for the dimensions \(D=10\) and \(D=30\), the proposed AHFADE algorithm is significantly better than SHADE, FA and LFA algorithms. However, the proposed AHFADE algorithm has higher \(R^+\) values than \(R^-\) when compared to the jDE algorithm on the dimensions \(D=10\) and \(D=30\). On the other hand, for the dimensions \(D=50\) and \(D=100\) the proposed algorithm achieves better performance than FA and LFA algorithms. Additionally, compared to the jDE and SHADE algorithms it is observed that for \(D=50\) and \(D=100\) the AHFADE algorithm provides higher \(R^+\) values than \(R^-\). Therefore, from the results, it can be concluded that the proposed AHFADE algorithm can achieve better performance compared to the individual performances of SHADE and FA algorithms. This shows the effectiveness of the proposed hybridization between DE and FA algorithms which is achieved by introducing the dynamically adapted parameter into the mutation operator, and incorporating the crossover and selection operators of DE algorithm into the FA.

Furthermore, to compare the overall performance of multiple algorithms, the Friedman test has been applied. In this regard, the Table  3 presents the average ranks according to Friedman test, for the proposed AHFADE and other considered algorithms on different CEC2014 benchmark functions for \(D=10, 30, 50\) and 100 dimensions. In this table, the algorithm with the best rank is shown in bold and second best is underlined.

Table 3 Average ranks computed through the Friedman test for all considered algorithms across all functions and over all dimensions using CEC2014, at the significance level 0.05

From the results of the statistical analysis using the Friedman test, given in Table 3, it can be concluded that the proposed AHFADE algorithm achieves the best performance and has the lowest rank among all considered algorithms. Furthermore, it can be observed that the p values computed through Friedman test for all considered test functions in all dimensions are less than significance level \(\alpha = 0.05\). Therefore, based on the hypothesis, it can be concluded that there is a significant difference between the performances of the considered algorithms. According to the statistical analysis, it can be concluded that the proposed AHFADE algorithm has an effective mutation strategy and improved adaptive techniques for setting the control parameter values. Furthermore, the obtained results also show that the hybridization between DE and FA algorithms significantly enhances the performance of the AHFADE algorithm, which demonstrates the effectiveness of the modifications proposed in this paper.

10.2 Localization performance

This section presents the simulation results of the proposed AHFADE algorithm and the well-known algorithms such as SDP, DE and FA, which are performed to verify and compare their performance under the same noisy environment. In the following simulations, the CRLB is employed as a benchmark for evaluating the RMSE performance of the considered algorithms.

The numerical simulations are conducted for the passive target localization system in LOS environment, which consists of one transmitter, four receivers located at known positions and the passive target located at different positions. The evaluation should show how sensitive is performance of all considered algorithms under the noisy TOA measurements. Besides, this work demonstrates the influence of changing the position of the target with respect to the geometry of transmitter and receivers on the localization performance of the existing algorithms.

In this regard, three simulation scenarios are conducted here based on different positions of the target such as: i) the true position of the target lies inside the convex hull of the four receivers; ii) the position of the target is outside of the convex hull of the four receivers; iii) the position of the target is randomly deployed in the \({200\mathrm{{m}}\times 200\mathrm{{m}}}\) square area, and its position varies for each simulation run.

For the proposed passive target localization system, each simulation scenario has the same configuration with transmitter which is located at the origin of the coordinate system and the four receivers forming the convex hull, at: \(\mathbf {x}_1^r = {\left[ {100\;100} \right] ^T}\mathrm{{m}}\), \(\mathbf {x}_2^r = {\left[ {100 \;- 100} \right] ^T}\mathrm{{m}}\), \(\mathbf {x}_3^r = {\left[ { - 100 \;- 100} \right] ^T}\mathrm{{m}},\) and \(\mathbf {x}_4^r = {\left[ { - 100\;100} \right] ^T}\mathrm{{m}}\).

In this way, the localization accuracy is evaluated in terms of RMSE, which can be defined as follows:

$$\begin{aligned} \mathrm{RMSE} = \sqrt{\frac{1}{{{N_m}}}\sum \limits _{n = 1}^{{N_m}} {\left\| {{{{{{\hat{\mathbf{x}}}}}}^{\left( n \right) }} - \mathbf{{x}}} \right\| _2^2} } , \end{aligned}$$
(61)

where \(\mathbf {x}\) is the true target position, \({{{{\hat{\mathbf{x}}}}}^{\left( n \right) }}\) is the estimated target position and \(N_m\) is the number of simulation runs for a given variance of the noise \(\sigma _{ni}^2\). A Monte Carlo simulation, which consists of \(N_m=1000\) runs, is conducted to determine the RMSE performance of the considered algorithms.

Figure 6 shows the simulation results of the comparison between CRLB and the RMSEs of all considered algorithms as a function of the measurement noise \(p = 10\log \left( {\sigma _{ni}^2} \right) \), for the first simulation scenario, when the true position of the target is located at \(\mathbf {x} = {\left[ {20\;30} \right] ^T}\mathrm{{m}}\).

Fig. 6
figure 6

RMSE curves in terms of p of the considered algorithms for the first simulation scenario

As shown in Fig. 6, it can be observed that the RMSE of the proposed AHFADE algorithm attains the CRLB accuracy for all the considered range of p. Furthermore, from this figure it should be noted that the SDP, FA and DE algorithms can achieve performance several dBs above the CRLB. However, numerical results of this simulation show that the SDP method has significant deviation from the CRLB for high values of p (i.e., \(p > 20\,\mathrm{{dB}}\)), compared with other algorithms.

Figure 7 shows the results of the comparison between CRLB and the RMSEs of all considered algorithms in terms of measurement noise p for the second simulation scenario, when the true position of target is located at \(\mathbf {x} = {\left[ {120\;100} \right] ^T}\mathrm{{m}}\).

Fig. 7
figure 7

RMSE curves in terms of p of the considered algorithms for the second simulation scenario

From Fig. 7, it is observed that the proposed AHFADE algorithm can attain the CRLB accuracy in Gaussian noise environment and outperforms all other considered algorithms as p increases. The simulation results also show that the FA and DE algorithms have degradation of localization performance compared to the proposed AHFADE algorithm. It can be also observed that the SDP method has the worst performance, especially when the measurement noise becomes very large ( i.e., \(p \ge 30\,\mathrm{{dB}}\)).

Figure 8 shows the results of comparison between CRLB and the RMSEs of all considered algorithms in terms of measurement noise p for the third simulation scenario, where the true position of the passive target is randomly located within the considered area for each simulation run.

Fig. 8
figure 8

RMSE curves in terms of the p of the considered algorithms for the third simulation scenario

As it could be expected, the results show that the proposed AHFADE algorithm attains the CRLB accuracy for all the considered range of p, and provides superior performance over the FA, DE and SDP algorithms, as shown in Fig. 8.

The comparison between numerical results of the above three simulation scenarios, clearly shows that the RMSE of each considered algorithm is lowest when the true position of the target lies inside the convex hull of the four receivers. On the other hand, the obtained RMSEs of all the considered algorithms are highest when the position of the target is outside of the convex hull of the four receivers. It should be noticed that the proposed AHFADE algorithm has the best performance and its RMSE attains the CRLB accuracy for all the considered range of p in every considered scenario. However, the RMSE performance of the SDP method rapidly deviates from the CRLB, especially in the presence of significant measurement noise.

The average computational time in searching the global optimal solution of each algorithm is another important factor that has a strong influence on the localization performance. To this aim, the average time consumed to obtain the global optimal solution of the considered algorithms can be determined on the same computer with 3.2 GHz CPU and 16 GB of RAM. Based on the proposed simulation scenarios, Table 4 shows the corresponding average computational time to achieve the global optimal solution taken by each algorithm to satisfy the stopping condition.

Table 4 The average computational time of the considered algorithms

It can be seen from Table 4 that the SDP method has the highest average computation time among all examined algorithms, while the FA and DE algorithms have the best average execution time. Generally, it can be concluded that the AHFADE algorithm achieves the best trade-off between the benefit of the localization accuracy and computational complexity, and in this way becomes attractive for applications in sensor networks in noisy environment.

In order to further evaluate the localization performance, the cumulative distribution functions (CDFs) of localization errors of the considered algorithms are conducted for measurement noise \(\sigma _{ni}^2 = 1\;{\mathrm{{m}}^{\mathrm{{2}}}}\) . The localization error (LE) is defined as the Euclidean distance between the estimated and the true position of the passive target, i.e.,

$$\begin{aligned} LE = {\left\| {{{{{{\hat{\mathbf{x}}}}}}^{\left( n \right) }} - \mathbf{{x}}} \right\| _2},\forall n \in \{ 1, \ldots ,{N_m}\}. \end{aligned}$$
(62)

Figure 9 represents the simulation results for the second scenario with the corresponding CDFs in terms of localization error obtained for each algorithm.

Fig. 9
figure 9

CDFs of passive target localization error of the considered algorithms for the second simulation scenario

From Fig. 9, it is observed that the proposed AHFADE algorithm provides the lowest localization error compared to the other considered algorithms. Additionally, it is evident that the AHFADE algorithm has localization error less than \(1.18\; \mathrm{{m}}\) in \(90 \%\) of the cases, while the FA, DE and SDP have localization error less than \(1.4 \;\mathrm{{m}}\), \(1.45 \;\mathrm{{m}}\) and \(1.64 \;\mathrm{{m}}\), respectively. Therefore, the proposed AHFADE algorithm has smaller localization error in terms of CDFs compared to other considered algorithms.

Finally, the influence of the number of receivers N at known positions on the localization accuracy of all algorithms is also considered in this section. In this context, the corresponding coordinates of the ith receiver can be obtained as follows:

$$\begin{aligned} {\mathbf {x}}_i^r = \begin{bmatrix} {R\cos {\varphi _i}} \\ {R\sin {\varphi _i}} \end{bmatrix},\quad \forall i \in \left\{ {1, \ldots ,{N}} \right\} , \end{aligned}$$
(63)

where \(R = 100\sqrt{2} \;\mathrm{{m}}\) is the radius of the circle and the angular separation between receivers can be expressed as \({\varphi _i} = {{2\pi } / i}\) .

Hence, Fig.10 depicts the RMSE curves of all considered algorithms as a function of the number of receivers N, for the first simulation scenario where the target position is set at \(\mathbf {x} = {\left[ {20\;30} \right] ^T}\mathrm{{m}}\) inside the convex hull of receivers, when the variance of measurement noise is \(\sigma _{ni}^2 = 1\;{\mathrm{{m}}^{\mathrm{{2}}}}\).

Fig. 10
figure 10

RMSEs of the considered algorithms as function of the number of receivers

As can be seen from Fig. 10, by increasing the number of receivers from 4 to 10, the RMSEs of all considered algorithms significantly decreased. As shown, the proposed AHFADE algorithm can achieve the CRLB under the given noise environment. It can also be noticed that with increasing the number of receivers from 4 to 12, the proposed AHFADE algorithm provided approximately \(42.77\%\) improvement in localization accuracy.

Based on the above obtained simulation results, it can be concluded that the performance of the proposed AHFADE algorithm shows better robustness to measurement noise than all other considered algorithms in solving the passive target localization problem.

11 Conclusion

In this paper, the passive target localization problem based on the noisy TOA measurements, utilizing multiple receivers and a single transmitter, has been considered and investigated. In this context, an improved robust AHFADE algorithm has been proposed, based on the hybridization of DE and FA, to solve this nonlinear and non-convex localization problem with high accuracy even in highly noisy environments. In the proposed algorithm, a dynamically adapted parameter has been introduced to select the appropriate mutation operator for achieving a proper balance between global exploration and local exploitation during the evolution process. Moreover, the crossover and selection operators of DE algorithm are incorporated into FA, where the control parameters of both DE and FA are dynamically adjusted during the evolution process. To evaluate the performance of the proposed algorithm, the CRLB for the passive target localization problem has been derived. In addition, to compare the optimization performance, the statistical analysis has been conducted between AHFADE and several well-known algorithms on CEC2014 benchmark test problems.

From the numerical results of statistical comparison between AHFADE algorithm and other well-known algorithms, such as SHADE, jDE, FA and LFA, it can be concluded that the hybridization proposed in this paper can improve the overall optimization performance, especially compared to the individual optimization performance of SHADE and FA algorithms. Furthermore, simulation results have shown that the proposed AHFADE algorithm exhibits better robustness in high-noise environments and provides better localization performance in comparison with well-known algorithms such as SDP, DE and FA. Additionally, the results have shown that the AHFADE algorithm can attain the CRLB accuracy and is robust against changes in the topology. Finally, the AHFADE algorithm can provide a proper balance between localization accuracy and computational complexity, thus making it more attractive for applications in sensor networks in noisy environment.

In the future research, the focus will be on the problem of finding the optimal geometry of the receiver configuration for the passive target localization in the presence of non-line-of-sight (NLOS) propagation. In addition, the research can be extended to parameter sensitivity analysis and influence of parameter variations during the evolution process to further enhance the optimization performance for the given localization problem.