1 Introduction

Space–time adaptive processing (STAP) has been viewed as an effective tool in an airborne surveillance radar since it was first proposed by Brennan et al. [1]. It adaptively suppresses the clutter received by the airborne radar from both the temporal domain and spatial domain and also detects moving targets that are relatively weak compared with the surrounding clutter. One of the fundamental problems in STAP is the estimation of the clutter covariance matrix (CCM). As stated in [8], the sample covariance matrix (SCM) is a practical estimator to compute CCM. However, in the full-dimension STAP (f-STAP), SCM requires excessive number of training samples, which is difficult to be realized in the real airborne radar environment. Moreover, f-STAP needs to conduct matrix inverse operation in adaptive processing, which is quite computationally intensive and intractable. These two main limitations motivate the development of STAP algorithms with low computational cost and small training sample demanding. Sub-optimal dimension/rank-reduced STAP algorithms, which maintain nearly the same performance as f-STAP and address the above issues, have been studied in various literature [2,3,4, 6, 7, 9, 12,13,14,15, 17, 18]. Among them, the post-Doppler adaptive processing method, named extended factored approach (EFA or mDT), which is expected to be one of the most efficient and practical STAP algorithms, is mainly concerned in this paper [3, 16].

The basic principle of EFA is that it transforms f-STAP of a NK-dimensional adaptive filtering problem into K separate PN-dimensional adaptive processing problems, where N is the number of antennas, K is the number of pulses, and P is an integer satisfied \(P>\)1. As a consequence, the required number of homogeneous training samples is reduced to 2PN according to the RMB rule [10]. Besides, the computational cost of EFA is also decreased and the applicability of it is increased in practice. Unfortunately, insufficient homogeneous training sample support is a frequent condition in the airborne radar system with large-scale antenna array elements, which results in considerable degeneration of the clutter suppression ability of EFA.

Actually, nowadays the persymmetry property of CCM is usually used as the a priori knowledge to increase the training sample support for accurate CCM estimation. It is initially exploited for communications. Tong et al. [12] extended the application of persymmetry to EFA (Per-EFA) and showed that the required number of training samples is reduced. Per-EFA is demonstrated to be an effective algorithm in training-limited scenarios. However, the computational cost of Per-EFA in adaptive processing increases because of the extra computational cost in CCM estimation. Besides, it still requires excessive quantity of training samples in the airborne radar system with large spatial degrees of freedom (DoFs).

In order to reduce both the computational cost and the training sample requirement in the post-Doppler adaptive processing, a bi-iterative algorithm based on Per-EFA is proposed in this paper. Firstly, according to the persymmetric property, CCM is estimated by using the original data, the constructed spatial transformed data, the constructed temporal transformed data and the constructed spatial–temporal transformed data. Secondly, the spatial weight vector in EFA is decomposed as the Kronecker products of two short weight vectors. Then the cost function of EFA is transformed into the cost function with multiple weight vectors. Finally, the bi-iterative algorithm [11] is exploited to obtain the desired weight vectors.

Notation Throughout this paper, matrix and vector are represented by boldface uppercase letter and boldface lowercase letter, respectively. Superscripts \(^{T}\), \(^{\mathrm {*}}\) and \(^{H}\) separately denote transpose, conjugate and conjugate transpose of a matrix or a vector. \(\otimes \) and \(\circ \) denote the Kronecker product operator and Hadamard product operator, respectively. A \(c\times c\) identity matrix is represented as \(\mathbf{I}_c \). \(\left\| {\bullet } \right\| _F \) denotes the F-norm of a matrix, and \([{\bullet }]^{-1}\) means the inverse of a matrix.

2 STAP Fundamentals

In this paper, we assume that the uniform linear antenna array is paralleled to the ground. The configuration of a traditional airborne phased-array radar system to be studied is shown in Fig. 1. The airborne radar system moves at a constant speed \(v_a \) along the positive x-direction with N antenna array elements and processes K pulses on the repetition of \(T_r \) in one coherent processing interval (CPI). Let the crab angle between the antenna array line and the flight velocity be \(\theta _{c}\). The azimuth angle and elevation angle at the ith clutter scatter in the lth range cell are denoted as \(\theta _i \) and \(\varphi _l \), respectively.

Fig. 1
figure 1

The airborne phased-array radar system

For a given range cell, the clutter, generated by the terrain scatterings, can be considered as a collection of independent scatters as [8]

$$\begin{aligned} {\mathbf{z}}(l)=\sum _{i=1}^{N_c } \beta _i {\mathbf{a}}_{{\mathrm{t}}} (\theta _i ,\varphi _l )\otimes {\mathbf{a}}_{{\mathrm{s}}}(\theta _i ,\varphi _l), \end{aligned}$$
(1)

where \(N_c \) scatters are divided in an iso-range cell, \(\beta _i \) is clutter echo’s random complex amplitude, \(\mathbf{a}_\mathrm{t} (\theta _i ,\varphi _l )\) and \(\mathbf{a}_\mathrm{s} (\theta _i ,\varphi _l )\) are temporal steering-vector and spatial steering-vector, respectively. They can be expressed as

$$\begin{aligned} \mathbf{a}_\mathrm{t} (\theta _i ,\varphi _l )=[1,\exp (j2\pi \omega _i ),\ldots ,\exp (j2\pi (K-1)\omega _i )]^{T}, \end{aligned}$$
(2)
$$\begin{aligned} \mathbf{a}_\mathrm{s} (\theta _i ,\varphi _l )=[1,\exp (j2\pi \upsilon _i ),\ldots ,\exp (j2\pi (N-1)\upsilon _i )]^{T}, \end{aligned}$$
(3)

where \(\omega _i =2T_r v_a \cos (\theta _i+\theta _c )\cos \varphi _l /\lambda \) is the normalized Doppler frequency, \(\upsilon _i =d\cos \theta _i \cos \varphi _l /\lambda \) is the normalized spatial frequency, \(\lambda \) is the operating wavelength, and d is the array element spacing. The received clutter signal also contains noise, namely

$$\begin{aligned} \mathbf{x}(l)=\mathbf{z}(l)+\mathbf{n}, \end{aligned}$$
(4)

where \(\mathbf{n}\) is the Gaussian noise vector. Let \(\mathbf{s}\in \mathbf{C}^{KN\times 1}\) denote the normalized target vector. The objective of f-STAP is to maintain the output energy of the target while minimizing that of clutter. This implies the following cost function

$$\begin{aligned} \left\{ {\begin{array}{l} \min \ E\left[ \left| {\mathbf{w}^H \mathbf{x}(l)} \right| ^{2}\right] \\ \hbox {s.t.}\ \mathbf{w}^{H} \mathbf{s}=1 \\ \end{array}} \right. . \end{aligned}$$
(5)

The Lagrange multiplier methodology is adopted to obtain the optimal weight

$$\begin{aligned} \mathbf{w}=\mathbf{R}_x^{-1} \mathbf{s}/\left( \mathbf{s}^{H}{} \mathbf{R}_x^{-1} \mathbf{s}\right) , \end{aligned}$$
(6)

where \(\mathbf{R}_x =E[\mathbf{x}(l)\mathbf{x}^{H}(l)]\) is the CCM. In general, since the statistics of the range cell under test is never known, CCM is replaced by the SCM estimator, namely \({\tilde{\mathbf{R}}}_x=\frac{1}{L}\sum \nolimits _{l=1}^L {\mathbf{x}(l)\mathbf{x}^{H}(l)} \), where L is the number of training samples obtained in adjacent range cells. RMB rule shows that, in the Gaussian noise environment, if the number of homogeneous training samples exceeds twice the dimension of CCM, the output SCNR (signal-to-clutter-plus-noise ratio) loss will be within 3dB. The training sample requirement of f-STAP is evidently unpractical. Moreover, as the inverse operation of a \(KN\times KN\) matrix, the computational cost of f-STAP is unbearable of approximate \(O(K^{3}N^{3})\). These two major issues prevent f-STAP from being put into effect. Hence, EFA, one of the most efficient and practical STAP algorithms, is mainly concerned in this paper. It performs Doppler filtering on data of each array element and then spatial filtering separately within P (\(P\ge 2 )\) Doppler bins. Generally, P is set to 3 in EFA. Therefore, the EFA with three Doppler bins is referred in the following. Let \(\mathbf{f}_k =\exp (-\,j2\pi (k-1)[0,1,2,\ldots ,K-1]^{T}\hbox {/}K)/\hbox {sqrt}(K){\begin{array}{ll} ,&{} {k=1,\ldots ,K} \\ \end{array} }\), which are actually the coefficients of discrete Fourier transform and compose the Doppler filter group, be the Doppler filter coefficient vector. Then the EFA dimension-reduced transformation matrix according to the kth Doppler bin can be expressed as

$$\begin{aligned} \mathbf{B}_k =[\mathbf{f}_{k-1} ,\mathbf{f}_k ,\mathbf{f}_{k+1} ]^{T}\otimes \mathbf{I}_N . \end{aligned}$$
(7)

The dimension-reduced space–time data and target can be achieved as

$$\begin{aligned} \begin{array}{l} {\hat{{\mathbf{x}}}}_k (l)=\mathbf{B}_k \mathbf{x}(l) \\ {\hat{{\mathbf{s}}}}_k=\mathbf{B}_k \mathbf{s} \\ \end{array}. \end{aligned}$$
(8)

The cost function in EFA is

$$\begin{aligned} \left\{ {\begin{array}{l} \min \ E\left[ \left| {\hat{{\mathbf{w}}}_k^H {\hat{{\mathbf{x}}}}_k (l)} \right| ^{2}\right] \\ \hbox {s.t.}\ {\hat{{\mathbf{w}}}}_k^H {\hat{{\mathbf{s}}}}_k =1 \\ \end{array}} \right. . \end{aligned}$$
(9)

Similarly, the solution to Eq. (9) is obtained by using the Lagrange multiplier methodology and the CCM is estimated by

$$\begin{aligned} {\hat{{\mathbf{R}}}}_{\hat{{x}}_k }=\frac{1}{L}\sum _{l=1}^L {{\hat{{\mathbf{x}}}}_k(l){\hat{{\mathbf{x}}}}_k^H (l)} . \end{aligned}$$
(10)

Evidently, the dimension of CCM is reduced to 3N. As a consequence, the required number of training samples and computational cost in adaptive processing are separately reduced to 6N and \(O( 3 ^{3}N^{3})\). Although EFA reduces the training sample requirement, its clutter suppression ability can still be heavily degraded in large-scale antenna array elements system.

3 The Bi-iterative Algorithm Based on Per-EFA

Through the persymmetric property of CCM, Ref. [12] increases the number of training samples by constructing the spatial transformed data, the temporal transformed data and spatial–temporal transformed data. Let \(f_{k-1} \), \(f_k \), \(f_{k\hbox {+1}} \) be the normalized Doppler frequency of three adjacent Doppler bins, \(\mathbf{U}_\mathrm{t} =\text {diag}([\text {e}^{j2\pi f_{k-1} (N-1)},\text {e}^{j2\pi f_k (N-1)},\text {e}^{j2\pi f_{k+1} (N-1)}])\), \(\mathbf{U}_\mathrm{st} =\text {diag}([\text {e}^{-j2\pi f_{k-1} (N-1)},\text {e}^{-j2\pi f_k (N-1)},\text {e}^{-j2\pi f_{k+1} (N-1)}])\) and \(\mathbf{J}_\mathrm{s} =\text {fliplr}(\mathbf{I}_N )\) with \(\text {fliplr}({\bullet })\) being the MATLAB operator that flips matrix in left/right direction. Then CCM can be estimated by using the constructed data as

(11)

where

$$\begin{aligned} \mathbf{B}_\mathrm{s} =\mathbf{I}_p \otimes \mathbf{J}_\mathrm{s} ,\quad \mathbf{B}_\mathrm{t} =\mathbf{U}_\mathrm{t} \otimes \mathbf{I}_N ,\quad \mathbf{B}_\mathrm{st} =\mathbf{U}_\mathrm{st} \otimes \mathbf{J}_\mathrm{s}, \end{aligned}$$
(12)

in which \(\mathbf{B}_\mathrm{s} \) is the spatial transformation matrix that exploits the persymmetric property of CCM in the spatial domain, \(\mathbf{B}_\mathrm{t} \) is the temporal transformation matrix that exploits the persymmetric property of CCM in the temporal domain, and \(\mathbf{B}_\mathrm{st} \) is the spatial–temporal transformation matrix that jointly exploits the persymmetric property of CCM in the spatial and temporal domains.

Then, the proposed bi-iterative adaptive method will be introduced after constructing the CCM estimator by Eq. (11). As a matter of fact, \({\hat{{\mathbf{x}}}}_k (l)=[\hat{{x}}_1 (l),\hat{{x}}_2 (l),\ldots ,\hat{{x}}_{PN} (l)]\) can be rearranged as the matrix with the following formula

$$\begin{aligned} {\hat{{\mathbf{X}}}}_k (l)\hbox {=}\left[ {{\begin{array}{llll} {\hat{{X}}_{1,1} (l),}&{}\quad {\hat{{X}}_{1,2} (l),}&{}\quad {\ldots ,}&{}\quad {\hat{{X}}_{1,b} (l)} \\ {\hat{{X}}_{2,1} (l),}&{}\quad {\hat{{X}}_{2,2} (l),}&{}\quad {\ldots ,}&{}\quad {\hat{{X}}_{2,b} (l)} \\ \vdots &{}\quad \vdots &{}\quad \vdots &{}\quad \vdots \\ {\hat{{X}}_{a,1} (l),}&{}\quad {\hat{{X}}_{a,2} (l),}&{}\quad {\ldots ,}&{}\quad {\hat{{X}}_{a,b} (l)} \\ \end{array} }} \right] , \end{aligned}$$
(13)

where \(\hat{{X}}_{n,m} (l)=\hat{{x}}_{(n-1)\times b+m} (l), n=1,\ldots ,a, m=1,\ldots ,b\) and \(a\times b=P\times N\) with scalars a and b both being integers. The objective function in Eq. (9) can be transformed into the following formula by using Eq. (13)

$$\begin{aligned} E\left[ \left| {\hat{{\mathbf{w}}}}_k^H {\hat{{\mathbf{x}}}}_k (l)\right| ^{2}\right] =E\left[ \left\| {\hat{{\mathbf{W}}}}_k^{\mathrm{*}} \circ {\hat{{\mathbf{X}}}}_k (l)\right\| _F^2 \right] =E\left[ \left| {\sum _{m=1}^b{\sum _{n=1}^a {\hat{{w}}_{n,m}^*\hat{{X}}_{n,m} (l)} } }\right| ^{2}\right] . \end{aligned}$$
(14)

According to the separation of variables, let weight coefficient \(\hat{{w}}_{n,m} =u_n v_m \) be expressed as the separable form and be introduced into Eq. (14), and the following bi-quadratic cost function can be acquired

$$\begin{aligned} E\left[ \left| {\sum _{m=1}^b {\sum _{n=1}^a {\hat{{w}}_{n,m}^{*} \hat{{X}}_{n,m} (l)} } } \right| ^{2}\right]= & {} E\left[ \left| {\sum _{m=1}^b {\sum _{n=1}^a {u_n^{*} \hat{{X}}_{n,m} (l)v_m^{*} } } } \right| ^{2}\right] \nonumber \\= & {} E\left[ \left| {\mathbf{u}^{H}{\hat{{\mathbf{X}}}}_k (l)\mathbf{v}^{{*}}} \right| ^{2}\right] =E\left[ \left| {(\mathbf{u}\otimes \mathbf{v})^{H}{\hat{{\mathbf{x}}}}_k (l)} \right| ^{2}\right] ,\nonumber \\ \end{aligned}$$
(15)

where \(\mathbf{u}\in \mathbf{C}^{a\times 1}\) and \(\mathbf{v}\in \mathbf{C}^{b\times 1}\) represent the separable weight vectors. Therefore, after substituting \({\hat{{\mathbf{R}}}}_{\hat{{x}}_k }\) by , Eq. (9) can be converted into the following cost function

(16)

Equation (16) means to minimize the output energy of the clutter while maintain that of the target by jointly optimizing the weight vectors \(\mathbf{u}\) and \(\mathbf{v}\). By using the Lagrange multiplier methodology, Eq. (16) can be transformed into the following unconstrained one

(17)

where \(\mu \) is the Lagrange multiplier. Generally, the solution to Eq. (17) is not easy to be analytically expressed. Fortunately, the bi-iterative algorithm [11] can be conveniently applied to numerically solve this cost function. And the procedure of the algorithm is as follows.

To obtain the numerical solution to Eq. (17), \(\mathbf{v}\) is initialized with \(\mathbf{v}(0)\). Then Eq. (17) becomes an unconstrained cost function with respect to \(\mathbf{u}\) and \(\mu \) after substituting \(\mathbf{v}(0)\) into it. Let the gradient of \(J(\mathbf{u},\mathbf{v}(0))\) with respect to \(\mathbf{u}\) and \(\mu \) be zero, the following iterative value can be obtained

$$\begin{aligned} \mathbf{u}(1)=\mathbf{R}_u^{-1} \mathbf{s}_u /\left( \mathbf{s}_u^H \mathbf{R}_u^{-1} \mathbf{s}_u \right) , \end{aligned}$$
(18)

where

(19)

Again, with fixing \(\mathbf{u}(1)\), let the gradient of \(J(\mathbf{u}(1),\mathbf{v})\) with respect to \(\mathbf{v}\) and \(\mu \) be zero, and the iterative value of \(\mathbf{v}(1)\) can be achieved by

$$\begin{aligned} \mathbf{v}(1)=\mathbf{R}_v^{-1} \mathbf{s}_v /\left( \mathbf{s}_v^H \mathbf{R}_v^{-1} \mathbf{s}_v \right) , \end{aligned}$$
(20)

where

(21)

By repeating the above iterative procedure, \(\mathbf{u}(2)\) and \(\mathbf{v}(2)\), \(\mathbf{u}(3)\) and \(\mathbf{v}(3)\), ..., can be obtained in turn. The iterative procedure will be stopped if the termination condition \(\left\| {\mathbf{v}(t)-\mathbf{v}(t-1)} \right\| {/}\left\| {\mathbf{v}(t)} \right\| <\delta \) is satisfied at the tth iteration, where \(\delta \) is the threshold and \(\left\| {\bullet } \right\| \) is the Euclidean norm of a vector. Finally, \({\hat{{\mathbf{w}}}}_k\) is achieved by the Kronecker products of two short vectors

$$\begin{aligned} {\hat{{\mathbf{w}}}}_k=\mathbf{u}(t)\otimes \mathbf{v}(t). \end{aligned}$$
(22)

The dimensions of \(\mathbf{R}_u \) and \(\mathbf{R}_v \) produced in the iterative procedure are \(a\times a\) and \(b\times b\), respectively. After weight decomposition, both the required number of training samples and computational cost in the spatial adaptive processing are reduced. Furthermore, in order to satisfy the small computational cost and training sample requirements, the constraints \(\min \left| {a-b} \right| \) and \(a<b\) are added. Hereafter, for simplicity, the bi-iterative algorithm based on the persymmetry of CCM is denoted as BiPer-EFA.

It is worth noting that \(J(\mathbf{u},\mathbf{v})=J(\alpha \mathbf{u},\alpha ^{-1}{} \mathbf{v})\) is satisfied for any nonzero constant \(\alpha \). In order to eliminate the scale indeterminacy, \(\left\| \mathbf{v} \right\| \) is set to unity in each iteration procedure.

4 Convergence and Computational Cost Analysis

The convergence of the proposed bi-iterative algorithm is verified in this section. Firstly, let \(\left\| \mathbf{v} \right\| =1\), the relation \(J(\mathbf{u}(t),\mathbf{v}(t))=\mathbf{u}^{H}{} \mathbf{R}_u \mathbf{u}\ge \kappa _{\min } \left\| \mathbf{u} \right\| ^{2}\) can be acquired and the set \(\left\{ {\mathbf{u}(t)|\kappa _{\min } \left\| {\mathbf{u}\hbox {(}t\hbox {)}} \right\| ^{2}\le J(\mathbf{u}(t),\mathbf{v}(t))\le c} \right\} \) is bounded for any constant \(0<c<\infty \), where \(\kappa _{\min } \) is the smallest eigenvalue of \(\mathbf{R}_u \). Therefore the set \(\left\{ {\mathbf{u}(t),\mathbf{v}(t)|J(\mathbf{u}(t),\mathbf{v}(t))\le c} \right\} \), which is bounded for any constant \(0<c<\infty \) and iteration number t, can be inferred. Secondly, the cost function \(J(\mathbf{u},\mathbf{v})\) in Eq. (17) is continuous for its differentiability. Thirdly, \(J({\varvec{u}}(t-1),{\varvec{v}}(t-1))\ge J({\varvec{u}}(t),{\varvec{v}}(t))\) since

$$\begin{aligned} J({\varvec{u}}(t-1),{\varvec{v}}(t-1))\ge & {} J({\varvec{u}}(t-1),{\varvec{v}}(t))=\mathop {\min }\limits _v J({\varvec{u}}(t-1),{\varvec{v}})\nonumber \\\ge & {} J({\varvec{u}}(t),{\varvec{v}}(t))=\mathop {\min }\limits _u J({\varvec{u}},{\varvec{v}}(t)) \end{aligned}$$
(23)

The three points just mentioned above satisfy the definition of Lyapunov function [5, 7]. Accordingly, \(J(\mathbf{u},\mathbf{v})\) is a Lyapunov function and the convergence of the bi-iterative algorithm is proved.

Since the computational costs of the Doppler filtering are nearly the same in each method, we mainly consider the computational cost of CCM estimation and matrix inversion in the spatial adaptive processing. The multiplication and division number (MDN) is used as an index of computational cost. First of all, the computational cost of BiPer-EFA is analyzed in detail. It should be noted that the MDN of reducing the data dimension and constructing the transformed data is \(L_1 \left[ {3N^{2}K+3(3N)^{2}} \right] \). In the iteration algorithm, the MDN of CCM estimation and weight vector computation is \(12(a+b)NL_1 +\frac{2}{3}(a^{3}+b^{3})+2(a^{2}+b^{2})+a+b\). As shown in Fig. 5, the bi-iterative algorithm can reach the convergence value within 10 steps. Therefore, the total MDN of BiPer-EFA is

$$\begin{aligned} L_1 \left[ {3N^{2}K+3(3N)^{2}} \right] +10\left[ 12(a+b)NL_1 +\frac{2}{3}(a^{3}+b^{3})+2(a^{2}+b^{2})+a+b\right] .\nonumber \\ \end{aligned}$$
(24)

Similarly, the MDNs of EFA and Per-EFA are

$$\begin{aligned} L_3 \left[ 3N^{2}K+(3N)^{2}\right] +\frac{ 2 }{ 3 }(3N)^{3}+2(3N)^{2}+3N, \end{aligned}$$
(25)
$$\begin{aligned} L_2 \left[ {3N^{2}K+7(3N)^{2}} \right] +\frac{ 2 }{ 3 }(3N)^{3}+2(3N)^{2}+3N, \end{aligned}$$
(26)

respectively, where \(L_1 \), \(L_2 \) and \(L_3 \) are the required number of training samples in each method.

5 Simulation Results

The linear array element mounted on the aircraft is composed of 64 antennas with equally spaced by a distance of \(d=0.1\) m. The echo amplitude of each clutter source is a complex Gaussian random variable and has been weighted by the transmit beam pattern. The other parameters are listed in Table 1. The experimental platform is on a PC notebook with 2.4 GHz Core i7-5500U processor, 4 GB RAM. The algorithms are coded as MATLAB programs and are run in MATLAB R2011a.

Table 1 Parameters of the simulation

One important metric to measure the clutter suppression ability of a STAP method is the SCNR loss, which is defined as the loss between the implemented processor and the optimal one [8]. According to the parameters given above, the weight vector as discussed in Eq. (15) is decomposed as the Kronecker products of two short vectors \(\mathbf{u}\in \mathbf{C}^{12\times 1}\) and \(\mathbf{v}\in \mathbf{C}^{16\times 1}\). Figure 2 shows the normalized SCNR loss (normalized by the maximum SCNR loss among EFA, Per-EFA and BiPer-EFA) curves of EFA, Per-EFA and BiPer-EFA versus the normalized Doppler frequency with small training sample support. For the sake of estimated CCMs being invertible, the number of training samples separately conducted in each method is 30, 50 and 192, which are marked in the parentheses in Fig. 2. In addition, \(\theta _c =0\) that represents the sidelooking scenario and \(\theta _c =\pi /6\) that represents the nonsidelooking scenario are considered in plotting the curves of Fig. 2. It is demonstrated that though less training samples are exploited in BiPer-EFA, it obviously possesses better SCNR loss performance than those of Per-EFA and EFA in both main clutter region and sidelobe clutter region. Figure 3 shows the normalized SCNR loss curves of EFA, Per-EFA and BiPer-EFA versus the normalized Doppler frequency with large training sample support (\(L=500)\) under sidelooking scenario and nonsidelooking scenario. Under this condition, the SCNR performance of BiPer-EFA is a slightly lower than those of Per-EFA and EFA. Evidently, when adequate number of training samples can be acquired, EFA possesses the best SCNR performance.

Fig. 2
figure 2

SCNR loss curves with small training sample support. a \(\theta _c =0\), b \(\theta _c =\pi /6\)

Fig. 3
figure 3

SCNR loss curves with large training sample support (\(L=500)\). a \(\theta _c =0\), b \(\theta _c =\pi /6\)

Fig. 4
figure 4

The comparison of convergence rates versus the number of training samples

The SCNR loss curve, varying with the number of training samples, describes the training sample convergence rate of a STAP method. It also demonstrates the minimum number of training samples required in a STAP method. Figure 4 shows the convergence rates of BiPer-EFA, Per-EFA and EFA with the initial number of training samples individually being 10, 50 and 192 at normalized Doppler frequency 0.3 and normalized spatial frequency 0. Evidently, BiPer-EFA enjoys far faster training sample convergence rate than its counterparts. In other words, BiPer-EFA owns lower training sample requirement than Per-EFA and EFA. The experimental result of iterative convergences of BiPer-EFA with five different random initializations is shown in Fig. 5. The optimal value in Fig. 5 means the SCNR loss value obtained by the optimal spatial weight. The BiPer-EFA with all the five initializations can be convergent within 10 iterations, which demonstrates its fast iterative convergence rate. However, all the five convergent values are around \(-\,5\) dB, which probably means that the bi-iterative algorithm may be attracted by the local minimum. Besides, when the number of training sample is large enough, the SCNR loss performances of EFA and Per-EFA are superior to that of BiPer-EFA. The probable reason for this result is that when adequate number of training samples is available, Per-EFA and EFA can achieve the optimal SCNR values, while BiPer-EFA can only achieve an inferior SCNR value for the limitation of its special structure. Through the experimental results shown above, the advantage of BiPer-EFA over Per-EFA and EFA under small training sample support is exhibited.

Fig. 5
figure 5

The iterative convergence rate with five different initializations

According to Eqs. (24), (25) and (26), the computational cost comparison is shown in Fig. 6. Here, \(L_1 =30\), \(L_2 =100\) and \(L_3 =250\) are adopted since each method can acquire nearly the same SCNR loss with the chosen number of training samples. Compared with EFA and Per-EFA, one conclusion is obtained that BiPer-EFA can achieve the smallest computational cost when \(N>10\). According to the simulation result, the computational cost of BiPer-EFA is almost 2 times smaller than that of Per-EFA and that of EFA when the array element scale approaches large. In addition, for the additional CCM estimation by Eq. (11), the computational cost of Per-EFA is a little higher than that of EFA.

Fig. 6
figure 6

Computational cost comparison

The processing times for drawing the curves of each method in Fig. 2 are compared in Table 2. Obviously, in accordance with the aforementioned computational cost analysis, BiPer-EFA possesses the minimum processing time among all methods. In Table 3, the processing times for drawing the curves of each method in Fig. 3 are compared. When \(L_1 =L_2 =L_3 =500\), BiPer-EFA achieves the maximum computational cost, while EFA achieves the minimum computational cost. As a consequence, when adequate number of training samples can be obtained, EFA enjoys better clutter suppression ability and smaller computational cost.

Table 2 Processing time comparison under small training sample support
Table 3 Processing time comparison under large training sample support

6 Conclusion

In this paper, an algorithm that decreases the training sample requirement and computational cost has been proposed. The persymmetry property of CCM is utilized, and the weight vector of EFA is decomposed. Then, the bi-iterative algorithm is applied to find the desired weight vectors. Experimental results show that the proposed algorithm has superiority in training sample convergence rate and computational cost in the airborne radar system with large antenna array.