1 Introduction

Wireless sensor network (WSN) has been adopted as one of the most potential applications in many research fields, such as environmental monitoring, factory instrumentation, power transmission and distribution systems, and military surveillance [1]. In many applications, the objective is to estimate a desired unknown parameter employing measurements collected from the sensor nodes and the distributed solution is usually used to estimate the parameter. The distributed estimation extracts the information collected from the sensors distributed in an area. The distribution of the sensors, in addition to the temporal dimension, provides spatial diversity to improve the robustness of the processing tasks [2]. In the distributed solution, each sensor node relies on its local measurements, communicates with its immediate neighbors, and the processing is distributed between all sensors. In this solution, the processing and communications measurements are significantly decreased.

Recently, distributed adaptive estimation approaches (or adaptive networks) have been proposed to solve the linear estimation problem in a distributed and adaptive manner. The distributed adaptive approach extends the adaptive filters to the network domain, which is performed without prior knowledge of data statistics. The adaptive networks’ performance strongly depends on the cooperation modes of incremental and diffusion [3]. In the incremental scheme, a cyclic path is needed, and the sensors communicate with their neighbors inside the path. The incremental LMS algorithm [4], incremental RLS algorithm [5], incremental affine projection-based adaptive (APA) algorithm [6] are distributed adaptive estimations that use incremental cooperation. When more communications and energy resources are accessible, a cooperative diffusion mode is utilized. In this mode, each node communicates with all its neighbors as managed by the network topology, and it is not required any cyclic path. The diffusion LMS [7], diffusion RLS algorithm [8], and diffusion APA [9] are distributed adaptive estimation algorithms that use diffusion cooperation.

In most of the previous studies of distributed adaptive estimation algorithms [4,5,6,7,8,9], the tap-length of the adaptive filters is constant for each node. This issue is not proper in general because the optimal value of tap-length is variable or unknown. The tap-length of a linear filter significantly influences its performance. The deficient tap-length likely increases the mean-squared error (MSE), whereas the excess mean-squared error (EMSE) and computational cost will be high when the tap-length is too large. Hence, there is an optimal value for tap-length to balance the conflicting requirements of performance and complexity. Furthermore, in many applications, the optimal tap-length may be time-variant. So, a variable tap-length algorithm is required to determine the optimal length. In the single filter domain, many algorithms have been introduced for this purpose [10,11,12,13,14,15,16,17,18,19,20,21]. The most efficient algorithm among them is the FT algorithm [21]. This algorithm, like the LMS and its variants [22,23,24] is simple and has an excellent performance. Thus, it is considered a popular algorithm, and in the adaptive networks domain, it is more suitable than all variable tap-length algorithms. In [25], the FT algorithm has developed in the context of incremental learning for distributed networks. Two recursive algorithms run concurrently, one for estimating the tap-length and the other for estimating the tap-weights. The authors in [25] only provide a steady-state analysis to examine the steady-state performance of the tap-weight vector. This issue is performed by deriving the theoretical expressions for the mean-squared deviation (MSD), MSE, and EMSE for each node within the network. For this purpose, the authors assume a fixed tap-length at the steady-state scenario, but they have not provided any theoretical expression for this steady-state tap-length. In this paper, we investigate a theoretical expression for steady-state tap-length achieved in the incremental structure tap-length adaption for each node. The achieved results indicate that this algorithm overestimates the optimal tap-length. Simulation results validate the derived theoretical expressions.

The remains of the paper are organized as follows. In Sect. 2, we provide the background that includes the estimation problem and its incremental-based solution. In part  3, we analyze the performance of distributed incremental variable FT LMS algorithm. In this part, we also calculate the mathematical expression for the steady-state tap-length at each sensor. The comparison of the simulation results and theoretical analyses is given in Sect. 4, while conclusion remarks are presented in Sect. 5.

Notation: For ease of reference, the main symbols used in this paper are listed in Table 1.

Table 1 The main symbols

2 Incremental FT-LMS algorithm

We consider N sensor nodes spatially distributed in a region. The objective is to estimate the coefficients and the length \(L_{opt}\) of an unknown vector \(\mathbf{w}^o_{L_{opt}}\) from multiple measurements obtained from N sensors of the network. The adaption rules for the coefficients and the length of \(\mathbf{w}^o_{L_{opt}}\) are decoupled, where the selection of each node does not depend on the other. As we will present later, the tap-length L is calculated using the tap-length estimation algorithm. We also suppose that each node k has access to time realizations \(\{d_k(i), \mathbf{u}_{k,i}\}\) of zero-mean spatial data \(\{d_k,\mathbf{u}_k\}\), where \(d_k\) and \(\mathbf{u}_k\) are respectively the scalar and \(1 \times L\) row regression vector.

By collecting the measurement data and regression into global matrices, the following definitions are made:

$$\begin{aligned} \begin{array}{l} \mathbf{d}\triangleq col \{ d_1, d_2,\dots , d_N\} \ \ \ \ \ (N\times 1) \\ \mathbf{U}\triangleq col \{\mathbf{u}_1,\mathbf{u}_2,\dots \mathbf{u}_N\} \ \ \ \ \ (N\times L) \end{array} \end{aligned}$$
(1)

First, we consider the unknown parameter coefficients estimation. The aim is the estimation of \(L\times 1\) vector \(\mathbf{w}\) that solves \(arg \ min {\mathop {}_\mathbf{w}{J_L(\mathbf{w})}}\), where

$$\begin{aligned} J_L(\mathbf{w})=E\{{\Vert \mathbf{d}-\mathbf{{Uw}}\Vert }^2\} \end{aligned}$$
(2)

Let \({\varvec{\psi } }^{(i)}_k\) be the local estimation of the unknown parameter at node k in time instant i. In [4], a distributed incremental LMS (DILMS) approach is reported to solve the above optimization problem as follows:

$$\begin{aligned}&{\varvec{\psi }}^{(i)}_k={\varvec{\psi }}^{(i)}_{k-1}+\mu _k\mathbf{u}^*_{k,i}(d_k(i)-\mathbf{u}_{k,i}{\varvec{\psi }}^{(i)}_{k-1})\nonumber \\&\quad k \in {\mathcal {N}} \end{aligned}$$
(3)

where the parameter \(\mu _k\) is the proper selected positive local step-size. Each sensor k emploies the local data \(d_k(i)\), \(\mathbf{u}_{k,i}\) and \({\varvec{\psi }}^{(i)}_{k-1}\) received from the sensor \(k-1\). Then, \({\varvec{\psi }}^{(i)}_N\) is used as the initial condition for the next time at sensor node \(k=1\).

Considering the tap-length L, the segmented cost function is defined as [25]:

$$\begin{aligned} J^{(L)}_M(\mathbf{w}) \triangleq E \{{\Vert \mathbf{d}-\mathbf{U}_M\mathbf{w}_M \Vert }^2 \} \end{aligned}$$
(4)

where \(1 \le M \le L\), \(\mathbf{U}_M\) and \(\mathbf{w}_M\) respectively contain the initial M column vectors of \(\mathbf{U}\) and M elements of \(\mathbf{w}\) such as:

$$\begin{aligned}&\mathbf{w}_M \triangleq col\{w(1),w(2),\dots , w(M)\} \ \ \ \ \ \ \ (1\times M) \nonumber \\&\quad \mathbf{U}_M \triangleq col \{\mathbf{u}_1(1:M),\mathbf{u}_2(1:M),\dots , \mathbf{u}_N(1:M)\}\nonumber \\&\quad (N\times M) \end{aligned}$$
(5)

where w(j) is the j’th element of \(\mathbf{w}\) and \(\mathbf{u}_k(1:M)\) consists of the M initial elements of \(\mathbf{u}_k\). In [25], to explore the optimal length \(L_{opt}\), the smallest difference of MSE estimation is proposed as:

$$\begin{aligned} min \{\ L \ | \ J^{(L)}_{L-\Delta }(\mathbf{w})-J^{(L)}_L(\mathbf{w})\le \varepsilon \} \end{aligned}$$
(6)

where \(\varepsilon \) is a small positive value and it is predetermined by the system requirements. The integer value \(\Delta \) prevents the tap-length to be suboptimal.

In [25], to solve the optimization problem (6) a solution by developing the FT algorithm within the context of incremental cooperation has been proposed. With the assumption that \({\ell }_{k,f}(i)\) denotes the local estimation of the FT at sensor k in time instant i, in the specific topological circle, sensor k receives the calculated FT \({\ell }_{k-1,f}(i)\) from sensor \(k-1\) and updates its local estimation as:

$$\begin{aligned}&{\ell }_{k,f}(i)=({\ell }_{k-1,f}(i)-{\alpha }_k)-{\gamma }_k\nonumber \\&\quad \left[ {(e^{(L_k(i))}_{k,L_k(i)}({\varvec{\psi }}^{(i)}_{k-1}))}^2-{(e^{(L_k(i))}_{k,L_k(i)-\Delta }({\varvec{\psi }}^{(i)}_{k-1}))}^2\right] \end{aligned}$$
(7)

where

$$\begin{aligned} e^{(L_k(i))}_{k,L_k(i)}\left( {\varvec{\psi }}^{(i)}_{k-1}\right) =d_k(i)-\mathbf{u}_{k,i}{\varvec{\psi }}^{(i)}_{k-1} \end{aligned}$$
(8)

and

$$\begin{aligned}&e^{(L_k(i))}_{k,L_k(i)-\Delta }\left( {\varvec{\psi }}^{(i)}_{k-1}\right) =d_k(i)\nonumber \\&\quad -\mathbf{u}_{k,i}(1:L_k(i)-\Delta ){\varvec{\psi }}^{(i)}_{k-1}(1:L_k(i)-\Delta ) \end{aligned}$$
(9)

In which \(\mathbf{u}_{k,i}(1:L_k(i)-\Delta )\) and \({\varvec{\psi }}^{(i)}_{k-1}(1:L_k(i)-\Delta )\) consist of the initial \(L_k(i)-\Delta \) elements of \(\mathbf{u}_{k,i}\) and \({\varvec{\psi }}^{(i)}_{k-1}\) respectively. In (7) \({\alpha }_k\) denotes the local leakage factor and \({\gamma }_k\) indicates the local step-size for \({\ell }_{k,f}(i)\) adaption at sensor k. \({\ell }_{k,f}(i)\) is no longer forced to be an integer value, and the local estimation of integer tap-length \(L_k(i)\) is achieved from the FT as follows:

$$\begin{aligned} L_{k+1}(i)= {\left\{ \begin{array}{ll} \lfloor {\ell }_{k,f}(i)\rfloor &{} if \ \ |L_k(i)-{\ell }_{k,f}(i)| \ge {\delta }_k \\ L_k(i) &{} o.w \end{array}\right. } \end{aligned}$$
(10)

where \(\lfloor .\rfloor \) indicates the floor operation, which rounds down to the nearest integer, and \({\delta }_k\) is a small integer. Also the tap-length during its evolution is conditioned to be not less than a lower floor value \(L_{min}\), where \(L_{min}>\Delta \) (usually \(L_{min}={\Delta }+{\delta }_k+1\)), i.e., when the tap-length fluctuates under \(L_{min}\), it is set to \(L_{min}\). The necessity of this operation is due to using \(L_k(i)-\Delta \) as a tap-length in the incremental FT algorithm.

With \(L_k(i)\), the tap coefficients are then updated by (3) in which the length of vectors \(\mathbf{u}_{k,i}\) and \({\varvec{\psi }}^{(i)}_{k-1}\) are adjusted by \(L_k(i)\).

3 Steady-state analysis of incremental variable FT LMS networks

In this part, we intend to obtain the mathematical expression for the steady-state tap-length of incremental FT algorithm for each sensor node. First, we assume a linear model as:

$$\begin{aligned} d_k(i)=\mathbf{u}_{L_{opt}k,i}{} \mathbf{w}^o_{L_{opt}}+v_k(i) \end{aligned}$$
(11)

where \(v_k(i)\) is temporally and spatially white noise element with zero mean and variance \({\sigma }^2_{v,k}\) and independent of \(\mathbf{u}_{L_{opt}\ell ,j}\) and \(d_{\ell }(j)\) for all \(\ell \), j. In (11), \(\mathbf{u}_{L_{opt}k,i}\) is a vector of length \(L_{opt}\). Since in this scenario, the length of the vectors varies in each iteration, so it is convenient to define an upper bound for the length and pad the vectors by zeros to achieve this length. So, we obtain vectors with the same length that simplifies the calculations. Thus, we define the upper bound for length \(L_{ub}\) as:

$$\begin{aligned} L_{ub}>max \{L_{opt}, L_k(i)\} \ \ \ \ \text {for} \ \ \ \text {all} \ \ \ i \end{aligned}$$
(12)

By this definition, using \(\mathbf{w}_{L_{ub}}\) we denote the unknown parameter \(\mathbf{w}^\mathbf{o}_{L_{opt}}\), where \(\mathbf{w}_{L_{ub}}\) is obtained by padding \(\mathbf{w}^\mathbf{o}_{L_{opt}}\) with \(L_{ub}-L_{opt}\) zeros. In this way, we define the vector \({\varvec{\psi }}^{(i)}_{L_{ub},k-1}\) that is obtained by padding \({\varvec{\psi }}^{(i)}_{k-1}\) with \(L_{ub}-L_k(i)\) zeros. Now we partition the unknown parameter \(\mathbf{w}_{L_{ub}}\) into three parts as:

$$\begin{aligned} \left[ \begin{array}{l} \dot{\mathbf{w}} \\ \ddot{\mathbf{w}} \\ \mathop {\mathbf{w}}\limits ^{...} \end{array} \right] \end{aligned}$$
(13)

where \(\dot{\mathbf{w}}\) is the part modeled by \({\dot{\varvec{\psi }}}^{(i)}_{k-1}\), and \(\ddot{\mathbf{w}}\) is the part modeled by \({\ddot{\varvec{\psi }}}^{(i)}_{k-1}\) such that:

$$\begin{aligned} \begin{array}{l} \dot{\varvec{\psi }}^{(i)}_{k-1}={\varvec{\psi }}^{(i)}_{k-1}(1:L_k(i)-\Delta ) \\ {\ddot{\varvec{\psi }}}^{(i)}_{k-1}={\varvec{\psi }}^{(i)}_{k-1}(L_k(i)-\Delta +1:L_k(i)) \end{array} \end{aligned}$$
(14)

where \({\varvec{\psi }}^{(i)}_{k-1}(1:L_k(i)-\Delta )\) consists of the initial \(L_k(i)-\Delta \) elements of \({\varvec{\psi }}^{(i)}_{k-1}\) and \({\varvec{\psi }}^{(i)}_{k-1}(L_k(i)-\Delta +1:L_k(i))\) consists of the last \(\Delta \) elements of \({\varvec{\psi }}^{(i)}_{k-1}\) and \(\mathop \mathbf{w}\limits ^{...}\) is the undermodeled part. Now we define the vector \(\overline{\varvec{\psi }}^{(i)}_{L_{ub},k-1}\), that computes the difference among the estimated weight at sensor \(k-1\) and the desired solution \(\mathbf{w}_{L_{ub}}\) as:

$$\begin{aligned} \overline{\varvec{\psi }}^{(i)}_{L_{ub},k-1}=\mathbf{w}_{L_{ub}}-{\varvec{\psi }}^{(i)}_{L_{ub},k-1} \end{aligned}$$
(15)

and it is partitioned into three parts as:

$$\begin{aligned} \begin{array}{c} \overline{\varvec{\psi }}^{(i)}_{L_{ub},k-1}=\mathbf{w}_{L_{ub}}-{\varvec{\psi }}^{(i)}_{L_{ub},k-1}=\left[ \begin{array}{c} \dot{\mathbf{w}} \\ \ddot{\mathbf{w}} \\ \mathop \mathbf{w}\limits ^{...} \end{array} \right] -\left[ \begin{array}{l} \dot{\varvec{\psi }}^{(i)}_{k-1} \\ \ddot{\varvec{\psi }}^{(i)}_{k-1} \\ \mathbf{0}_{(L_{ub}-L_k(i))\times 1} \end{array} \right] \\ =\left[ \begin{array}{l} \dot{\mathbf{w}}-{\dot{\varvec{\psi }}}^{(i)}_{k-1} \\ \ddot{\mathbf{w}}-{\ddot{\varvec{\psi }}}^{(i)}_{k-1} \\ \mathop \mathbf{w}\limits ^{...} \end{array} \right] =\left[ \begin{array}{l} \dot{\overline{\varvec{\psi }}}^{(i)}_{k-1} \\ \ddot{\overline{\varvec{\psi }}}^{(i)}_{k-1} \\ {\mathop {\overline{\varvec{\psi }}}\limits ^{...}}^{(i)}_{k-1} \end{array} \right] \end{array} \end{aligned}$$
(16)

For convenience of analysis, similar to \(\mathbf{w}_{L_{ub}}\) and \(\overline{\varvec{\psi }}^{(i)}_{L_{ub},k-1}\), we partition the regression vector \(\mathbf{u}_{L_{ub},k,i}\) with length \(L_{ub}\) into three parts \(\dot{\mathbf{u}}_{k,i}\), \(\ddot{\mathbf{u}}_{k,i}\) and \({\mathop {{\varvec{u}}}\limits ^{...}}_{k,i}\). Considering the above notations and replacing (11) and (15) into (8) and (9) and padding all the vectors in (8), (9) and (11) with zeros to make their length equal to \(L_{ub}\), we have:

$$\begin{aligned} \begin{array}{l} e^{(L_k(i))}_{k,L_k(i)}({\varvec{\psi }}^{(i)}_{k-1})=\mathbf{u}_{L_{opt}k,i}{} \mathbf{w}^o_{L_{opt}}-\mathbf{u}_{k,i}{\varvec{\psi }}^{(i)}_{k-1}+v_k(i) \\ =\mathbf{u}_{L_{ub},k,i} \left[ \begin{array}{l} \mathbf{w}^{o}_{L_{opt}} \\ \mathbf{0}_{(L_{ub}-L_{opt})\times 1} \end{array} \right] -\mathbf{u}_{L_{ub},k,i} \left[ \begin{array}{l} {\varvec{\psi }}^{(i)}_{k-1} \\ \mathbf{0}_{(L_{ub}-L_k(i))\times 1} \end{array} \right] \\ +v_k(i) \\ =\mathbf{u}_{L_{ub},k,i}{} \mathbf{w}_{L_{ub}}-\mathbf{u}_{L_{ub},k,i}{\varvec{\psi }}^{(i)}_{L_{ub},k-1}+v_k(i) \\ =\mathbf{u}_{L_{ub},k,i}(\mathbf{w}_{L_{ub}}-{\varvec{\psi }}^{(i)}_{L_{ub},k-1})+v_k(i) \\ =\mathbf{u}_{L_{ub},k,i}{\overline{\varvec{\psi }}}^{(i)}_{L_{ub},k-1}+v_k(i) \\ =\left[ \begin{array}{lll} {\dot{\mathbf{u}}}_{k,i} &{} {\ddot{\mathbf{u}}}_{k,i} &{} {\mathop \mathbf{u}\limits ^{...}}_{k,i} \end{array} \right] \left[ \begin{array}{l} {\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1} \\ {\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1} \\ {\mathop {\overline{\varvec{\psi }}}\limits ^{...}}^{(i)}_{k-1} \end{array} \right] +v_k(i) \\ {={\dot{\mathbf{u}}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}+{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}+{\mathop \mathbf{u}\limits ^{...}}_{k,i}{\mathop {\overline{\varvec{\psi }}}\limits ^{...}}^{(i)}_{k-1}+v_k(i) \end{array} \end{aligned}$$
(17)

and

$$\begin{aligned} \begin{array}{l} e^{(L_k(i))}_{k,L_k(i)-\Delta }({\varvec{\psi }}^{(i)}_{k-1})=\mathbf{u}_{L_{opt}k,i}{} \mathbf{w}^o_{L_{opt}} \\ -\mathbf{u}_{k,i}(1:L_k(i)-\Delta ){\varvec{\psi }}^{(i)}_{k-1}(1:L_k(i)-\Delta )+v_k(i) \\ =\mathbf{u}_{L_{ub},k,i}\left[ \begin{array}{l} \mathbf{w}^{o}_{L_{opt}} \\ \mathbf{0}_{(L_{ub}-L_{opt})\times 1} \end{array} \right] -\mathbf{u}_{L_{ub},k,i}\left[ \begin{array}{l} {\dot{\varvec{\psi }}}^{(i)}_{k-1} \\ \mathbf{0}_{(L_{ub}-L_k(i)+\Delta )\times 1} \end{array} \right] \\ +v_k(i) \\ =\mathbf{u}_{L_{ub},k,i}\left[ \begin{array}{l} \dot{\mathbf{w}}-{\dot{\varvec{\psi }}}^{(i)}_{k-1} \\ \ddot{\mathbf{w}} \\ \mathop \mathbf{w}\limits ^{...} \end{array} \right] +v_k(i) \\ =\left[ \begin{array}{lll} {\dot{\mathbf{u}}}_{k,i} &{} {\ddot{\mathbf{u}}}_{k,i} &{} {\mathop \mathbf{u}\limits ^{...}}_{k,i} \end{array} \right] \left[ \begin{array}{l} {\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1} \\ \ddot{\mathbf{w}} \\ \mathop \mathbf{w}\limits ^{...} \end{array} \right] +v_k(i) \\ {=\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}+{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}+{\mathop \mathbf{u}\limits ^{...}}_{k,i}\mathop \mathbf{w}\limits ^{...}+v_k(i) \end{array} \end{aligned}$$
(18)

The main term in the FT updating (7), is written as:

$$\begin{aligned} \begin{array}{l} {(e^{(L_k(i))}_{k,L_k(i)}({\varvec{\psi }}^{(i)}_{k-1}))}^2-{(e^{(L_k(i))}_{k,L_k(i)-\Delta }({\varvec{\psi }}^{(i)}_{k-1}))}^2 \\ =(e^{(L_k(i))}_{k,L_k(i)}({\varvec{\psi }}^{(i)}_{k-1})-e^{(L_k(i))}_{k,L_k(i)-\Delta }({\varvec{\psi }}^{(i)}_{k-1})) \\ \times (e^{(L_k(i))}_{k,L_k(i)}({\varvec{\psi }}^{(i)}_{k-1})+e^{(L_k(i))}_{k,L_k(i)-\Delta }({\varvec{\psi }}^{(i)}_{k-1})) \end{array} \end{aligned}$$
(19)

Substituting (17) and (18) in (19) results:

$$\begin{aligned}&{(e^{(L_k(i))}_{k,L_k(i)}({\varvec{\psi }}^{(i)}_{k-1}))}^2-{(e^{(L_k(i))}_{k,L_k(i)-\Delta }({\varvec{\psi }}^{(i)}_{k-1}))}^2 \nonumber \\&\quad =({\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}-{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}) \nonumber \\&\quad \times (2{\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}+{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\nonumber \\&\quad +{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}+2{\mathop \mathbf{u}\limits ^{...}}_{k,i}\mathop \mathbf{w}\limits ^{...}+2v_k(i)) \end{aligned}$$
(20)

By multiplying the parentheses in (20), we get:

$$\begin{aligned} \begin{array}{l} {(e^{(L_k(i))}_{k,L_k(i)}({\varvec{\psi }}^{(i)}_{k-1}))}^2-{(e^{(L_k(i))}_{k,L_k(i)-\Delta }({\varvec{\psi }}^{(i)}_{k-1}))}^2 \\ =2v_k(i){\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}-2v_k(i){\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}} \\ +2{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}-2{\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}} \\ \quad+{\left[ {\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\right] }^2-{\left[ {\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}\right] }^2 \\ \quad +2{\mathop \mathbf{u}\limits ^{...}}_{k,i}\mathop \mathbf{w}\limits ^{...}{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}-2{\mathop \mathbf{u}\limits ^{...}}_{k,i}\mathop \mathbf{w}\limits ^{...}{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}} \end{array} \end{aligned}$$
(21)

For the ease of presentation, we define the following parameters:

$$\begin{aligned} \begin{array}{l} A=2v_k(i){\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}, \ \ \ \ \ \ \ \ \ \ \ \ B=2v_k(i){\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}} \\ C=2{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}, \ \ \ \ \ \ \ D=2{\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}} \\ E={\left[ {\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\right] }^2, \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ F={\left[ {\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}\right] }^2 \\ G=2{\mathop \mathbf{u}\limits ^{...}}_{k,i}\mathop \mathbf{w}\limits ^{...}{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}, \ \ \ \ \ \ \ \ \ H=2{\mathop \mathbf{u}\limits ^{...}}_{k,i}\mathop \mathbf{w}\limits ^{...}{\ddot{{\varvec{u}}}}_{k,i}\ddot{\mathbf{w}} \end{array} \end{aligned}$$
(22)

From (7), (21) and (22) we have:

$$\begin{aligned}{\ell }_{k,f}(i)&=({\ell }_{k-1,f}(i)-{\alpha }_k)\nonumber \\&\quad -{\gamma }_k[A-B+C-D+E-F+G-H] \end{aligned}$$
(23)

Now, according to the update equation, we analyze the steady-state performance. For simplicity of analysis, the following assumptions are made.

  1. 1.

    The regressors \(\mathbf{u}_{L_{ub},k,i}\) are spatially and temporally independent.

  2. 2.

    The components of \(\mathbf{u}_{L_{ub}k,i}\) are obtained from white Gaussian random process with zero mean and variance \({\sigma }^2_{u,k}\). The covariance matrix of \(\mathbf{u}_{L_{ub}k,i}\) is \(\mathbf{R}_{u,k}={\sigma }^2_{u,k}\mathbf{I}\).

  3. 3.

    In the steady-state, the tap-length in node k approximately converges to the constant value \(L_k(\infty )\). Also we assume that at steady-state, \(L_k(\infty )>L_{opt}\). This overestimate phenomenon will be justified by the results of the analysis.

  4. 4.

    The elements of the unknown parameter are obtained from the zero-mean random sequence with variance \({\sigma }^2_o\).

  5. 5.

    We assume that in the steady-state \(\{\left[ \begin{array}{c} {\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1} \\ {\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1} \end{array} \right] {\left[ \begin{array}{c} {\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1} \\ {\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1} \end{array} \right] }^T\}={\sigma }^2_{\overline{\varvec{\psi }},k-1}I, \)where \({\sigma }^2_{\overline{\varvec{\psi }},k-1}\) is the variance of the elements of \({\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\) and \({\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\).

As a result of these assumptions, we can say that \({\varvec{\psi }}^{(i)}_{k-1}\) and consequently \({\overline{\varvec{\psi }}}^{(i)}_{L_{ub},k-1}\) (or \({\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\) and \({\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\)) are independent from \(\mathbf{u}_{L_{ub},k,i}\) and \(v_k(i)\).

Considering the assumptions 1-5 and performing the mathematical expectations from both sides of (23) in steady-state, we have:

$$\begin{aligned}&E\{{\ell }_{k,f}(i)\}=(E\{{\ell }_{k-1,f}(i)\}-{\alpha }_k)\nonumber \\&\quad -{\gamma }_kE\{A-B+C-D+E-F+G-H\}, i\longrightarrow \infty \end{aligned}$$
(24)

With the assumption of \(L_k(\infty )\) and \(L_{k-1}(\infty )\) as steady-state tap-length of nodes k and \(k-1\) we have:

$$\begin{aligned}&L_k(\infty )=(L_{k-1}(\infty )-{\alpha }_k)\nonumber \\&\quad -{\gamma }_kE\{A-B+C-D+E-F+G-H\}, i\longrightarrow \infty \end{aligned}$$
(25)

To continue, we evaluate the moments in (25). From the assumptions 1–5 and their results we have:

$$\begin{aligned} \begin{array}{l} E\{A\}=E\{2v_k(i){\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\}=2E\{v_k(i)\}E\{{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\}=0 \\ E\{B\}=E\{2v_k(i){\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}\}=2E\{v_k(i)\}E\{{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}\}=0 \\ E\{C\}=E\{2{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\}\\ =2E\{{\ddot{\mathbf{u}}}_{k,i}\}E\{{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\}E\{{\dot{\mathbf{u}}}_{k,i}\}E\{{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\}=0 \\ E\{D\}=E\{2{\dot{\mathbf{u}}}_{k,i}{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}\}=2E\{{\dot{\mathbf{u}}}_{k,i}\}E\{{\dot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\}E\{{\ddot{\mathbf{u}}}_{k,i}\} E\{\ddot{\mathbf{w}}\}=0 \\ E\{E\}=E\{{\left[ {\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\right] }^2\}=tr(E\{{\left[ {\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\right] }^2\}) \\ =E(tr\{{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\}) \\ =E(tr\{{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{{\ddot{\overline{\varvec{\psi }}}}^{(i)^T}_{k-1}}{{\ddot{\mathbf{u}}}_{k,i}}^T\}) \\ =E(tr\{{{\ddot{\mathbf{u}}}_{k,i}}^T{\ddot{\mathbf{u}}}_{k,i}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{{\ddot{\overline{\varvec{\psi }}}}^{(i)^T}_{k-1}}\}) \\ =tr(E\{{{\ddot{\mathbf{u}}}_{k,i}}^T{\ddot{\mathbf{u}}}_{k,i}\}E\{{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{{\ddot{\overline{\varvec{\psi }}}}^{(i)^T}_{k-1}}\}) \\ =tr({\sigma }^2_{u,k}I_{\Delta }E\{{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}{{\ddot{\overline{\varvec{\psi }}}}^{(i)^T}_{k-1}}\}) \\ ={\sigma }^2_{u,k}E(tr\{{{\ddot{\overline{\varvec{\psi }}}}^{(i)^T}_{k-1}}{\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\})={\sigma }^2_{u,k}E\{{\Vert {\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\Vert }^2\} \\ E\{F\}=E\{{\left[ {\ddot{\mathbf{u}}}_{k,i}\ddot{\mathbf{w}}\right] }^2\}={\sigma }^2_{u,k}E\{{\Vert \ddot{\mathbf{w}}\Vert }^2\} \end{array} \end{aligned}$$
(26)

According to assumption in the steady-state, \(L_k(\infty )>L_{opt}\), so, in steady-state \(\mathop \mathbf{w}\limits ^{...}=0\) and it is resulted that in the steady-state the terms G and H will disappear. From this and substituting(26) in (25) result:

$$\begin{aligned}&L_k(\infty )=(L_{k-1}(\infty )-{\alpha }_k)-{\gamma }_k{\sigma }^2_{u,k} \{E\{{\Vert {\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\Vert }^2\}\nonumber \\&\quad -E\{{\Vert \ddot{\mathbf{w}}\Vert }^2\}\}, i\longrightarrow \infty \end{aligned}$$
(27)

If \((L_{k-1}(\infty )-{\Delta }){>}L_{opt}\), then in the steady-state \(\ddot{\mathbf{w}}=0\), but, if \((L_{k-1}(\infty )-{\Delta }){\le }L_{opt}\), then \(\ddot{\mathbf{w}}\) consists of \(L_{opt}-L_{k-1}(\infty )+{\Delta }\) nonzero elements of \(\mathbf{w}^{o}_{L_{opt}}\), so we have:

$$\begin{aligned} E\{{\Vert \ddot{\mathbf{w}}\Vert }^2\}= {\left\{ \begin{array}{ll} (L_{opt}-L_{k-1}(\infty )+{\Delta }){\sigma }^2_o &{}if \ \ \ \ {\ L}_{opt}{\le }L_{k-1}(\infty ){\le }L_{opt}+ \mathrm {\Delta } \\ 0 &{} if \ \ \ \ \ L_{k-1}(\infty ){>}L_{opt}+{\Delta } \end{array}\right. } \end{aligned}$$
(28)

In fact, if \(L_k(\infty )\) and \(L_{k-1}(\infty )\) are not exactly equal, but are almost equal. We can temporarily assume \(L_k(\infty )=L_{k-1}(\infty )\), and rewrite (27) as:

$$\begin{aligned}&\{E\{{\Vert {\ddot{\overline{\varvec{\psi }}}}^{(i)}_{k-1}\Vert }^2\}-E\{{\Vert \ddot{\mathbf{w}}\Vert }^2\}\}\nonumber \\&\quad =-{{\alpha }_k}/{{\gamma }_k{\sigma }^2_{u,k}} \ \ \ \ \ i\longrightarrow \infty \end{aligned}$$
(29)

Now if we assume that \(L_{k-1}(\infty ){>}(L_{opt}+{\Delta })\), then according to (28) discrepancy is seen between the parties of (29). Because of this assumption, the left-hand side of (29) is higher than zero, but the right-hand side has a negative value. Thus, we conclude that \({L}_{opt}{\le }L_{k-1}(\infty ){\le }(L_{opt}+{\Delta })\) will hold true. According to this and by defining \(\mathbf{P}_{L_{ub},k-1}={\overline{\varvec{\psi }}}^{({\infty })}_{L_{ub},k-1}\) and partitioning of \(\mathbf{P}_{L_{ub},k-1}\) similar to \({\overline{\varvec{\psi }}}^{(i)}_{L_{ub},k-1}\) in (16) into three parts \({\dot{\mathbf{P}}}_{k-1}\), \({\ddot{\mathbf{P}}}_{k-1}\) and \({\mathop \mathbf{P}\limits ^{...}}_{k-1}\) we can rewrite (27) as:

$$\begin{aligned} \begin{array}{l} L_k(\infty )=(L_{k-1}(\infty )-{\alpha }_k) \\ -{\gamma }_k{\sigma }^2_{u,k}\{E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}-(L_{opt}-L_{k-1}(\infty )+{\Delta }){\sigma }^2_o\} \end{array} \end{aligned}$$
(30)

In order to evaluate the term \(E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}\) we need the steady-state MSD of the deficient length DILMS algorithm. In [26], we have introduced the concept of deficient length DILMS algorithm and derived an expression for its steady-state MSD. The main results are included in appendix A to make the paper self-contained. Here we rewrite the resulting MSD that appears in appendix A, (71), as:

$$\begin{aligned} E\{{\Vert \mathbf{P}_{L_{ub},k-1}\Vert }^2\}={(1-\Pi _{k,1})}^{-1}s_k \end{aligned}$$
(31)

In comparison with (71) here the subscript \(L_{opt}\) in \(E\{{\Vert \mathbf{P}_{L_{opt}, k-1}\Vert }^2\}\) is modified as \(L_{ub}\), since here we are assuming that the length of the unknown parameter equals to \(L_{ub}\) . In (31), \(\Pi _{k,1}\) and \(s_k\) are defined as:

$$\begin{aligned}&\Pi _{k,\ell }{\triangleq \beta }_{k-1}{\beta }_{k-2}\dots {\beta }_1{\beta }_N{\beta }_{N-1}\dots \nonumber \\&\quad {\beta }_{k+\ell }{\beta }_{k+\ell -1}\ \ \ \ \ \ell =1,\dots ,N \nonumber \\&\quad s_k\triangleq \Pi _{k,2}f_k+\Pi _{k,3}f_{k+1}+\dots + \Pi _{k,N-1} f_{k-3}\nonumber \\&\quad +\Pi _{k,N}f_{k-2}+f_{k-1} \end{aligned}$$
(32)

with

$$\begin{aligned}&{\beta }_k=1-2{\mu }_k{\sigma }^2_{u,k}\nonumber \\&\quad +{\mu }^2_k{\sigma }^4_{u,k}(L_{k-1}(\infty )+2) \end{aligned}$$
(33)

and

$$\begin{aligned}f_k&=2{\mu }_k{\sigma }^2_{u,k}(1-{\mu }_k{\sigma }^2_{u,k}){\Vert \mathop \mathbf{w}\limits ^{...}\Vert }^2\nonumber \\&\quad +{\mu }^2_k{\sigma }^2_{v,k}{\sigma }^2_{u,k}L_{k-1}(\infty ) \nonumber \\&\quad ={\mu }^2_k{\sigma }^2_{v,k}{\sigma }^2_{u,k}L_{k-1}(\infty ) \end{aligned}$$
(34)

By comparing (33) and (34) with (66) and (69), we observe that in (33) and (34), M has changed to \(L_{k-1}(\infty )\), since here in the steady-state the length of adaptive filter in sensor \(k-1\) is equal to \(L_{k-1}(\infty )\), while in appendix A, length M is considered for the adaptive filter of each sensor. Also \({\Vert {\ddot{\mathbf{w}}}_{L_{opt}-M}\Vert }^2\) in (69) has changed to \({\Vert \mathop \mathbf{w}\limits ^{...}\Vert }^2\) in (34), since according to partitioning in (13), here \({\ddot{\mathbf{w}}}_{L_{opt}-M}\) is equivalent to \(\mathop \mathbf{w}\limits ^{...}\), which is also consistent with \(\mathop \mathbf{w}\limits ^{...}=0\) in the steady-state. Now using (31)-(34) we want to compute \(E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}\). In order to do so, first we divide the steady-state MSD into three parts as:

$$\begin{aligned}E\{{\Vert \mathbf{P}_{L_{ub},k-1}\Vert }^2\}&= E\{{\Vert {\dot{\mathbf{P}}}_{k-1}\Vert }^2\}+E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}\nonumber \\&\quad + E\{{\Vert {\mathop \mathbf{P}\limits ^{...}}_{k-1}\Vert }^2\} \end{aligned}$$
(35)

Using the assumption 5, and since \({{\mathop \mathbf{P}\limits ^{...}}_{k-1}=\mathop {\overline{\varvec{\psi }}}\limits ^{...}}^{(\infty )}_{k-1} =\mathop \mathbf{w}\limits ^{...}=0\), we have:

$$\begin{aligned} \begin{array}{l} E\{{\Vert {\dot{\mathbf{P}}}_{k-1}\Vert }^2\}=(L_{k-1}(\infty )-{\Delta }){\sigma }^2_{\overline{\varvec{\psi }}, k-1} \\ E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}={\Delta }{\sigma }^2_{\overline{\varvec{\psi }},k-1} \\ E\{{\Vert {\mathop \mathbf{P}\limits ^{...}}_{k-1}\Vert }^2\}=0 \end{array} \end{aligned}$$
(36)

Substituting (36) into (35), we get:

$$\begin{aligned} E\{{\Vert \mathbf{P}_{L_{ub},k-1}\Vert }^2\}= L_{k-1}(\infty ){\sigma }^2_{\overline{\varvec{\psi }},k-1} \end{aligned}$$
(37)

From (37) and (36) we have:

$$\begin{aligned} E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}=\frac{\Delta }{L_{k-1}(\infty )}E\{{\Vert \mathbf{P}_{L_{ub},k-1}\Vert }^2\} \end{aligned}$$
(38)

According to (31), the eqution (38) is rewritten as:

$$\begin{aligned} E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}=\frac{\Delta }{L_{k-1}(\infty )}{(1-\Pi _{k,1})}^{-1}s_k \end{aligned}$$
(39)

or

$$\begin{aligned} E\{{\Vert {\ddot{\mathbf{P}}}_{k-1}\Vert }^2\}=\Delta {(1-\Pi _{k,1})}^{-1}{{\dot{s}}}_k \end{aligned}$$
(40)

where

$$\begin{aligned} {{\dot{s}}}_k\triangleq \frac{s_k}{L_{k-1}(\infty )}&=\Pi _{k,2}{{\dot{f}}}_k +\Pi _{k,3}{{\dot{f}}}_{k+1}+\dots \\ &\quad+\Pi _{k,N-1}{{\dot{f}}}_{k-3} \\&\quad +\Pi _{k,N}{{\dot{f}}}_{k-2}+{{\dot{f}}}_{k-1} \end{aligned}$$
(41)

with

$$\begin{aligned} {{\dot{f}}}_k=\frac{f_k}{L_{k-1}(\infty )}={\mu }^2_k{\sigma }^2_{v,k}{\sigma }^2_{u,k} \end{aligned}$$
(42)

Substituting (40) into (30) results:

$$\begin{aligned} L_k(\infty )&=L_{k-1}(\infty )-{\alpha }_k \\&\quad -{\gamma }_k{\sigma }^2_{u,k}\{\Delta {(1-\Pi _{k,1})}^{-1}{{\dot{s}}}_k\\&\quad -(L_{opt}-L_{k-1}(\infty )+{\Delta }){\sigma }^2_o\} \end{aligned}$$
(43)

Equation (43) can be arranged as the following:

$$\begin{aligned} L_k(\infty )&=(1-{\gamma }_k{\sigma }^2_{u,k}{\sigma }^2_o)L_{k-1}(\infty )\\ &\quad+{\gamma }_k{\sigma }^2_{u,k}{\sigma }^2_o\{L_{opt}+{\Delta }-\frac{\Delta }{{\sigma }^2_o}{(1-\Pi _{k,1})}^{-1} {{\dot{s}}}_k\}-{\alpha }_k \end{aligned}$$
(44)

In the resulting equation, \(\Pi _{k,1}\) depends on \({\beta }_k\) and according to (33) depends on \(L_{k-1}(\infty )\). But, on the other hand, as previously explained, we know \(L_{opt}{\le }L_{k-1}(\infty ){\le }(L_{opt}+{\Delta })\), and in practice \(\Delta {\ll } L_{opt}\); thus, \(L_{k-1}(\infty )\) is very close to \(L_{opt}\). The equation (33) can be rewritten as:

$$\begin{aligned} {\beta }_k=1-2{\mu }_k{\sigma }^2_{u,k}+{\mu }^2_k{\sigma }^4_{u,k}(L_{opt}+2) \end{aligned}$$
(45)

By defining \(h_k\) and \(g_k\) as:

$$\begin{aligned} \begin{array}{l} h_k=1-{\gamma }_k{\sigma }^2_{u,k}{\sigma }^2_o \\ g_k={\gamma }_k\\ {\sigma }^2_{u,k}{\sigma }^2_o\{L_{opt}+{\Delta }-\frac{\Delta }{{\sigma }^2_o} {(1-\Pi _{k,1})}^{-1}{{\dot{s}}}_k\}-{\alpha }_k \end{array} \end{aligned}$$
(46)

(44) can be written as:

$$\begin{aligned} L_k(\infty )=h_kL_{k-1}(\infty )+g_k \end{aligned}$$
(47)

It should be noted that (47) is a coupled equation: it contains \(L_k(\infty )\) and \(L_{k-1}(\infty )\), i.e., information of two different locations. To mitigate this issue, we use ring topology [4]- [6]. Thus, by iterating (47), we have:

$$\begin{aligned} \begin{array}{l} L_1(\infty )=h_1L_N(\infty )+g_1 \\ L_2(\infty )=h_2L_1(\infty )+g_2 \\ \vdots \\ L_{k-2}(\infty )=h_{k-2}L_{k-3}(\infty )+g_{k-2} \\ L_{k-1}(\infty )=h_{k-1}L_{k-2}(\infty )+g_{k-1} \\ \vdots \\ L_N(\infty )=h_NL_{N-1}(\infty )+g_N \end{array} \end{aligned}$$
(48)

Observe that according to (48), \(L_{k-1}(\infty )\) can be express in term of \(L_{k-3}(\infty )\) as:

$$\begin{aligned} L_{k-1}(\infty )=h_{k-1}h_{k-2}L_{k-3}(\infty )+h_{k-1}g_{k-2}+g_{k-1} \end{aligned}$$
(49)

By iterating in this manner, we have:

$$\begin{aligned} L_{k-1}(\infty )&=h_{k-1}h_{k-2}\dots h_1h_Nh_{N-1}\dots h_kL_{k-1}(\infty ) \\ &\quad +h_{k-1}h_{k-2}\dots h_1h_Nh_{N-1}\dots h_{k+1}g_k \\ &\quad+h_{k-1}h_{k-2}\dots h_1h_Nh_{N-1}\dots h_{k+2}g_{k+1} \\ &\quad+\dots +h_{k-1}h_{k-2}\dots h_1h_Ng_{N-1} \\ +h_{k-1}h_{k-2}\dots h_1g_N+h_{k-1}h_{k-2}\dots h_2g_1 \\ &\quad+h_{k-1}h_{k-2}\dots {h_4h}_3g_2+\dots \\&\quad +h_{k-1}h_{k-2}g_{k-3}+h_{k-1}g_{k-2}+g_{k-1} \end{aligned}$$
(50)

We use N quantities for each sensor k as:

$$\begin{aligned}&G_{k,\ell }{\triangleq h}_{k-1}h_{k-2}\dots h_1h_Nh_{N-1}\dots \nonumber \\&\quad h_{k+\ell } h_{k+\ell -1},\ \ \ \ell =1,\dots ,N \end{aligned}$$
(51)

and \(m_k\) is defined as:

$$\begin{aligned}&m_k\triangleq G_{k,2}g_k+G_{k,3}g_{k+1}+\dots \nonumber \\&\quad +G_{k,N-1}g_{k-3}+G_{k,N}g_{k-2} + g_{k-1} \end{aligned}$$
(52)

From (50)-(52) we have:

$$\begin{aligned} L_{k-1}(\infty )=G_{k,1}L_{k-1}(\infty )+m_k \end{aligned}$$
(53)

From (53), we illustrate the desired steady-state tap-length as:

$$\begin{aligned} L_{k-1}(\infty )={(1-G_{k,1})}^{-1}m_k \end{aligned}$$
(54)

The equation (54) renders a mathematical expression of the steady-state tap-length. Because of the complicated form of the obtained results for the steady-state tap-length, it is almost difficult to make beneficial information about the steady-state tap-length of each sensor node. To achieve a clear view of the steady-state tap-length, we simplify the equation (54). For this issue, we assume that all sensors employ the same step size for the tap-length adaption, i.e., \({\gamma }_k=\gamma ,\ \forall k\epsilon {\mathcal {N}}\), then, we assume that \(\mathbf{R}_{u,k}={\sigma }^2_u\mathbf{I}\). With these assumptions, we have:

$$\begin{aligned} h_k=1-\gamma {\sigma }^2_u{\sigma }^2_o \end{aligned}$$
(55)

Furthermore, we assume \(\gamma {\sigma }^2_u{\sigma }^2_o\ll 1\) such that:

$$\begin{aligned} G_{k,1}= & {} h_1h_2\dots h_N\nonumber \\= & {} {(1-\gamma {\sigma }^2_u{\sigma }^2_o)}^N\approx 1-N \gamma {\sigma }^2_u{\sigma }^2_o \end{aligned}$$
(56)

Also, \(m_k\) can be approximate as:

$$\begin{aligned} \begin{array}{l} m_k \approx \sum ^N_{k=1}{g_k}=N\gamma {\sigma }^2_u{\sigma }^2_o(L_{opt}+{\Delta }) \\ -\gamma {\sigma }^2_u{\Delta }\sum ^N_{k=1}{{(1-\Pi _{k,1})}^{-1}{{\dot{s}}}_k} -\sum ^N_{k=1}{{\alpha }_k} \end{array} \end{aligned}$$
(57)

Substituting (56) and (57) in (54) result:

$$\begin{aligned}L_{k-1}(\infty )&=(L_{opt}+{\Delta })\nonumber \\&\quad -\frac{\Delta }{N{\sigma }^2_o}\sum ^N_{k=1}{{(1-\Pi _{k,1})}^{-1}{{\dot{s}}}_k} \nonumber \\&\quad -\frac{\sum ^N_{k=1}{{\alpha }_k}}{N\gamma {\sigma }^2_u{\sigma }^2_o} \end{aligned}$$
(58)

Since the value of steady-state tap-length in (58) rarely is an integer, the steady-state tap-length usually varies nearby this value. The last two terms of the right-hand side of (58) are small and can be ignored. So, the steady-state tap-length is close to the value \(L_{opt}+\Delta \), as will be presented in the simulations. This gives the result that the incremental FT algorithm overestimates the steady-state tap-length, and the third assumption is confirmed.

4 Simulation results

In this part, simulations are presented to verify the theoretical results of the paper. Numerical simulations are made under three low noise, high noise, and large step sizes for tap-length adaption conditions.

4.1 Low-noise scenario (SNR=20dB)

We assume a network with N=12 nodes exploring an unknown parameter with length \(L_{opt}\)=20. The elements of the unknown parameter are obtained from a white and zero-mean random sequence with variance \({\sigma }^2_o\)=0.01. The regressors are independent zero-mean Gaussian variables, and their covariance matrix is \(\mathbf{R}_{u,k}=\mathbf{I}\). The observation noise is zero-mean white Gaussian and scaled to have the SNR=20dB. The step size for the DILMS algorithm is set to \({\mu _k}\)=0.05 for all nodes. The parameters of the incremental FT algorithm are chosen as \({\delta }_k\)=1, \({\alpha }_k\)=0.001, \({\gamma }_k\)=6, \(\Delta \)=6 with initial tap-length that are set in the minimum value \(L_{min}\)=8. Figures 1 and 2 respectively show the evolutions of the fractional and integer tap-length, under a low noise condition for node k=1. To obtain the integer values for the theoretical tap-lengths, we applied the floor operator \(\lfloor .\rfloor \) to (54). As shown in the figures, this node converges to a steady-state tap-length that corresponds well with the value obtained for the theoretical steady-state tap-length. We observe through the simulations that this is true for all nodes.

Fig. 1
figure 1

Evolution of the fractional tap-length in low-noise scenario

Fig. 2
figure 2

Evolution of the integer tap-length in low-noise scenario

4.2 High-noise scenario (SNR=0dB)

In this case, a high-noise environment is considered for simulations. The parameters used in the simulations are the same as that of the previous case, but the observation noise is scaled to reach the SNR=0dB. The step size of the DILMS and tap-length adaption algorithm are set to \({\mu }_k\)=0.001, and \({\gamma }_k\)=2, respectively. The evolutions of the fractional and integer tap-length, under this condition, are shown in Figs. 3 and 4 for node k=1. It is observed that the simulation results support the theoretical analyses.

Fig. 3
figure 3

Evolution of the fractional tap-length in high-noise scenario

Fig. 4
figure 4

Evolution of the integer tap-length in high-noise scenario

4.3 Large step size for tap-length adaption

The simulation setup of this scenario is the same as that of the first scenario, except that, in this case, we use a significant value for the tap-length adaption step size \({\gamma }_k\)=30. However, increasing the step size \({\gamma }_k\) increases the convergence rate, but a large \({\gamma }_k\) causes a significant fluctuation of the evolution curve. The evolution curves of the fractional and integer tap-length, under this condition, are presented in Figs. 5 and 6 for node k=1. It is easily seen from the figures that the steady-state tap-length fluctuates are about the theoretical values.

Fig. 5
figure 5

Evolution of the fractional tap-length for large tap-length adaption step size, \({\gamma }_k\)=30

Fig. 6
figure 6

Evolution of the integer tap-length for large tap-length adaption step size, \({\gamma }_k\)=30

Table 2 shows the simulation results for different SNR values and compares them with the theoretical results. All parameters are following the previous experiments except for \(\mu _k\)=0.001 and \(\gamma _k\)=2. As the table shows, for different values of SNRs, the theoretical results are well compatible with the simulation results.

Table 2 The steady-state tap-length performance in different SNRs

Table 3 also shows the results for different \(L_{opt}\) values and compares them with the theoretical results. In this simulation, \(\mu _k\)=0.001 and \(\gamma _k\)=10. As shown, first, the simulation results are compatible with the theoretical results. Secondly, all these results show that the length of the steady-state is overestimated almost as much as \(\Delta \). This overestimation of length is important because it shows that length is not estimated deficiently. As shown in [26], deficient length estimation increases the steady-state error. On the other hand, a length much larger than the optimal value increases the computational complexity, which is not proper for a network of low-power sensors. However, the overestimation of the length is not significant and almost equal to \(\Delta \), which is much smaller than the optimal length.

Table 3 The steady-state tap-length performance in different values of \(L_{opt}\)

5 Conclusions

In this paper, we investigated the performance of the distributed incremental variable fractional tap-length (FT) LMS algorithm. Since in this scenario, the length of the vectors varies in each iteration, so we defined an upper bound for the size and padded the vectors by zeros to achieve this length. We obtained the vectors with the same sizes that simplified the calculations. Based on these findings, we analyzed the incremental variable FT LMS algorithm and derived the mathematical expression for the steady-state tap-length at each sensor node. The most important result is that this algorithm overestimates the optimal value of the tap-length. Computer simulation results confirmed the mathematical analysis and the derived closed-form expression of tap-length.