Keywords

1 Introduction

Recently, the problem of deconvolution estimation for linear systems has attracted much attention due to its extensive applications in image processing [1], oil exploration [2, 3] and so on. The precursory work on the deconvolution problems can be traced back to [2], which contains the study of white noise estimation according to the Kalman filter approach. As for the deconvolution estimation problems, another approach based on the modern time series analysis method is presented in [4], which includes both the input white noise estimator and measurement white noise estimator. However, communication networks are usually unreliable and may lead to time delay and packet dropout, so that many results about time delay and packet dropout are presented in [517]. For the time delay problems, the authors apply the state augmentation approach to solve time delay problem in [5, 6]. There is another approach in [7], where the authors use the partial difference Riccati equation approach to deal with time delay problem.

We retrospect the pioneer work on the Kalman filter about packet losses to [8], in which the author uses the statistics of the unobserved Bernoulli process to describe the observation uncertainty and to derive the estimator. In [9], the sensor measurements are encoded together and sent over the network in a single packet, so that the Kalman filter receives either the complete observation if the packet is received or none of the observation if the packet is lost. In [10], the measurements are transmitted to the filter via two communication channels, while the authors consider the measurements may be sent through different communication channels in the networked filter systems in [11], thus [11] is more general than [10]. Some results about multiple packet dropout have been published in [1217].

For nonlinear stochastic systems with multi-step transmission delays, multiple packet dropouts and correlated noises, the authors study the recursive estimations by using the innovation analysis approach in [15], in which the noises are assumed to be one-step autocorrelated and cross-correlated. In [16], the multiple packet dropouts phenomenon is considered to be random and is described by a binary switching sequence, which obeys a conditional probability distribution. The authors calculate the recursive estimators based on an innovation analysis method and the orthogonal projection theorem.

In previous works about time delay and packet dropout, the authors have dealt with time delay problem using the partial difference Riccati equation approach or the state augmentation approach. While, those approaches may bring tremendous computation when time delay \(d_{l-1}\) is large. The authors have used a scalar independent and identically distributed (i.i.d) Bernoulli process to describe the packet dropout phenomenon. In the real word, scalar can not satisfy realistic. Stimulated by above works about measurement dropout and time delay systems, we will investigate deconvolution estimation for discrete-time systems with measurement delay and packet dropout in this paper. According to the projection formula and the reorganized innovation analysis approach, the state estimation is first obtained. Then, we gain the white noise estimation on the basis of the state estimation obtained above. The major contributions of this paper are as follows: (i) we describe the multiplicative noise using a diagonal matrix dropout, which is described by an independent and identically distributed (i.i.d) Bernoulli process. A closed-form result is gained using the Hadamard product flexibly. (ii) according to the reorganized innovation analysis approach, we derive the optimal state estimator and noise estimator utilizing l Raccati difference equations and one Lyapunov difference equation. When the delay \(d_{l-1}\) is large, our proposed approach is more efficient than the classical augmentation approach in [5, 6] and partial difference Riccati equation approach in [7].

The organization of this paper is following. In Sect. 2, we put forward the problem formulae, some assumptions and remarks. In Sect. 3, we deduce the optimal state estimation according to the reorganized innovation analysis approach. Then, we obtain the white noise estimation based on the optimal state estimation gained above and the projection formula. A numerical example is given to explain the effectiveness of our approach in Sect. 4. Finally, some conclusions are provided in Sect. 5.

Notation: From beginning to end in this paper, the superscripts “\(-1\)” and “T” represent the inverse and transpose of a matrix. \(\mathcal {R}^{n}\) denotes an Euclidean space of n-dimension. \(\mathcal {R}^{n\,\times \,m}\) means the linear space of all \(n\times m\) real matrices. The measurement sequence \(\{y(0), \ldots , y(k)\}\) can be represented as \(\mathcal {L}\left\{ y(s)^{k}_{s=0}\right\} \). \(\odot \) denotes the Hadamard product. The diagonal matrix with diagonal elements \(\lambda _{1},\ldots , \lambda _{n}\) is expressed as \(diag\{\lambda _{1},\ldots , \lambda _{n}\}\). In addition, the mathematical expectation operator is denoted as E.

2 Problems Statement and Preliminary

Consider the linear system following

$$\begin{aligned} x(k+1)= & {} Ax(k)+n(k), \end{aligned}$$
(1)
$$\begin{aligned} y_{i}(k)= & {} \xi _{i}(k)B_{i}x(k-d_{i})+v_{i}(k),k\ge d_{i}, i=0,\ldots , l-1, \end{aligned}$$
(2)

here \({x}(k)\in \mathcal {R}^{n}\) is an unknown state and \({y}_{i}(k)\in \mathcal {R}^{m_{i}}\) is delayed measurement, respectively. n(k) and \({ v}_{i}(k)\) are white Gaussian noises with zero mean and covariances \(E \{{n}(k) {n}^T (j)\} ={Q}\delta _{k,j}, E\{{ v}_{i}(k){v}^T_{i} (j)\} ={ R_{i}}\delta _{k,j}\) respectively. Here, \(\delta _{k,j}\) is the Kronecker function. We describe the packet dropout phenomenon with the mutually uncorrelated and identically distributed (i.i.d.) Bernoulli random variables \(\xi _{ij}(k)\), in the \(m_{i}\) channels with \(Pr\{\xi _{ij}(k)= 1\} =\alpha _{ij}\), \(Pr\{\xi _{ij}(k)= 0\} =1-\alpha _{ij}\). \(d_{i}\) satisfies that \(0=d_{0}<d_{1}<\cdots < d_{l-1}\). The initial state x(0) is a random vector with zero mean and covariance matrix D(0). The random process n(k), \(v_{i}(k)\), \(\xi _{i}(k)\) for all k and the initial state x(0) are uncorrelated mutually.

Let

$$\begin{aligned} y(k)={\left\{ \begin{array}{ll} y_{0}(k), \quad ~~~~ 0\le k< d_{1}, \\ ~~~~\vdots \\ \left[ \begin{array}{c} y_{0}(k) \\ \vdots \\ y_{i}(k) \end{array} \right] ,\quad d_{i} \le k< d_{i+1}, \\ ~~~~\vdots \\ \left[ \begin{array}{c} y_{0}(k) \\ \vdots \\ y_{l-1}(k) \end{array} \right] ,\quad d_{l-1} \le k. \end{array}\right. } \end{aligned}$$
(3)

Problem

For the given measurements \(\{y(k)\}^{N}_{k=0}\), find a LMMSE estimator \(\hat{x}(k\mid k)\) of x(k) and \(\hat{n}(k\mid k+T)\) of n(k), such that

$$\begin{aligned} E\{[n(k)-\hat{n}(k\mid k+T)][n(k)-\hat{n}(k\mid k+T)]^{T}\} \end{aligned}$$

is minimized. Note that \(T = 0\) is the filter, \(T >0\) is the smoother and \(T <0\) is the predictor.

Remark 1

From the distribution of \(\xi _{ij}(k)\), it is readily to deduce that \(E\{\xi _{ij}(k)\}=\alpha _{ij}\), \(E\{(\xi _{ij}(k)-\alpha _{ij})^{2}\}=\alpha _{ij}(1 -\alpha _{ij})\), \(E\{\xi _{ij}(k)(1-\xi _{ij}(k))\}=0\), \(E\{[\xi _{ij}(k)-\alpha _{ij}][\xi _{il}(s)-\alpha _{il}]\}\) \(=\alpha _{ij}(1-\alpha _{ij})\delta _{k,s}\delta _{j,l}\). \(E\{\xi _{ij}(k)\xi _{il}(s)\}\) \(=\alpha _{ij}\alpha _{il}\) for \(k\ne s\) or \(j\ne l\).

Remark 2

In the previous references, the authors usually use a scalar independent and identically distributed (i.i.d) Bernoulli process to describe the packet dropout phenomenon. In this paper, we describe the packet dropout via a diagonal matrix independent and identically distributed (i.i.d) Bernoulli process, which is more realistic. Regarding to the time delay systems, one can use the state augmentation approach to solve the optimal state estimation and white noise estimation problems, but the state augmentation approach may bring tremendous computational when the delay \(d_{l-1}\) is large. Therefore, in this paper we will deduce the problem of deconvolution estimation using the reorganized innovation analysis approach to avoid tremendous computation.

3 Main Results

In this section, we will present an analytical solution to the optimal state estimation and the white noise estimation according to the projection formula.

Lemma 1

([18]) Let \(\sigma = diag\{\sigma _{1},\ldots , \sigma _{n} \}\) and \(\rho = diag\{\rho _{1},\ldots ,\rho _{m}\}\) be two diagonal stochastic matrices, and A be any \(n \times m\) matrix. Then

$$\begin{aligned} E\{\sigma A \rho \}=\left[ \begin{array}{ccc} E\{\sigma _{1}\rho _{1}\} &{} \cdots &{} E\{\sigma _{1}\rho _{m}\}\\ \vdots &{} \ddots &{} \vdots \\ E\{\sigma _{n}\rho _{1}\} &{} \cdots &{} E\{\sigma _{n}\rho _{m}\} \\ \end{array} \right] \odot A. \end{aligned}$$
(4)

3.1 Design of the Optimal State Estimator

In the next, we first provide an optimal state estimator, which will lead to the optimal white noise estimator. We note that \(y_{i}(k)\) is an additional measurement of the state \(x(k-d_{i})\), which is gained at time instant k with time delay \(d_{i}\), so the measurement y(k) contains time delay when \(k\ge d_{i}\). On the basis of [19], the linear space \(\mathcal \{{y(s)}^{k}_{s=0}\}\) includes the same information as

$$\begin{aligned} \mathcal L \left\{ \{Y_{l-1}(s)\}^{k-d_{l-1}}_{s=0}, \ldots ,\{Y_{0}(s)\}^{k}_{s=k-d_{1}+1}\right\} , \end{aligned}$$

where the new observations

$$\begin{aligned} Y_{l-1}(s)= & {} \left[ \begin{array}{c} y_{0}(s) \\ ~~~~\vdots \\ y_{l-1}(s+d_{l-1}) \end{array}\right] , 0\le s \le k-d_{l-1}, \end{aligned}$$
(5)
$$\begin{aligned} \vdots \nonumber \quad \\ Y_{i}(s)= & {} \left[ \begin{array}{c} y_{0}(s) \\ \vdots \\ y_{i}(s+d_{i}) \end{array}\right] , k-d_{i+1}+1 \le s \le k-d_{i}, \end{aligned}$$
(6)
$$\begin{aligned} \vdots \nonumber \quad \\ Y_{0}(s)= & {} y_{0}(s), k-d_{1}+1\le s \le k. \end{aligned}$$
(7)

Obviously, \(Y_{0}(s)\), \(Y_{1}(s)\),..., \(Y_{l-1}(s)\) satisfy

$$\begin{aligned} Y_{i}(s)=H_{i}x(s)+V_{i}(s),i=0,1,2,\ldots , l-1, \end{aligned}$$
(8)

where \(H_{i}=diag\{\xi _{0}(s),\xi _{1}(s+d_{1}),\ldots , \xi _{i}(s+d_{i})\}\bar{B}_{i}\), and

$$\begin{aligned} \bar{B}_{i}=\left[ \begin{array}{cccc} B^{T}_{0},B^{T}_{1},\ldots ,B^{T}_{i} \end{array}\right] ^{T},~V_{i}(s)=\left[ \begin{array}{c} v^{T}_{0}(s),v^{T}_{1}(s+d_{1}),\ldots ,v^{T}_{i}(s+d_{i}) \end{array} \right] ^{T}. \end{aligned}$$

Obviously the new measurements \(Y_{0}(s)\), \(Y_{1}(s), \ldots , Y_{i}(s)\) are delay-free and the associated measurement noises \(V_{0}(s)\), \(V_{1}(s), \ldots , V_{i}(s)\) are white noises with zero mean and covariance matrices \(R_{V_{0}(s)}\,{=}\,R_{0}\), \(R_{V_{1}(s)}\,{=}\,diag\{R_{0},R_{1}\}, \ldots , R_{V_{i}(s)}\,{=}\,diag\{R_{0},R_{1},\ldots ,R_{i}\}\). The filter \(\hat{x}(k\mid k)\) is the projection of x(k) onto the linear space of

$$\begin{aligned} \mathcal L \left\{ \{Y_{l-1}(s)\}^{k-d_{l-1}}_{s=0}, \ldots , \{Y_{i}(s)\}^{k-d_{i}}_{s=k-d_{i+1}+1},\ldots , \{Y_{0}(s)\}^{k}_{s=k-d_{1}+1}\right\} . \end{aligned}$$

In order to compute the projection, we define the innovation sequence as follows:

$$\begin{aligned} \varepsilon (s,i)=Y_{i}(s)-\hat{Y}(s,i), i=0,1,2,\ldots , l-1, \end{aligned}$$
(9)

where \(\hat{Y}(s,i)\) is the projection of \(Y_{i}(s)\) onto the linear space of

$$\begin{aligned} \mathcal L \left\{ \{Y_{l-1}(m)\}^{k-d_{l-1}}_{m=0}, \ldots , \{Y_{i}(m)\}^{s-1}_{m=k-d_{i+1}+1} \right\} . \end{aligned}$$

From (8)–(9), the innovation sequence is given as follows

$$\begin{aligned} \varepsilon (s,i)= & {} diag\{\xi _{0}(s)-\phi _{0}, \ldots ,\xi _{i}(s)-\phi _{i}\} \bar{B}_{i}x(s)\nonumber \\&+\,diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\} \bar{B}_{i}\tilde{x}(s,i)+V_{i}(s), \end{aligned}$$
(10)

here \(\phi _{i}=diag\{\alpha _{i1},\alpha _{i2},\ldots ,\alpha _{im_{i}}\}\), \(\tilde{x}(s,i)=x(s)-\hat{x}(s,i)\), the definition of \(\hat{x}(s,i)\) is same to \(\hat{Y}(s,i)\). We know that white noises \(\varepsilon (s,0)\), \(\varepsilon (s,1), \ldots , \varepsilon (s,l-1)\) are independent. Conveniently, we define that

$$\begin{aligned} P_{i}(s)= & {} E\{\tilde{x}(s,i)\tilde{x}^{T}(s,i)\},\nonumber \\ D(s)= & {} E\{x(s)x^{T}(s)\}.\nonumber \end{aligned}$$

Based on (10) and Lemma 1, the recognized innovation covariance matrices are calculated by

$$\begin{aligned} R_{\varepsilon (s,i)}= & {} E\{{diag\{\xi _{0}(s)-\phi _{0},\ldots ,\xi _{i}(s)-\phi _{i}\}}\bar{B}_{i}x(s)x^{T}(s) \bar{B_{i}}^{T}\nonumber \\&diag\{\xi _{0}(s)-\phi _{0},\ldots ,\xi _{i}(s)-\phi _{i}\}\nonumber \\&+\,diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\bar{B}_{i}\tilde{x}(s,i)\tilde{x}^{T}(s,i) \bar{B}_{i}^{T}\nonumber \\&diag\{\phi _{0},\phi _{1},\ldots \phi _{i}\}\}+diag\{R_{0},R_{1},\ldots , R_{i}\}\nonumber \\= & {} diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\bar{B}_{i}P_{i}(s)\bar{B}_{i}^{T} diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\nonumber \\&+\,diag\{R_{0},R_{1},\ldots , R_{i}\}\nonumber \\&+\,diag\{\varPi _{0},\varPi _{1},\ldots ,\varPi _{i}\}\odot \left( \bar{B}_{i}D(s)\bar{B}_{i}^{T}\right) , \end{aligned}$$
(11)

where

$$\begin{aligned} \varPi _{i}=\left[ \begin{array}{cccc} \alpha _{i1}(1-\alpha _{i1}) &{} 0 &{}\cdots &{} 0\\ 0&{} \alpha _{i2}(1-\alpha _{i2}) &{}\cdots &{} 0\\ \vdots &{} \vdots &{} \ddots &{} \vdots \\ 0&{} 0&{}\cdots &{} \alpha _{im_{i}}(1-\alpha _{im_{i}}) \end{array} \right] . \nonumber \end{aligned}$$

Now, we deduce the covariance matrices of one-step ahead state estimation error using the lemma as follows.

Lemma 2

The covariance matrices \(P_{i}(s+1)\) submits to the Riccati difference equations following,

$$\begin{aligned} P_{i}(s+1)= & {} AP_{i}(s)A^{T}+Q-AP_{i}(s)\bar{B}_{i}^{T} diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\nonumber \\&R_{\varepsilon (s,i)}^{-1}diag\{\phi _{0},\phi _{1}, \ldots ,\phi _{i}\}\bar{B}_{i}P_{i}(s)A^{T}, \end{aligned}$$
(12)
$$\begin{aligned} P_{l-1}(0)= & {} D(0), \end{aligned}$$
(13)
$$\begin{aligned} P_{i}(k-d_{i}+1)= & {} P_{i-1}(k-d_{i}+1), \end{aligned}$$
(14)

where \(R_{\varepsilon (s,i)}\) is the one in (11), and \(D(s+1)\) can be calculated using

$$\begin{aligned}&D(s+1)=AD(s)A^{T}+Q \end{aligned}$$
(15)

with initial value D(0).

Proof

According to (1), it is readily to yield (15). On the basis of projection formula, \(\hat{x}(s+1,i)\) is calculated

$$\begin{aligned} \hat{x}(s+1,i)= & {} proj\{x(s+1)\mid \{\varepsilon (m,l-1)\}_{m=0}^{k-d_{l-1}},\ldots , \{\varepsilon (m,i)\}_{m=k-d_{i+1}+1}^{s}\}\nonumber \\= & {} A\hat{x}(s,i)+E\{x(s+1)\varepsilon ^{T}(s,i)\}R^{-1}_{\varepsilon (s,i)}\varepsilon (s,i)\nonumber \\= & {} A\hat{x}(s,i)+K_{P}(s,i)\varepsilon (s,i), \end{aligned}$$
(16)

here \(K_{P}(s,i)=AP_{i}(s)\bar{B}_{i}^{T}diag\{\phi _{0}, \phi _{1}, \ldots ,\phi _{i}\}R_{\varepsilon (s,i)}^{-1}.\) From (1) and (16), one has that

$$\begin{aligned} \tilde{x}(s+1,i)=A\tilde{x}(s,i)+n(s)-K_{P}(s,i)\varepsilon (s,i). \end{aligned}$$
(17)

Therefore, the prediction error covariance is obtained by

$$\begin{aligned} P_{i}(s+1)= & {} E\{\tilde{x}(s+1,i)\tilde{x}^{T}(s+1,i)\}\nonumber \\= & {} AP_{i}(s)A^{T}+Q-AP_{i}(s)\bar{B}_{i}^{T}diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}R_{\varepsilon (s,i)}^{-1}\nonumber \\&diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\bar{B}_{i}P_{i}(s)A^{T},\nonumber \end{aligned}$$

which is (12). On the basis of the definitions of \(\hat{x}(k-d_{i}+1,i)\), we have \(\hat{x}(k-d_{i}+1,i)=\hat{x}(k-d_{i}+1,i+1)\). The proof is accomplished.\(\nabla \)

Now, we introduce the filter according to the projection formula in Hilbert space.

Theorem 1

In view of the system (1)–(2), the filter \(\hat{x}(k\mid k)\) can be computed by

$$\begin{aligned} \hat{x}(k\mid k)=\hat{x}(k,0)+P_{0}(k)B_{0}^{T}\phi _{0} R_{\varepsilon (k,0)}^{-1}\varepsilon (k,0), \end{aligned}$$
(18)

where, we calculate the estimation \(\hat{x}(k,0)\) by

$$\begin{aligned} \hat{x}(s+1,0)= & {} A\hat{x}(s,0)+K_{P}(s,0)\varepsilon (s,0) \end{aligned}$$
(19)

with initial value \(\hat{x}(k-d_{1}+1,0)=\hat{x}(k-d_{1}+1,1)\). \(\hat{x}(k-d_{1}+1,i)\) is gained by

$$\begin{aligned} \hat{x}(s+1,i)= & {} A\hat{x}(s,i)+K_{P}(s,i)\varepsilon (s,i) \end{aligned}$$
(20)

with \(\hat{x}(k-d_{i}+1,i)=\hat{x}(k-d_{i}+1,i+1)\). And one calculates \(\hat{x}(k-d_{l-1}+1,l-1)\) by

$$\begin{aligned} \hat{x}(s+1,l-1)= & {} A\hat{x}(s,l-1) +K_{P}(s,l-1)\varepsilon (s,l-1) \end{aligned}$$
(21)

with initial value \(\hat{x}(0,l-1)=0\).

Proof

Because \(\hat{x}(k\mid k)\) is the projection of x(k) onto the linear space of

$$\mathcal L \left\{ \{\varepsilon (s,l-1)\}^{k-d_{l-1}}_{s=0},\ldots , \{\varepsilon (s,i)\}^{k-d_{i}}_{s=k-d_{i+1}+1}, \ldots , \{\varepsilon (s,0)\}^{k}_{s=k-d_{1}+1}\right\} ,$$

according to the projection theory, one has

$$\begin{aligned} \hat{x}(k\mid k)= & {} proj\{x(k)\mid \{\varepsilon (s,l-1)\}^{k-d_{l-1}}_{s=0},\ldots , \{\varepsilon (s,i)\}^{k-d_{i}}_{s=k-d_{i+1}+1},\ldots , \nonumber \\&\{\varepsilon (s,0)\}^{k}_{s=k-d_{1}+1}\}\nonumber \\= & {} \hat{x}(k,0)+P_{0}(k)B_{0}^{T}\phi _{0}R_{\varepsilon (k,0)}^{-1}\varepsilon (k,0), \end{aligned}$$
(22)

therefore, (18) is proved. The proof of (19) and (20) can be yielded from (16). The proof is finished. \(\nabla \)

Remark 3

Applying the reorganized innovation approach, we have deduced the optimal filter by calculating Riccati equations (12) and one Lyapunov equation (15) of \(\mathbf {n\times n}\) dimension. On the other hand, if we let

$$X(k)=\small \left[ \begin{array}{cccc} x^{T}(k),x^{T}(k-d_{1}),\ldots ,x^{T}(k-d_{l-1}) \\ \end{array} \right] _{\mathbf {(d_{l-1}+1)n \times 1}}^{T},$$

the delayed measurement equation (2) can be converted into a delay-free equation. Further, the LMMSE estimator can be designed in terms of one Lyapunov equation and one Riccati equation of \(\mathbf {(d_{l-1}+1)n \times (d_{l-1}+1)n}\) dimension. Hence the high dimension Riccati equation related to the augmentation approach is avoided.

3.2 Design of the Optimal White Noise Estimator

We can export the optimal white-noise estimator \(\hat{n}(s\mid s+T)\) according to the innovation sequences \(\varepsilon (0)\), \(\varepsilon (1),\ldots , \varepsilon (s+T)\). When \(T\le 0 \), it can be observed that n(s) is independent of \(\varepsilon (0)\), \(\varepsilon (1),\ldots , \varepsilon (s+T)\). Then the estimator \(\hat{n}(s\mid s+T)\) is 0. When \(N>0\), the optimal input white-noise smoother \(\hat{n}(s\mid s+T)\) is defined following

$$\begin{aligned} \hat{n}(s\mid s+T)= & {} \hat{n}(s\mid s+T-1)+E\{n(s)\varepsilon ^{T}(s+T,i)\}R^{-1}_{\varepsilon (s+T,i)}\varepsilon (s+T,i), \end{aligned}$$
(23)

where \(E\{n(s)\varepsilon ^{T}(s+T,i)\}R^{-1}_{\varepsilon (s+T,i)}\) is to be determined, such that

$$\begin{aligned} E\{[n(s)-\hat{n}(s\mid s+T)][n(s)-\hat{n}(s\mid s+T)]^{T}\} \end{aligned}$$

is minimized. On the basis of the projection formula, we can calculate the optimal recursive input white-noise smoother \(\hat{n}(s\mid s+T)\).

Theorem 2

Consider the system (1)–(2), the optimal recursive input white-noise smoother is given following

$$\begin{aligned} \hat{n}(s\mid s+T)= & {} \hat{n}(s\mid s+T-1)+M_{n}(s+T,i)\varepsilon (s+T,i), T=1,2\ldots , \end{aligned}$$
(24)

the initial value \(\hat{n}(s\mid s)=0\), and the smoother gain \(M_{n}(s+T)\) satisfies the equation as follows

$$\begin{aligned} M_{n}(s+1,i)= & {} Q\bar{B}^{T}_{i}diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}R^{-1}_{\varepsilon (s+1,i)}, \nonumber \\ M_{n}(s+T,i)= & {} Q\prod _{l=1}^{T-1}\{\varPsi _{p}^{T}(s+l,i)\}\bar{B}^{T}_{i} diag\{\phi _{0}, \phi _{1},\ldots ,\phi _{i}\}R^{-1}_{\varepsilon (s+T,i)}, T>1 \nonumber \\ \varPsi _{p}(s,i)= & {} A-K_{p}(s,i)diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\bar{B}_{i}.\nonumber \end{aligned}$$

The covariance matrix \(P_{n}(s+T)\) is derived following

$$\begin{aligned} P_{n}(s+T)= & {} P_{n}(s+T-1)+M_{n}(s+T,i)R_{\varepsilon (s+T,i)}M^{T}_{n}(s+T,i) \end{aligned}$$
(25)

with initial value \(P_{n}(s)=0\).

Proof

Using the projection formula, one obtains

$$\begin{aligned} \hat{n}(s\mid s+T)= & {} \hat{n}(s\mid s+T-1)\nonumber \\&+E\{n(s)\varepsilon ^{T}(s+T,i)\}R^{-1}_{\varepsilon (s+T,i)}\varepsilon (s+T,i) \end{aligned}$$
(26)

where \(E\{n(s)\varepsilon ^{T}(s+T,i)\}R^{-1}_{\varepsilon (s+T,i)}\) is determined. We know that

$$\begin{aligned} \varepsilon (s+T,i)= & {} Y_{i}(s+T)-\hat{Y}_{i}(s+T)\nonumber \\= & {} diag\{\xi _{0}(s+T)- \phi _{0},\ldots ,\xi _{i}(s+T)-\phi _{i}\}\bar{B}_{i}x(s+T)\nonumber \\&+\,diag\{\phi _{0},\phi _{1},\ldots , \phi _{i}\}\bar{B}_{i}\tilde{x}(s+T,i) +V_{i}(s+T) \end{aligned}$$
(27)
$$\begin{aligned} \tilde{x}(s+1,i)= & {} x(s+1,i)-\hat{x}(s+1,i)\nonumber \\= & {} A\tilde{x}(s,i)+n(s)-K_{p}(s,i)\varepsilon (s,i)\nonumber \\= & {} \varPsi _{p}(s,i)\tilde{x}(s,i)+n(s)\nonumber \\&-\,K_{p}(s,i) [diag\{\xi _{0}(s)-\phi _{0},\ldots ,\xi _{i}(s)-\phi _{i}\} \bar{B}_{i}x(s)\nonumber \\&+\,V_{i}(s)] \end{aligned}$$
(28)

here, \(\varPsi _{p}(s,i)=A-K_{p}(s,i)diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\bar{B}_{i}.\) From (28), \(\tilde{x}(s+T,i)\) is derived as follows

$$\begin{aligned} \tilde{x}(s+T,i)= & {} \varPsi _{i}(s+T,s)\tilde{x}(s,i)+\sum _{l=s+1}^{s+T}\varPsi _{i}(s+T,l)[n(l-1)\nonumber \\&-\,K_{p}(l-1,i)(diag\{\xi _{0}(l-1)-\phi _{0},\ldots ,\xi _{i}(l-1)-\phi _{i}\}\nonumber \\&\bar{B}_{i}x(l-1)+V_{i}(l-1))],T>1, \end{aligned}$$
(29)

where,

$$\begin{aligned} \varPsi _{i}(s+T,s+T)= & {} I,\nonumber \\ \varPsi _{i}(s+T,l)= & {} \varPsi _{p}(s+T-1,i)\ldots \varPsi _{p}(l,i), l<s+T.\nonumber \end{aligned}$$

Putting (28), (29) into (27), one has

$$\begin{aligned} \varepsilon (s+1,i)= & {} diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\bar{B}_{i}\{\varPsi _{p}(s,i)\tilde{x}(s,i)+n(s)\nonumber \\&-\,K_{p}(s,i)[diag\{\xi _{0}(s)-\phi _{0},\ldots ,\xi _{i}(s)-\phi _{i}\}\bar{B}_{i}x(s)\nonumber \\&+\,V_{i}(s)]\}+V_{i}(s+1)\nonumber \\&+\,diag\{\xi _{0}(s+1)-\phi _{0},\ldots ,\xi _{i}(s+1)-\phi _{i}\}\bar{B}_{i}x(s+1), \nonumber \\ \varepsilon (s+T,i)= & {} diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}\bar{B}_{i} \{\varPsi _{i}(s+T,s)\tilde{x}(s,i)+\sum _{l=s+1}^{s+T}\varPsi _{i}(s+T,l)\nonumber \\&[n(l-1)-K_{p}(l-1,i) (diag\{\xi _{0}(l-1)-\phi _{0},\ldots ,\xi _{i}(l-1) -\phi _{i}\}\nonumber \\&\bar{B}_{i}x(l-1)+V_{i}(l-1))]\}+V_{i}(s+T)+diag\{\xi _{0}(s+T)-\phi _{0},\ldots ,\nonumber \\&\xi _{i}(s+T)-\phi _{i}\}\bar{B}_{i}x(s+T), T>1.\nonumber \end{aligned}$$

From above equations, it is readily to obtain

$$\begin{aligned} E\{n(s)\varepsilon ^{T}(s+1,i)\}= & {} Q\bar{B}^{T}_{i}diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}, \nonumber \\ E\{n(s)\varepsilon ^{T}(s+T,i)\}= & {} Q\prod _{l=1}^{T-1}\{\varPsi _{p}^{T}(s+l,i)\}\bar{B}^{T}_{i}diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}, T>1.\nonumber \end{aligned}$$

Considering that \(R_{\varepsilon (s+T, i)}\) is invertible, we define

$$\begin{aligned} M_{n}(s+1,i)= & {} Q\bar{B}^{T}_{i}diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}R^{-1}_{\varepsilon (s+1,i)}, \nonumber \\ M_{n}(s+T,i)= & {} Q\prod _{l=1}^{T-1}\{\varPsi _{p}^{T}(s+l,i)\}\bar{B}^{T}_{i}diag\{\phi _{0},\phi _{1},\ldots ,\phi _{i}\}R^{-1}_{\varepsilon (s+T, i)},T>1.\nonumber \end{aligned}$$

Next, we begin to derive the expression of \(P_{n}(s+T)\). According to (24), \(\tilde{n}(s\mid s+T)\) is gained by

$$\begin{aligned} \tilde{n}(s\mid s+T)= & {} \tilde{n}(s\mid s+T-1)-M_{n}(s+T,i)\varepsilon (s+T,i), \end{aligned}$$
(30)

so one gets (25). The proof is finished. \(\nabla \)

Remark 4

According to projection formula and the optimal state estimator in Theorem 1, one can gain the smoother in Theorem 2 directly. What has we finished in this paper is the state estimator and white noise estimator for finite horizon, while the related work for infinite horizon has not been finished. In the future, we will solve the problems about state estimator and white noise estimator for infinite horizon under the condition that A is a stable matrix.

Fig. 1
figure 1

The first state component \(x_{1}(k)\) and the filter \(\hat{x}(k\mid k)\)

Fig. 2
figure 2

The second state component \(x_{2}(k)\) and the filter \(\hat{x}(k\mid k)\)

Fig. 3
figure 3

The first state component \(x_{1}(k)\) and the filter \(\hat{x}(k\mid k)\)

Fig. 4
figure 4

The second state component \(x_{2} (k)\) and the filter \(\hat{x}(k\mid k)\)

4 Numerical Example

Consider the following linear measurement-delay systems

$$\begin{aligned} x(k+1)= & {} \left[ \begin{array}{cc} 0.2&{}0.12\\ 0.4&{}0.18 \end{array}\right] x(k)+n(k),\end{aligned}$$
(31)
$$\begin{aligned} y_{0}(k)= & {} \xi _{0}(k)\left[ \begin{array}{cc} 1&{}1.5\\ 2&{}1 \end{array}\right] x(k)+v_{0}(k),\end{aligned}$$
(32)
$$\begin{aligned} y_{1}(k)= & {} \xi _{1}(k) \left[ \begin{array}{cc} 1.2 &{}1.4\\ 2.3 &{}1.8 \end{array}\right] x(k-d)+v_{1}(k), \end{aligned}$$
(33)

with

\(x(0)=\left[ \begin{array}{c} -1\\ 1 \end{array}\right] ,\) \(\hat{x}(0,1)=\left[ \begin{array}{c} 0 \\ 0 \end{array}\right] \), \(D(0)=P(0)=\left[ \begin{array}{cc} 1 &{}-1\\ -1&{}1 \end{array}\right] \), \(\phi _{0}=diag\{0.85,0.73\}, \) \(\phi _{1}=diag\{0.79,0.89\} \), \(R_{0}=R_{1}=I_{2\times 2}\), \(Q=\left[ \begin{array}{cc} 1&{}1\\ 1&{}1\end{array}\right] \). n(k), \(v_{0}(k)\) and \(v_{1}(k)\) are white noises with zero mean and covariances Q, \(R_{0}\) and \(R_{1}\), respectively. We adopt d = 15, N = 100 in this simulation.

Primarily, according to Theorem 1, we give the simulation results in Figs. 1 and 2, where we can observe that the filter with additional measurement-delay channel represented by red line tracks the true state x(k) better than the filter without measurement-delay channel shown by green line. It suggests that the information coming from the measurement-delay channel is also important.

Then, we consider the influence of different packet dropping rate case,

$$\begin{aligned} Case~ 1: \phi _{0}=diag\{0.45,0.43\}, \phi _{1}=diag\{0.37,0.5\},\\ Case~ 2: \phi _{0}=diag\{0.83,0.75\}, \phi _{1}=diag\{0.78,0.9\}.\ \end{aligned}$$

The results are appeared in Figs. 3 and 4, where the filter for case 1 is represented by green line and the filter for case 2 is shown by the red line. It is shown in Figs. 3 and 4 that the filter for case 2 tracks the true state better. The reason is that the filter in case 2 is gained with more information than the filter in case 1.

5 Conclusion

In this paper, we have investigated the LMMSE state estimation and noise estimation for discrete-time system with time delay and random packet loss, which is modeled by an independent and identically distributed Bernoulli process. First, the optimal state estimator has been designed by using the reorganized innovation analysis approach. The solution to state estimator has been provided by calculating l Riccati difference equations and one Lyapunov difference equation. Next, an optimal input white-noise estimation has been put forward based on innovation analysis approach and state estimation mentioned above. Finally, a numerical example has been offered to show the effectiveness of the proposed approach. These works about the state estimator and noise estimator are only for finite horizon. In the future, we will solve problems about the state estimator and noise estimator for infinite horizon under the condition that A is a stable matrix.