1 Introduction

Multi-sensor fusion estimation is how to make the best use of the information contained in local sensors to obtain an optimal estimation of an event or a parameter [17]. Some local and fusion estimation methods have been studied (see [1, 9, 11, 13,14,15, 18, 21, 23, 24], and the references therein). Due to the rapid development of communication technology, communication network is widely used to connect individual sensors, because it can effectively improve the scalability and reduce wiring complexity. In this case, some networked multi-sensor fusion systems (NMFSs) are put forward for solving fusion problems, and they have found applications in a wide range, such as cyber-physical systems [5], Internet of Things [3] and SINR estimation [25]. Then, a great number of fusion estimation algorithms have been developed based on NMFSs (see [4, 6, 28, 29], and the references therein). However, along with the introduction of networks, some inevitable problems which will directly degrade the fusion estimation performance arise [2, 7, 22], such as bandwidth constraints, packet losses and time delays. Under this context, this paper will focus on the distributed fusion estimation for a class of NMFSs with limited communication capacity.

In general, there are two strategies to solve the problem of limited bandwidth: quantization [12, 16] and dimensionality reduction [10, 19]. The core idea of the dimensionality reduction is to convert a high-dimensional signal into a low-dimensional signal, while the idea of the quantization is to reduce the number of coding bits, but the dimension of the quantized signal is the same as that of the original signal. In this regard, to more greatly reduce communication traffic, the dual data compression strategy (DDCS) which includes both quantization and dimensionality reduction was proposed [8]. However, most of the above methods are based on stable systems, where the state estimates can be effectively quantized, while that of unstable systems cannot. Therefore, instead of directly quantizing the local estimates/measurements, the innovation quantization strategy was utilized to design the distributed fusion estimator in [26], but the dimensionality reduction scheme which can largely reduce the traffic was not considered.

Motivated by the aforementioned analysis, we shall study the distributed dimensionality reduction fusion estimation problem under quantized innovations, where the addressed dynamical systems are unstable. The main contributions of this paper can be summarized as follows: (1) The DDCS based on innovation is adopted to reduce the communication traffic for unstable NMFSs. (2) An uncertain approach is given to process the uncertainty of quantization, and a robust distributed dimensionality reduction fusion Kalman filter is derived under quantized innovations.

Notations The superscript “\({\mathrm {T}}\)” stands for the transpose, and “I” stands for the identity matrix. The notation \({\mathrm {diag\{}}\cdot {\mathrm {\}}}\) represents a block diagonal matrix, and \({\mathrm {col}}\{{a_1},\dots ,{a_n}\}\) represents a column vector whose elements are \({a_1},\dots ,{a_n}\). \({\mathrm {E}}\{\cdot \}\) denotes the mathematical expectation, while \({\mathrm {Tr}}\{\cdot \}\) represents the trace of matrix. The notation \({\mathrm {rank}}\{\cdot \}\) denotes the rank of matrix. \(X>(<)0\) means a positive-definite (negative-definite) matrix, while \(X\ge (\le )0\) means a non-negative definite (non-positive definite) matrix.

2 Problem Statement

Consider the following discrete-time state-space model:

$$\begin{aligned}&x(t+1)=Ax(t)+Bw(t) \end{aligned}$$
(1)
$$\begin{aligned}&y_i(t)=C_ix(t)+v_i(t)(i=1,2,\dots ,L). \end{aligned}$$
(2)

\(x(t)\in \mathbb {R}^n\) is the system state, and \(y_i(t)\in \mathbb {R}^{m_i}\) is the measurement of the ith sensor at time t. The addressed system is not necessarily stable, i.e., \(\lim \nolimits _{t\rightarrow \infty } x(t)\rightarrow \infty \) and \(\lim \nolimits _{t\rightarrow \infty } y_{i}(t)\rightarrow \infty \). Moreover, let L denote the number of sensors. A, B and \(C_i\) are constant matrices with appropriate dimensions. w(t) and \(v_i(t)\) are uncorrelated white Gaussian noises satisfying

$$\begin{aligned} \begin{array}{c} \mathrm{E}\{[w^{\mathrm{T}}(t)\ v_{i}^\mathrm{T}(t)]^{\mathrm{T}}[w^{\mathrm{T}}(t_{1})\ v_{j}^{\mathrm{T}}(t_{1})]\} =\delta _{tt_{1}}\mathrm{diag}\{Q,\delta _{ij}R_{i}\} \end{array} \end{aligned}$$
(3)

where \(\delta _{ij}=0\) and \(\delta _{tt_{1}}=0\) if \(i\ne j\) and \(t\ne t_{1}\); otherwise, \(\delta _{ij}=1\) and \(\delta _{tt_{1}}=1\). Assume that (AB) is controllable and \((A,C_{i})\) is observable, that is

$$\begin{aligned} \left\{ \begin{array}{ll} \mathrm{rank}\{[B\ AB\ \cdots \ A^{n-1}B]\}=n \\ \mathrm{rank}\{\mathrm{col}\{C_{i},C_{i}A,\dots ,C_{i}A^{n-1}\}\}=n \end{array}. \right. \end{aligned}$$
(4)

In the sequel, the local estimate \(\hat{x}_i(t)\) at the ith sensor is given by the Kalman structure:

$$\begin{aligned} \begin{array}{ll} \hat{x}_i(t+1)=A\hat{x}_i(t)+K_{i}(t+1)\varepsilon _i(t+1) \end{array} \end{aligned}$$
(5)

where \(\varepsilon _i(t)\) represents the innovation sequence, and it can be described by the following form:

$$\begin{aligned} \begin{array}{ll} \varepsilon _i(t)=y_{i}(t)-C_iA\hat{x}_i(t-1) \end{array}. \end{aligned}$$
(6)

2.1 DDCS Based on Innovation

Before each sensor message is transmitted to the FC via constrained communication channels, the following DDCS is proposed to satisfy the finite bandwidth.

According to the idea of dimensionality reduction in [10], it is similarly proposed in this paper that \(r_i (1\le r_i<m_i)\) components of the ith innovation \(\varepsilon _i(t)\) are selected to be transmitted at a particular time. In this case, the allowed sending components have \(\theta _i\) possible cases, where \(\theta _i\) is taken as

$$\begin{aligned} \begin{array}{l} \theta _i=\frac{m_i!}{r_{i}!(m_i-r_{i})!} \end{array}. \end{aligned}$$
(7)

Hence, the reorganized innovation (RI) \(\varepsilon _{i}^{r}(t)\) can only take one signal from the following finite set:

$$\begin{aligned} \begin{array}{l} \chi _{i}(t)= \{\varTheta _{1}^{i}(t)\varepsilon _{i}(t),\dots , \varTheta _{h_{i}}^{i}(t)\varepsilon _{i}(t),\dots , \varTheta _{\theta _{i}}^{i}(t)\varepsilon _{i}(t)\} \end{array} \end{aligned}$$
(8)

where \(\varTheta _{h_{i}}^{i}\) denotes a diagonal matrix which contains \(r_{i}\) diagonal elements “1” and \(m_{i}-r_{i}\) diagonal elements “0.”

To describe the RI in a simple way, suppose that \(\theta _{i}\) elements of set \(\chi _{i}\) are indexed from 1 to \(\theta _{i}\). Then, the following indicator function is introduced:

$$\begin{aligned} \begin{array}{c} \sigma _{h_{i}}^{i}(t)= \left\{ \begin{array}{c} 1\quad {\hbox {if}}\,\, \varepsilon _{i}^{r}(t)=\varTheta _{h_{i}}^{i}(t)\varepsilon _{i}(t) \\ 0\quad {\hbox {if}}\,\, \varepsilon _{i}^{r}(t)\ne \varTheta _{h_{i}}^{i}(t)\varepsilon _{i}(t) \end{array} \right. \end{array} \end{aligned}$$
(9)

It can be realized from (9) that if the \(h_{i}\)th element of set \(\chi _{i}(t)\) is chosen as \(\varepsilon _{i}^{r}(t)\), \(\sigma _{h_{i}}^{i}(t)=1\); otherwise, \(\sigma _{h_{i}}^{i}(t)=0\). Meanwhile, the binary variables \(\sigma _{h_{i}}^{i}(t)(h_{i}=1,2,\dots ,\theta _{i})\) satisfy

$$\begin{aligned} \begin{array}{c} \sigma _{h_{i}}^{i}(t)\sigma _{h_{i0}}^{i}(t)=0(h_{i}\ne h_{i0}), \sum \limits _{h_i=1}^{\theta _i}\sigma _{h_{i}}^{i}(t)=1 \end{array}. \end{aligned}$$
(10)

It can be guaranteed from (10) that the RI only takes one signal from the set \(\chi _{i}(t)\) at each time. Then, the RI \(\varepsilon _{i}^{r}(t)\) can be calculated by the following form:

$$\begin{aligned} \begin{array}{c} \varepsilon _{i}^{r}(t)=\varTheta _{i}(t)\varepsilon _{i}(t) \end{array} \end{aligned}$$
(11)

where \(\varTheta _{i}(t)\) is described by:

$$\begin{aligned} \begin{array}{c} \varTheta _i(t) =\sum \limits _{h_i=1}^{\theta _i}\sigma ^i_{h_i}(t)\varTheta ^{i}_{h_i}(t) \\ \end{array}. \end{aligned}$$
(12)

Since \(\varTheta _{i}(t)\) is diagonal, it can be simplified as follows:

$$\begin{aligned} \begin{array}{c} \varTheta _i(t) =\mathrm{diag}\{\xi _{i1}^{h_i}(t),\xi _{i2}^{h_i}(t),\dots ,\xi _{im_i}^{h_i}(t)\} \\ \end{array} \end{aligned}$$
(13)

where \(\xi _{ij}^{h_i}(t)(j=1,2,\dots ,m_{i})\) are the binary variables satisfying

$$\begin{aligned} \begin{array}{c} \xi _{ij}^{h_i}(t)\in \{0,1\}, \sum \limits _{j=1}^{m_i}\xi _{ij}^{h_i}(t)=r_i \end{array}. \end{aligned}$$
(14)
Fig. 1
figure 1

The relationship between the input and output of the uniform quantizer

Subsequently, the RI will be quantized after dimensionality reduction, and Fig. 1 shows the relationship between the input and the output of uniform quantizer. Here, the uniform quantizer can be described by [20]:

$$\begin{aligned} \begin{array}{c} Q(x)=\bar{x}+\mathrm{sgn}(x-\bar{x})\cdot \delta \cdot \lfloor {\frac{\left\| x-\bar{x}\right\| }{\delta }+\frac{1}{2}} \rfloor \end{array} \end{aligned}$$
(15)

where \(\mathrm{sgn}\{\cdot \}\) is the sign function. x denotes the input, while \(\bar{x}\) denotes the mid-value of certain interval. The parameter \(\delta =\frac{l}{2}\) is the maximum error, where l is the length of the quantization interval. Thus, the quantization error \(\varDelta \) is bounded, that is

$$\begin{aligned} \begin{array}{l} \left\| \varDelta \right\| =\left\| x-Q(x)\right\| <\delta =\frac{l}{2} \end{array}. \end{aligned}$$
(16)

Accordingly, Q(x) can be expressed as \((I+\varDelta )x\) for certain \(\varDelta \) which satisfies (16).

Under the above analysis, the RI after quantization becomes the quantized innovation (QI) \(\varepsilon _{i}^{q}(t)\), i.e., the innovation based on the DDCS, and the QI can be expressed by the following form:

$$\begin{aligned} \begin{array}{c} \varepsilon _{i}^{q}(t)=Q(\varepsilon _{i}^{r}(t)) =(I+\varDelta _i(t))\varTheta _i(t)\varepsilon _i(t) =(I+H_{i}(t)F_{i}(t)E_{i}(t))\varTheta _i(t)\varepsilon _i(t) \end{array} \end{aligned}$$
(17)

where \(H_{i}(t),F_{i}(t),E_{i}(t)\) are known matrices with appropriate dimensions.

Remark 1

Generally, local estimates were quantized for satisfying the limited bandwidth [10]. But in this paper, the innovations were proposed to be quantized through finite communication channels. In fact, the inputs of quantizer should be bounded at each time. This means that the quantization density and the upper bound should be determined in advance, and then the quantizer can correctly judge which interval the output is in. However, for unstable systems, the estimates/measurements will diverge as time goes to \(\infty \). As an alternative, innovation sequences can be bounded due to the stability of Kalman filtering when the system is controllable and observable. In this case, the innovation quantization strategy was adopted in [26] such that the quantization method can be applicable for unstable NMFSs. Although the fusion estimation problem was also discussed in [26], the dimensionality reduction strategy which greatly reduces the traffic more than the quantization strategy was not considered. Therefore, the DDCS based on innovation is proposed in this paper.

Remark 2

Obviously, with the dimension reduced by dimensionality reduction strategy, the amount of useful information will be simultaneously lost and the estimation performance in FC will be compromised. To overcome this shortage, the one-step prediction compensation strategy is commonly proposed to reduce the loss [24]. However, different from the above-mentioned view, there is no need to compensate the reorganized information caused by dimensionality reduction in this paper, because it is the innovations instead of the estimates that are going to be transmitted. In this case, whether the FC receives the innovations or not, the one-step prediction will be computed when recursing Kalman filter, i.e., the local estimates, will be potentially replaced by the one-step predictions. Moreover, the stability and the accuracy are influenced by the dimension \(r_{i}\). Hence, the dimension \(r_{i}\) needs to be carefully determined to balance the communication traffic and the accuracy.

2.2 Problem of Interests

Consider the following local filter with quantized innovation [26]:

$$\begin{aligned} \begin{array}{c} \hat{x}_i(t+1)=A\hat{x}_i(t)+K_i(t+1)\varepsilon _{i}^{q}(t+1) \end{array} \end{aligned}$$
(18)

where \(\hat{x}_i(t+1)\) is the local state estimate and \(K_i(t+1)\) is the optimal Kalman filter gain. Notice that the Kalman gain in the linear minimum variance sense can be solved in two steps. First, a filter with quantized innovation needs to be designed to eliminate the multiplicative uncertainties such that for all admissible uncertainties, there exists a positive-definite matrix \(\tilde{P}_{ii}(t)\) satisfying

$$\begin{aligned} \begin{array}{c} \bar{P}_{ii}(t) =\mathrm{E}\{(x_i(t)-\hat{x}_i(t))(x_i(t)-\hat{x}_i(t))^{\mathrm{T}}\} \le \tilde{P}_{ii}(t) \end{array} \end{aligned}$$
(19)

because

$$\begin{aligned} \begin{array}{c} \mathrm{Tr}\{\bar{P}_{ii}(t)\} =\mathrm{E}\{(x_i(t)-\hat{x}_i(t))^{\mathrm{T}}(x_i(t)-\hat{x}_i(t)\} \le \mathrm{Tr}\{\tilde{P}_{ii}(t)\} \end{array} \end{aligned}$$
(20)

where \(\bar{P}_{ii}(t)\) represents the state estimation error covariance matrix of ith sensor, and \(\tilde{P}_{ii}(t)\) is the corresponding finite upper bound. Second, we shall minimize \(\mathrm{Tr}(\tilde{P}_{ii}(t))\) to calculate the optimal gain \(K_{i}(t)\) and \(P_{ii}(t)\), the minimum of upper bound, and then construct the optimal recursive Kalman filter.

After QI has been sent to the FC through constrained communication channel, the local estimate recursively calculated in FC, denoted as \(\hat{x}_{fi}(t+1)\), will be described by the following form:

$$\begin{aligned} \begin{array}{c} \hat{x}_{fi}(t+1)=A\hat{x}_{fi}(t)+K_{fi}(t+1)\varepsilon _{i}^{q}(t+1) \end{array}. \end{aligned}$$
(21)

When \(\hat{x}_{fi}(0)=\hat{x}_{i}(0)\), there must be \(\hat{x}_{fi}(t+1)=\hat{x}_{i}(t+1)\) at each time. In the sequel, the weighting fusion estimate in FC can be calculated by:

$$\begin{aligned} \begin{array}{c} \hat{x}_{m}(t)=\sum \limits _{i=1}^L W_i(t)\hat{x}_{fi}(t) \end{array} \end{aligned}$$
(22)

where \(W_{i}(t)(i=1,\ldots ,L)\) are weight matrices meet \(\sum \nolimits _{i=1}^L W_i(t)=I\), and they need to be designed by the linear unbiased minimum variance criterion. Then, the distributed dimensionality reduction fusion structure with quantized innovations is shown in Fig. 2.

Fig. 2
figure 2

The distributed dimensionality reduction fusion structure with quantized innovations

Consequently, the problems to be solved in this paper are summarized as follows: (1) Design a dimensionality reduction Kalman filter with quantized innovation based on the model uncertainty processing method in each sensor. (2) Design a distributed dimensionality reduction fusion Kalman filtering with quantized innovations based on the local estimates and the fusion criterion.

Remark 3

In most cases, the stable local estimator with standard form of Kalman filter is designed by \(\hat{x}_{i}^{s}(t+1)=A\hat{x}_i(t)+K_{i}^{s}(t+1)\varepsilon _i(t+1)\). However, the stability of the local estimator \(\hat{x}_{fi}(t+1)\) as (21) in FC cannot be guaranteed due to the introduction of DDCS. Suppose that the Kalman gain in FC is also computed as \(K_{i}^{s}(t)\). Then, to prove the instability between the standard Kalman filter and filter (21), their relative error \(\delta _i(t+1)\) is described as follows:

$$\begin{aligned} \begin{array}{ll} \delta _i(t+1) &{}=\hat{x}_{i}^{s}(t+1)-\hat{x}_{fi}(t+1) \\ &{}=A\delta _i(t)+K_{i}^{s}(t+1)\varepsilon _i(t+1)(I-(I+\varDelta _i(t+1))\varTheta _i(t+1)) \\ &{}=A^{t+1}\delta _i(0)+\sum \limits _{n=0}^{t}A^{t-n}K_{i}^{s}(n+1)\varepsilon _i(n+1)\varDelta _i(n+1)\varTheta _i(n+1). \end{array} \end{aligned}$$
(23)

When \(\delta _i(0)=0\), \(\delta _i(t)\) cannot be 0 due to the existence of quantization noise \(\varDelta _i(t)\) and dimensionality reduction matrix \(\varTheta _i(t)\). Nevertheless, if the standard local Kalman filter is replaced by the proposed filter with quantized innovation as (18), \(\delta _i(t+1)\) will become

$$\begin{aligned} \begin{array}{ll} \delta _i(t+1)&{}=\hat{x}_i(t+1)-\hat{x}_{fi}(t+1) \\ &{}=A\delta _i(t) \\ &{}=A^{t+1}\delta _i(0). \end{array} \end{aligned}$$
(24)

It is obvious that when \(\delta _i(0)=0\), \(\delta _i(t)=0\). In other words, the local filters recursively calculated in FC will be consistent with that in corresponding sensors. Thus, in this paper, recursive form (18) is considered for local state estimation instead of the standard Kalman filter.

3 Main Results

In this section, the distributed dimensionality reduction fusion Kalman filter with quantized innovations is designed for unstable NMFSs. Define the estimation error \(e_{i}(t)=x(t)-\hat{x}_{i}(t)\). Then, the optimal weight matrices \(W_{i}(t)\) in (22) can be obtained by resorting to the results in [21]:

$$\begin{aligned} \begin{array}{c} W(t) =(I_{a}^{\mathrm{T}}{P}^{-1}(t)I_{a})^{-1}I_{a}^{\mathrm{T}}{P}^{-1}(t) \end{array} \end{aligned}$$
(25)

where

$$\begin{aligned}&\begin{array}{c} I_{a}= [ \underbrace{I_{n}^{\mathrm{T}}\cdots I_{n}^{\mathrm{T}}}_{L\ times} ] \end{array} \end{aligned}$$
(26)
$$\begin{aligned}&\begin{array}{c} W(t)= \begin{bmatrix} W_{1}(t)\dots W_{i}(t)\cdots W_{L}(t) \end{bmatrix} \end{array} \end{aligned}$$
(27)
$$\begin{aligned}&\begin{array}{c} P(t)=\mathrm{E}\{[e_{1}^{\mathrm{T}}(t)\dots e_{L}^{\mathrm{T}}(t)] [e_{1}^{\mathrm{T}}(t)\dots e_{L}^{\mathrm{T}}(t)]^{\mathrm{T}}\}. \end{array} \end{aligned}$$
(28)

Before deriving the main results, the following lemma is introduced.

Lemma 1

[27] Given matrices A, H and E with compatible dimensions such that \(FF^\mathrm{T}\le I\). Let X be a symmetric positive-definite matrix and \(\alpha >0\) be an arbitrary positive constant such that \(\alpha ^{-1}I-EXE^\mathrm{T}>0\), then the following inequality holds:

$$\begin{aligned}&(A+HFE)X(A+HFE)^{\mathrm{T}}\nonumber \\&\quad \le AXA^{\mathrm{T}}+AXE^{\mathrm{T}}(\alpha ^{-1}I-EXE^{\mathrm{T}})^{-1}EXA^{\mathrm{T}}+\alpha ^{-1}HH^{\mathrm{T}}. \end{aligned}$$
(29)

Theorem 1

Given positive parameters \(\alpha _{ii}\), \(\beta _{ii}\) and \(\gamma _{ii}\) which satisfy

$$\begin{aligned} \left\{ \begin{array}{l} \alpha _{ii}^{-1}I-\tilde{E}_{i1}(t)P_{ii}(t)\tilde{E}_{i1}^{\mathrm{T}}(t)>0 \\ \beta _{ii}^{-1}I-\tilde{E}_{i2}(t)Q\tilde{E}_{i2}^{\mathrm{T}}(t)>0 \\ \gamma _{ii}^{-1}I-\tilde{E}_{i3}(t)R_{i}\tilde{E}_{i3}^{\mathrm{T}}(t)>0 \end{array}. \right. \end{aligned}$$
(30)

Then, the optimal estimator gain \(K_{i}(t+1)\) in (18) is calculated by:

$$\begin{aligned}&K_i(t+1) =[AS_{ii1}(t)A^{\mathrm{T}}C_{i}^{\mathrm{T}}\varTheta _i(t+1)^{\mathrm{T}} \nonumber \\&\quad \quad +\,BS_{ii2}(t)B^{T}C_i^{\mathrm{T}}\varTheta _i(t+1)^{\mathrm{T}}] [C_{i}AS_{ii1}(t)A^{\mathrm{T}}C_i^{\mathrm{T}}\nonumber \\&\qquad +\,C_{i}BS_{ii2}(t)B^{\mathrm{T}}C_i^{\mathrm{T}}+S_{ii3}(t) +(\alpha _{ii}^{-1}+\beta _{ii}^{-1}+\gamma _{ii}^{-1}){H}_{i}(t){H}_{i}(t)^{\mathrm{T}}]^{-1} \end{aligned}$$
(31)
$$\begin{aligned}&P_{ii}(t+1) =\tilde{A}_{i1}(t)S_{ii1}(t)\tilde{A}_{i1}^{\mathrm{T}}(t) +\tilde{A}_{i2}(t)S_{ii2}(t)\tilde{A}_{i2}^{\mathrm{T}}(t)\nonumber \\&\quad \quad +\,\tilde{A}_{i3}(t)S_{ii3}(t)\tilde{A}_{i3}^{\mathrm{T}}(t) \nonumber \\&\qquad +\,\alpha _{ii}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_i(t) +\beta _{ii}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_i(t)\nonumber \\&\quad \quad +\,\gamma _{ii}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_i(t) \end{aligned}$$
(32)

where

$$\begin{aligned} \left\{ \begin{array}{l} \tilde{A}_{i1}(t)=A-K_{i}(t+1)\varTheta _i(t+1)C_iA \\ \tilde{A}_{i2}(t)=B-K_{i}(t+1)\varTheta _i(t+1)C_iB \\ \tilde{A}_{i3}(t)=-K_{i}(t+1)\varTheta _i(t+1) \\ \tilde{H}_{i}(t)=-K_{i}(t+1)H_{i}(t+1) \\ \tilde{E}_{i1}(t)=E_{i}(t+1)\varTheta _i(t+1)C_iA \\ \tilde{E}_{i2}(t)=E_{i}(t+1)\varTheta _i(t+1)C_iB \\ \tilde{E}_{i3}(t)=E_{i}(t+1)\varTheta _i(t+1) \\ \end{array} \right. \end{aligned}$$
(33)

and

$$\begin{aligned} \left\{ \begin{array}{l} S_{ii1}(t)=P_{ii}(t)+P_{ii}(t)\tilde{E}_{i1}^{\mathrm{T}}(t)[\alpha _{ii}^{-1}I -\tilde{E}_{i1}(t)P_{ii}(t)\tilde{E}_{i1}^{\mathrm{T}}(t)]^{-1}\tilde{E}_{i1}(t)P_{ii}(t) \\ S_{ii2}(t)=Q+Q\tilde{E}_{i2}^{\mathrm{T}}(t)[\beta _{ii}^{-1}I -\tilde{E}_{i2}(t)Q\tilde{E}_{i2}^{\mathrm{T}}(t)]^{-1}\tilde{E}_{i2}(t)Q \\ S_{ii3}(t)=R_{i}+R_{i}\tilde{E}_{i3}^{\mathrm{T}}(t)[\gamma _{ii}^{-1}I -\tilde{E}_{i3}(t)R_{i}\tilde{E}_{i3}^{\mathrm{T}}(t)]^{-1}\tilde{E}_{i3}(t)R_{i} \\ \end{array}. \right. \end{aligned}$$
(34)

Meanwhile, after similarly giving \(\alpha _{ij}\), \(\beta _{ij}\) and \(\gamma _{ij}\), the cross-covariance matrix is computed by:

$$\begin{aligned} P_{ij}(t+1)= & {} \tilde{A}_{i1}(t)S_{ij1}(t)\tilde{A}_{j1}^{\mathrm{T}}(t) +\tilde{A}_{i2}(t)S_{ij2}(t)\tilde{A}_{j2}^{\mathrm{T}}(t)\nonumber \\&+\,\tilde{A}_{i3}(t)S_{ij3}(t)\tilde{A}_{j3}^{\mathrm{T}}(t) \nonumber \\&+\,\alpha _{ij}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_j(t) +\beta _{ij}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_j(t)\nonumber \\&+\,\gamma _{ij}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_j(t) \end{aligned}$$
(35)

where

$$\begin{aligned} \left\{ \begin{array}{l} S_{ij1}(t)=P_{ij}(t)+P_{ij}(t)\tilde{E}_{j1}^{\mathrm{T}}(t)[\alpha _{ij}^{-1}I -\tilde{E}_{i1}(t)P_{ii}(t)\tilde{E}_{j1}^{\mathrm{T}}(t)]^{-1}\tilde{E}_{i1}(t)P_{ij}(t) \\ S_{ij2}(t)=Q+Q\tilde{E}_{j2}^{\mathrm{T}}(t)[\beta _{ij}^{-1}I -\tilde{E}_{i2}(t)Q\tilde{E}_{j2}^{\mathrm{T}}(t)]^{-1}\tilde{E}_{i2}(t)Q \\ S_{ij3}(t)=0 \end{array}. \right. \end{aligned}$$
(36)

Proof

Under the distributed fusion structure, the fusion algorithm is based on the local estimates and weighting criterion. To start, let us define the original local Kalman filter before minimization:

$$\begin{aligned} \begin{array}{c} \bar{x}_i(t+1)=A\hat{x}_i(t)+\bar{K}_i(t+1)\varepsilon _{i}^{q}(t+1) \end{array} \end{aligned}$$
(37)

where \(\bar{x}_i(t+1)\) and \(\bar{K}_i(t+1)\), respectively, represent the estimate and gain before minimization. The corresponding state estimation error is denoted by \(\bar{e}_{i}(t+1)\triangleq x_{i}(t+1)-\bar{x}_{i}(t+1)\). After expanding uncertain term as (17), \(\bar{e}_{i}(t+1)\) will become:

$$\begin{aligned} \begin{array}{ll} \bar{e}_i(t+1) &{}=(\tilde{A}_{i1}(t)+\tilde{H}_{i}(t)F_{i}(t+1)\tilde{E}_{i1}(t))e_{i}(t) \\ &{}\quad +\,(\tilde{A}_{i2}(t)+\tilde{H}_{i}(t)F_{i}(t+1)\tilde{E}_{i2}(t))w(t) \\ &{}\quad +\,(\tilde{A}_{i3}(t)+\tilde{H}_{i}(t)F_{i}(t+1)\tilde{E}_{i3}(t))v_{i}(t+1) \end{array} \end{aligned}$$
(38)

where \(\tilde{A}_{i1}(t),\tilde{A}_{i2}(t),\tilde{A}_{i3}(t), \tilde{H}_{i}(t),\tilde{E}_{i1}(t),\tilde{E}_{i2}(t)\) and \(\tilde{E}_{i3}(t)\) have been defined in (33). As analyzed in (17), \(H_{i}(t+1)F_{i}(t+1)E_{i}(t+1)=\varDelta _{i}(t+1)\) represents the quantization noise. More concretely, \(F_{i}(t+1)=\frac{\varDelta _{i}(t+1)}{\delta }\) meets the constraint \(F_{i}(t+1)F_{i}^\mathrm{T}(t+1)\le I\), and \(H_{i}(t+1)=\delta ,E_{i}(t+1)=I\). Then, after giving positive parameters \(\alpha _{ii}\), \(\beta _{ii}\) and \(\gamma _{ii}\) which satisfy (30), the upper bound \(\tilde{P}_{ii}(t+1)\) of \(\bar{P}_{ii}(t+1)\triangleq \mathrm{E}\{\bar{e}_{i}(t+1)\bar{e}_{i}^{\mathrm{T}}(t+1)\}\), the original estimation error covariance matrix, can be derived by utilizing Lemma 1:

$$\begin{aligned} \begin{array}{lll} \bar{P}_{ii}(t+1) &{}\le \tilde{A}_{i1}(t)S_{ii1}(t)\tilde{A}_{i1}^{\mathrm{T}}(t) +\tilde{A}_{i2}(t)S_{ii2}(t)\tilde{A}_{i2}^{\mathrm{T}}(t) +\tilde{A}_{i3}(t)S_{ii3}(t)\tilde{A}_{i3}^{\mathrm{T}}(t) \\ &{}\quad +\,\alpha _{ii}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_i(t) +\beta _{ii}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_i(t) +\gamma _{ii}^{-1}\tilde{H}_i(t)\tilde{H}^{\mathrm{T}}_i(t) \\ &{}=\tilde{P}_{ii}(t+1) \end{array} \end{aligned}$$
(39)

where \(S_{ii1}(t),S_{ii2}(t)\) and \(S_{ii3}(t)\) have been defined in (34). To obtain the minimum covariance \(P_{ii}(t)\), the first-order partial derivative about \(\bar{K}_{i}(t+1)\) of \({\mathrm{Tr}}\{\tilde{P}_{ii}(t+1)\}\) is given as follows:

$$\begin{aligned} \frac{\partial {\mathrm{Tr}}\{\tilde{P}_{ii}(t+1)\}}{\partial \bar{K}_i(t+1)}= & {} 2\bar{K}_i(t+1)\varTheta _i(t+1)C_iAS_{ii1}(t)A^{\mathrm{T}}C_{i}^{\mathrm{T}}\varTheta _i(t+1)^{\mathrm{T}} \nonumber \\&-2AS_{ii1}(t)A^{\mathrm{T}}C_{i}^{\mathrm{T}}\varTheta _i(t+1)^{\mathrm{T}} +2\alpha _{ii}^{-1}\tilde{H}_{i}(t)\tilde{H}_{i}(t)^{\mathrm{T}} \nonumber \\&+2\bar{K}_i(t+1)\varTheta _i(t+1)C_iBS_{ii2}(t)B^{\mathrm{T}}C_{i}^{T}\varTheta _i(t+1)^{\mathrm{T}} \nonumber \\&-2BS_{ii2}(t)B^{\mathrm{T}}C_{i}^{\mathrm{T}}\varTheta _i(t+1)^{\mathrm{T}} +2\beta _{ii}^{-1}\tilde{H}_{i}(t)\tilde{H}_{i}(t)^{\mathrm{T}} \nonumber \\&+2\bar{K}_i(t+1)\varTheta _i(t+1)S_{ii3}(t)\varTheta _i(t+1)^{\mathrm{T}} +2\gamma _{ii}^{-1}\tilde{H}_{i}(t)\tilde{H}_{i}(t)^{\mathrm{T}}\nonumber \\ \end{aligned}$$
(40)

When (40) equals zero, the optimal Kalman gain \(K_{i}(t+1)\) can be derived in the linear minimum variance sense as (31), and the minimum upper bound \(P_{ii}(t+1)\) can be obtained by substituting the optimal Kalman gain \(K_{i}(t+1)\) into (39).

On the other hand, when the FC receives the quantized innovations, (21) is exploited to calculate the local estimate \(\hat{x}_{fi}(t+1)\). According to Remark 3, the local estimates are relatively invariant during the transmission process. Thus, \(P_{ij}(t+1)\), the minimum upper bound of cross-covariance and the composite covariance matrix \(P(t+1)\) can be calculated by (35) and (28). Finally, the optimal weight matrix \(W(t+1)\) in the linear unbiased minimum variance sense is derived by (25), and the optimal fusion estimate is calculated by (22). This completes the proof. \(\square \)

Based on Theorem 1, the computation procedures for distributed dimensionality reduction fusion Kalman filtering with quantized innovations are summarized as Algorithm 1.

figure a

4 Simulation Examples

Consider a target tracking system with two sensors, and the system parameters in (1) and (2) are given by [1]:

$$\begin{aligned} \begin{array}{l} A= \begin{bmatrix} 1 &{} \quad T_0 &{} \quad \frac{T_0^2}{2} \\ 0 &{} \quad 1 &{} \quad T_0 \\ 0 &{} \quad 0 &{} \quad 1 \end{bmatrix}, B= \begin{bmatrix} \frac{T_0^2}{2} \\ T_0 \\ 1 \\ \end{bmatrix}, C_1= \begin{bmatrix} 1 &{} \quad 1 &{} \quad 0 \\ 0 &{} \quad 1 &{} \quad 1 \end{bmatrix}, C_2= \begin{bmatrix} 0 &{} \quad 1&{} \quad 1 \\ 1 &{} \quad 1&{} \quad 0 \end{bmatrix} \end{array} \end{aligned}$$
(41)

where \(T_0=0.8s\) is the sampling period. Denote the state vector \(x(t)=\mathrm{col}\{X(t),\dot{X}(t),\ddot{X}(t)\}\), where \(X(t),\dot{X}(t),\ddot{X}(t)\) are the position, velocity and acceleration of moving target at time t, respectively. w(t), \(v_1(t)\) and \(v_2(t)\) are respective uncorrelated white Gaussian noises with covariances

$$\begin{aligned} \begin{array}{l} Q=1, R_{1}=\begin{bmatrix} 0.9 &{} \quad 0 \\ 0 &{} \quad 0.7 \end{bmatrix}, R_{2}=\begin{bmatrix} 0.5 &{} \quad 0 \\ 0 &{} \quad 0.3 \end{bmatrix} \end{array}. \end{aligned}$$
(42)

These random variables can be generated by the function “rand()” of MATLAB.

Suppose that each sensor has enough processing capabilities to compute local Kalman filter. It is considered for DDCS that only \(r_{1}=1\) components of \(\varepsilon _{1}(t)\) are allowed to be transmitted to the FC at time t, and the quantization parameter is given by \(l=0.01\). For example, the original innovation \(\varepsilon _{1}(t)\) is \(\mathrm{col}\{-0.6412,-0.7150\}\) which can be represented by twelve decimal bits. After dimensionality reduction strategy with \(\left( \varTheta _{1} (t)={\mathrm {diag}}\{1,0\}\right) ,\) the RI \(\varepsilon _{1}^{r}(t)\) becomes \(\mathrm{col}\{-0.6412,0\}\). Then, after quantization, the QI \(\varepsilon _{1}^{q}(t)\) becomes \(\mathrm{col}\{-0.64,0\}\) which can only be represented by four decimal bits. Therefore, the transmission amount is greatly reduced.

To demonstrate Remark 1, the following figures are presented. Figure 3a shows the local estimates and innovations in sensor 1. It can be seen from this figure that in unstable systems, the state estimates diverge, while the innovations are bounded. Meanwhile, it is the same in Fig. 3b which shows the quantized value of local estimates and innovations. Accordingly, this result proves the reason why quantizing innovations instead of states in unstable NMFSs.

Fig. 3
figure 3

a The position and innovation in unstable systems. b The quantization of position and innovation in unstable systems

To demonstrate the effectiveness of the designed fusion estimation algorithm, the following figures are presented. Figure 4a shows the trajectories of target x(t) and fusion tracking \(\hat{x}_{m}(t)\), and it illustrates the designed fusion estimate can catch up with the target well under bandwidth constraints. Due to the existence of stochastic noises, the performance is assessed by its mean square errors (MSEs) over an average of 100 runs Monte Carlo method. It is obviously seen in Fig. 4b that the MSE of fusion estimator is lower than that of any local estimator, and this means the fusion estimate performance better than any local estimate. Moreover, Fig. 4c shows that the MSE decreases with the increase in the dimension of innovation. According to the accuracy variation shown in this figure, the dimension \(r_{i}\) can be determined suitably.

Fig. 4
figure 4

a The trajectories of target and fusion tracking. b The MSEs of local and fusion estimates. c The relationship between the MSE and the dimension of innovation

5 Conclusions

In this paper, the distributed fusion Kalman filtering problem has been studied for a class of unstable NMFSs with communication constraints, where the DDCS was introduced to deal with the innovations. By using the uncertainty processing method, the optimal local recursive form of minimum upper bound of covariance and Kalman filter gain were obtained. Then, the distributed dimensionality reduction fusion Kalman filter with quantized innovations was designed by utilizing the optimal weighted matrix fusion criterion. Finally, it was illustrated from the simulation examples that the proposed fusion estimation algorithm can efficiently solve the estimator design problem for unstable NMFSs.