1 Introduction

Structural health monitoring (SHM) is an emerging field which aims at implementing damage detection and characterization strategies in various engineering structures and systems. Examples of such systems include buildings, pipelines, highways, stadiums, and others. SHM is implemented by placing sensors on these structures to collect data that can be used to make important decisions about the condition and characteristics of these structures and systems. With the advancements in instrumentation and data acquisition techniques, monitoring of engineering structures has become cheaper, more feasible, and gained more popularity. In modal parameter estimation, the SHM system can be ill-conditioned due to uncertainties in the measurements [2, 3]. The potential errors in structural model updating combined with aleatory uncertainty in modal parameter estimation often results in inconsistencies between the real structural behavior and the finite element model predictions [4, 5]. However, in most of the published research in SHM, the ill-conditioned state and the presence of uncertainty are not considered [6]. It is also known that the dynamic responses of the structures are excitation amplitude dependent. Different levels of excitation amplitudes were reported to result in different modal damping ratios [7]. These inconsistencies in conjunction with the noisy data lead to analytical models that are sometimes far from the real behavior of the structure. This would result in wrong decisions by the SHM decision-makers and of course not contributing to the fundamental goals of the SHM process. There have been numerous SHM studies on the various uncertainty and quantification parameters such as [813]. Many researchers have agreed that the Bayesian approach is “relatively” strong for system identification and it is capable of addressing the uncertainty to some extent. Bayesian techniques for SHM have been researched for ambient modal identification [8, 12, 1416], recursive filtering [17], acoustic emission-based monitoring [18, 19], distributed fiber Bragg sensors [14], optimal sensor placement [20], reliability assessment [21], seismic excitations [22], and artificial neural networks [23]. Not only in civil engineering structures, but also in aerospace and mechanical engineering structures, the emphasis in SHM research has been heavily on developing methods to better estimate the modal parameters through recorded data [2, 3]. However, there are some distinctions between the mechanical systems and civil structures for the application of some estimation methods. In addition, the SHM measurements are usually contaminated with errors hiding the important structural parameters [4, 5, 24, 25]. In an attempt to clean the contaminated data, filtering operation is conducted with Bayesian approximation techniques available in the literature. The classical Kalman filter (KF) was developed in the 1960s [26], and has been widely used in various engineering and science applications, including communications, control, machine learning, neuroscience, and many others. When the model describing the system is assumed to be linear and Gaussian, the KF provides an optimal solution  [27]. It is known that the KF is computationally efficient; however, it is limited by the non-universal linear and Gaussian modeling assumptions. To relax these assumptions, the extended Kalman filter (EKF) [28], the unscented Kalman filter (UKF) [29, 30], the central difference Kalman filter (CDKF) [31], the square-root unscented Kalman filter (SRUKF) [32], the square-root central difference Kalman filter (SRCDKF) [33], the iterated unscented Kalman filter (IUKF) [34] and the iterated square-root unscented Kalman filter (ISRUKF) [1] have been developed.

The objectives of the current paper are threefold: (1) to develop an iterated square-root central difference Kalman particle filter (ISRCDKF-PF) algorithm for nonlinear and non-Gaussian estimation. In the case of the conventional PF, the latest observation is not used for the computation of the weights of the particles since the sampling distribution is equal to the prior function. This selection of the sampling distribution simplifies the evaluation, but can cause filtering divergence. In those cases where the likelihood distribution is too narrow compared to the prior one, very few particles will have significant weights. Hence, a proposal sampling distribution that takes the latest observation into account is required. In this paper, the sampling distribution is evaluated using the ISRCDKF. Hence, the ISRCDKF-PF algorithm consists of a PF based on ISRCDKF to generate the optimal importance proposal distribution. The ISRCDKF-PF algorithm allows the particle filter to incorporate the latest observations into a prior updating scheme using the estimator of the posterior distribution that matches the true posterior more closely by using the ISRCDKF algorithm. (2) To investigate the effects of practical challenges (such as measurement noise and number of estimated states/parameters) on the performances of the techniques. To study the effect of measurement noise on the estimation performances, several measurement noise levels will be considered. Then, the estimation performances of the techniques will be evaluated for different noise levels. Also, to study the effect of the number of estimated states/parameters on the estimation performances of all the techniques, the estimation performance will be studied for different numbers of estimated states and parameters. (3) To apply the proposed technique to estimate the state variables as well as the model parameters through two examples, one using synthetic data and the other using simulated three DOF damped system data. The performances of the estimation techniques will be compared to each others by computing the execution times as well as the estimation root mean square error (RMSE) with respect to the noise-free data.

The rest of the paper is organized as follows. We first discuss related work and the motivation for the need to develop the scheme in Sect. 2. In Sect. 3, the state estimation problem is presented. Then, in Sect. 4, some state estimation methods and the developed iterated square-root central difference Kalman filter method are described. After that, in Sect. 5, the performance of the various state estimation techniques are compared for their application to estimate the state variables and model parameters in structural health monitoring system. Conclusions are presented in Sect. 6.

2 Related work

Several works based on Bayesian inference are used for state estimation in SHM systems. Mariani and Ghisi [35] used the UKF to perform the joint estimation of model parameters and state components of the softening single degree-of-freedom structural systems. While, the work in [36] showed that the UKF gives better state estimation and parameter identification than the EKF for higher degree of freedom systems and it is also robust to measurement noise levels. Eleni and Andrew [37] compared the UKF and PF methods and evaluated their efficiencies through a three-DOF damped system data.

The UKF and CDKF methods give good and similar estimation accuracies. However, in those techniques, the computation of the matrix square root of the state covariance at each time step when generating the sigma point is a costly operation. To get more numerical stability, the square root forms of the UKF and CDKF methods are numerically derived. In the SRUKF and SRCDKF methods, the square-root of the state covariance is directly propagated and updated in Cholesky factored form, using various linear algebra techniques such as the QR decomposition and the Cholesky factor updating. The SRCDKF method has equal or slightly better estimation accuracy when compared to the standard UKF, but with the benefit of having a reduced computational cost and an increased numerical stability (the covariance matrices are guaranteed to be positive definite). Otherwise, the iterative form of the UKF method gives better performance than the non-iterative one, because in the iterative schemes the observation equation is updated several times using the measurements at a single sample step, as a result the newest state estimation we get every time will be more precise. Hence, the IUKF method is more precise than the UKF method and the SRCDKF method is marginally better than the UKF method. In the same sense, ISRCDKF method is practically a computationally (potentially) more efficient of the IUKF method [34]. The ISRCDKF method employs an iterative procedure within a single measurement update step by resampling the sigma points till a termination criterion, based on the minimization of the maximum likelihood estimate, is satisfied. Furthermore, the ISRCDKF method propagates and updates the square root of the state covariance iteratively and directly in Cholesky factored form. In addition to providing reduction in the computational complexity, ISRCDKF has as increased numerical stability and better (or at least equal) performance when compared to the other algorithms. Unfortunately, for most nonlinear systems and non-Gaussian noise observations, closed-form analytic expressions of the posterior distribution of the state vector are untractable [38]. To overcome this drawback, a non-parametric Monte Carlo sampling-based method called particle filter (PF) [39, 40] has recently gained popularity. The PF approximates the posterior probability distribution by a set of weighted samples, called particles [39, 41]. PF is widely used in nonlinear and non-Gaussian systems for several filtering applications such as estimation, prediction, modeling and monitoring [42]. In case of standard PF, the latest observation is not considered for the evaluation of the weights of the particles as the importance function is taken to be equal to the prior density function [43, 44]. This choice of importance sampling function simplifies the computation but can cause filtering divergence. In cases where the likelihood function is too narrow compared to the prior function, very few particles will have significant weights [30, 45]. Hence, a better proposal distribution that takes the latest observation into account is desired in order to gain better performance [45, 46]. In the current work, we extend our previous work [1] to deal with non-parametric Monte Carlo sampling-based method and propose to use an ISRCDKF-PF technique, which consists of a PF based on ISRCDKF to generate the optimal importance proposal distribution. The ISRCDKF-PF method allows the PF to incorporate the latest observations into a prior updating scheme using ISRCDKF-based approximated posterior distribution.

3 State estimation problem

Next, we present the formulation of the state estimation problem.

3.1 Problem formulation

Consider a generic discrete-time nonlinear dynamic system described by the following dynamic state-space model (DSSM):

$$\begin{aligned} \left\{ \begin{array}{l} x_{k+1}=f(x_{k}, u_{k}, \theta , v_{k})\\ y_{k}=h(x_{k}, u_{k}, \theta , w_{k}), \end{array} \right. \end{aligned}$$
(1)

where k is the discrete time index, \(x_{k} \in {\mathbb {R}}^{n}\) is the state vector, \(y_{k} \in {\mathbb {R}}^m\) is the measurement vector, \(\theta \in {\mathbb {R}}^q\) is the system parameters vector, \(u_{k} \in {\mathbb {R}}^p\) is the input vector, f and h are, respectively, the state and the observation functions, and \(v_{k} \in {\mathbb {R}}^{n}\) and \(w_{k} \in {\mathbb {R}}^m\) are mutually independent i.i.d noise processes.

In this paper, the input vector \(u_{k}\) is assumed to be known. However, the vector of system parameters \(\theta\) is taken to be unknown and it will be jointly estimated with the state vector. For the sake of computational expediency, we assume that the unknown system parameter vector evolves artificially in time according to a random walk as follows:

$$\begin{aligned} \theta _{k}=\theta _{k-1}+ \eta _{k-1}, \end{aligned}$$
(2)

where \(r_{k-1}\) is a sequence of i.i.d zero-mean Gaussian random variables. In such estimation problems, the state vector and the system parameter vector are usually concatenated into a single higher-dimensional joint state vector, \(z_{k} \in {\mathbb {R}}^{n+q}\), defined as:

$$\begin{aligned} z_{k}=\left[ \begin{array}{ll} x_{k}^{T}&\theta _{k}^{T} \end{array}\right] ^{T} \end{aligned}$$
(3)

Using the augmented state vector, the process equation of the DSSM (1) can be expanded to:

$$\begin{aligned} z_{k+1}= \left[ \begin{array}{l} x_{k+1} \\ \theta _{k+1} \end{array} \right] = \left[ \begin{array}{l} f(x_{k},u_{k},v_{k}, \theta _{k}) \\ \theta _{k}+\eta _{k} \end{array} \right] \end{aligned}$$
(4)

Also, defining the augmented noise vector as \(\tilde{w}_{k}= \left[ \begin{array}{ll} v_{k}^{T}&\eta _{k}^{T} \end{array} \right] ^{T}\), the DSSM (1) can be written as:

$$\begin{aligned} \left\{ \begin{array}{l} z_{k+1}={f}(z_{k}, u_{k}, \tilde{w}_{k}) \\ y_{k}={l}(z_{k}, u_{k}, v_{k}), \end{array} \right. \end{aligned}$$
(5)

where f and l are differentiable nonlinear functions.

4 Description of state estimation methods

4.1 Square-root central difference Kalman filter (SRCDKF) method

For nonlinear system, the EKF recursively calculates the mean and the covariance \({\mathbf {P}}_k\) of the random variable, while the UKF calculates the matrix square-root \({\mathbf {S}}_k {\mathbf {S}}^T_k={\mathbf {P}}_k\), at each time step. However, the SRCDKF directly propagate the square-root covariance matrix \({\mathbf {S}}_k\) avoiding the computational complexity to refactorize at each time step [32]. The SRCDKF firstly initializes the mean of the state vector and the square root of it covariance.

$$\begin{aligned} \hat{z}_0=E[z_0], \end{aligned}$$
(6)

and

$$\begin{aligned} {\mathbf {S}}_0={\rm chol}\, \left\{ E\left[ (z_0-\hat{z}_0)(z_0-\hat{z}_0)'\right] \right\} . \end{aligned}$$
(7)

The Cholesky factorization produces an upper triangular matrix from the decomposition of a symmetric, positive-definite matrix into the product of a lower-triangular matrix and its transpose. This new matrix is utilized directly to generate the sigma points:

$$\begin{aligned} \Psi _{k-1}=\left[ \hat{z}_{k-1} \ \ \hat{z}_{k-1}+h {\mathbf {S}}_{k-1} \ \ \hat{z}_{k-1}-h {\mathbf {S}}_{k-1} \right],\end{aligned}$$
(8)

where \(\hat{z}_{k-1}=\left[ \hat{z}_{k-1} \ \ \bar{v} \right]\), \({\mathbf {S}} _{k-1}=\mathrm{diag}\left[ {\mathbf {S}}_{k-1} \ \ {\mathbf {S}}_v \right]\) and h is a scaling parameter equal to the kurtosis of the prior random variable. The optimal value of h for Gaussian random variables is \(h=\sqrt{3}\).

Next, the sigma points are propagated through the nonlinear process system to predict the current attitude based on each sigma point.

$$\begin{aligned} \Psi _{k|k-1}=f\left[ \Psi _{k-1}\right] . \end{aligned}$$
(9)

Then the state mean and square-root covariance are estimated from the transformed sigma points using,

$$\begin{aligned} \hat{z}^-_{k}=\sum ^{2L}_{i=0}W_i^{(m)}\Psi _{i,k|k-1}, \end{aligned}$$
(10)

and

$$\begin{aligned} {\mathbf {S}}^-_{k}=qr \left\{ \left[ \sqrt{W_1^{(c1)}}(\Psi _{1:L,k|k-1}-\Psi _{L+1:2L,k|k-1}) \sqrt{W_1^{(c2)}}(\Psi _{1:L,k|k-1}+\Psi _{L+1:2L,k|k-1}-2 \Psi _{0,k|k-1}) \right] \right\}, \end{aligned}$$
(11)

where \(W_i^{(c1)}=\frac{1}{4h^2}\), \(W_i^{(c2)}=\frac{h^2-1}{4h^4}\), \(W_0^{(m)}=\frac{h^2-L}{h^2}\), \(W_i^{(m)}=\frac{1}{2h^2}\).

The next step, the sigma-point for measurement update is generated as,

$$\begin{aligned} \Psi _{k|k-1}=\left[ \hat{z}_{k|k-1} \ \ \hat{z}_{k|k-1}+h {\mathbf {S}}_{k|k-1} \ \ \hat{z}_{k|k-1}-h {\mathbf {S}}_{k|k-1} \right], \end{aligned}$$
(12)

where \(\hat{z}_{k|k-1}=\left[ \hat{z}_{k|k-1} \ \ \bar{w}\right].\), \({\mathbf {S}}_{k|k-1}=\mathrm{diag}\left[ {\mathbf {S}}_{k|k-1} \ \ {\mathbf {S}}_{w}\right].\)

The measurements sigma points are propagated through the measurement model:

$$\begin{aligned} Y_{k|k-1}=h \left[ \Psi _{k|k-1} \right] . \end{aligned}$$
(13)

To predict the measurements, the expected measurement \(\hat{y}^-_{k}\) and square-root covariance of \(\tilde{y}_k=y_k-\hat{y}^-_{k}\) (called the innovation) are calculated as:

$$\begin{aligned} \hat{y}^-_{k}=\sum ^{2L}_{i=0}W_i^{(m)}Y_{i,j}. \end{aligned}$$
(14)
$$\begin{aligned} \tilde{y}_k = y_{k} - \hat{y}^-_{k}, \end{aligned}$$
(15)
$$\begin{aligned} {\mathbf {S}}_{\tilde{y}_k}=qr \left\{ \left[ \sqrt{W_1^{(c1)}}(Y_{1:L,j}-Y_{L+1:2L,j})\sqrt{W_1^{(c2)}}(Y_{1:L,j}+Y_{L+1:2L,j}-2Y_{0,j})\right] \right\} . \end{aligned}$$
(16)

In order to find out how much to adjust the predicted state mean and covariance based on the actual measurement, the Kalman gain matrix \({\mathbf {K}} _k\) is calculated as follows:

$$\begin{aligned} {\mathbf {P}}_{{z}_k{y}_k}=W_i^{(c1)}{S}^-_k [ Y_{1:L,j}- Y_{L+1:2L,j}]^T. \end{aligned}$$
(17)
$$\begin{aligned} {\mathbf {K}}_{k,j}={\mathbf {P}}_{z_k y_k}/{\mathbf {S}}_{k}^T/\mathbf{S _{\overline{y}_k}}. \end{aligned}$$
(18)

Then, updated state mean and covariance are expressed using the actual measurement and the Kalman gain matrix as:

$$\begin{aligned} \hat{z}_{k}=\hat{z}^-_{k}+{\mathbf {K}}_k(y_k-\hat{y}^-_{k}). \end{aligned}$$
(19)
$$\begin{aligned} {\mathbf {U}} ={\mathbf {K}}_k {\mathbf {S}}_{\bar{y}_k}. \end{aligned}$$
(20)
$$\begin{aligned} {\mathbf {S}}_k={\rm cholupdate } \left\{ {\mathbf {S}}^-_{k}, {\mathbf {U}} ,-1\right\}, \end{aligned}$$
(21)

where \({\mathbf {S}}_v=\sqrt{\mathbf{R }^v}\) is the square root of the process noise covariance matrix, \({\mathbf {S}}_{w}=\sqrt{\mathbf{R }^{w}}\) is the square root of the measurement noise covariance matrix, chol - is Cholesky method of matrix factorization, qr is QR matrix decomposition and cholupdate is a Cholesky factor updating.

The benefit of the ISRCDKF method lies in its ability to provide accuracy-related advantages over other estimation methods since it re-linearizes the measurement equation by iterating an approximate maximum a posteriori estimate around the updated state, instead of relying on the predicted state. Next, ISRCDKF is presented.

4.2 Iterated square-root central difference Kalman filter (ISRCDKF) method

With the success of IUKF method development [34] and the superiority of SRCDKF, an improved performance would be expected if the iterates are implemented in SRCDKF. With the potential problems experienced with the IUKF method, precaution should be taken for effective performance of the iterated filter [47]. The development of the ISRCDKF method is due to the need to overcome this issue, utilizing a different iteration strategy.

In the ISRCDKF method, we apply the iterations on the process of measurement update where the updated state estimation is calculated using the predicted state and the observation. Instead of relying on the predicted state, the observation equation is re-linearized over times by iterating an approximate maximum a posteriori estimate, so the state estimate will be more accurate.

In the iterated measurement update step of the SRCDKF approach, the new sigma points are generated at each iteration from the latest estimation of the posterior state distribution.

The square-root central difference approach is applied again with the state estimation \(\hat{z}_k\) which will be expressed as \(z_j\) in the iteration and indicates the state estimation in the jth iteration. Then, the state estimate is performed by the measurement value. This is how the ISRCDKF works.

After the estimated mean \(\hat{z_k}^-\) and the covariance matrix \(S_k^-\) are obtained, \(z_{j}\) is obtained by the filtering algorithm. The sampling points are re-generated based on \(z_{j}\) and \(S_k^-\) by the following equation:

$$\begin{aligned} \Psi _{j}=\left[ z_{j} \ \ z_{j}+h {\mathbf {S}}^-_{k} \ \ z_{j}-h {\mathbf {S}}^-_{k} \right] \end{aligned}$$
(22)

The new generated sample points pass through the measurement equation:

$$\begin{aligned} Y_{j}=h \left[ \Psi _{j} \right] , \end{aligned}$$
(23)
$$\begin{aligned} \hat{y}^-_{j}=\sum ^{2L}_{i=0}W_i^{(m)}Y_{i,j}, \end{aligned}$$
(24)
$$\begin{aligned} \tilde{y}_k = y_{k} - \hat{y}^-_{k}, \end{aligned}$$
(25)
$$\begin{aligned} {\mathbf {S}}_{\tilde{y}_k}=QR \left\{ \left[ \sqrt{W_1^{(c1)}}(Y_{1:L,j}-Y_{L+1:2L,j}) \sqrt{W_1^{(c2)}}(Y_{1:L,j}+Y_{L+1:2L,j}-2Y_{0,j})\right] \right\} , \end{aligned}$$
(26)
$$\begin{aligned} {\mathbf {P}}_{{z}_k{y}_k}=W_i^{(c1)}{S}^-_k [ Y_{1:L,j}- Y_{L+1:2L,j}]^T. \end{aligned}$$
(27)

Then, the Kalman gain \(K_k\) is be re-calculated:

$$\begin{aligned} {\mathbf {K}}_{k,j}={\mathbf {P}} _{z_k y_k}/{\mathbf {S}} _{k}^T/\mathbf{S _{\overline{y}_k}}. \end{aligned}$$
(28)

Next, the state estimate update is improved using \(y_j\) instead of \(\hat{y}_k\):

$$\begin{aligned} \hat{z}_{k}=\hat{z}^-_{k}+{\mathbf {K}} _k(y_k-\hat{y}^-_{j}). \end{aligned}$$
(29)

Finally, the covariance matrix is re-calculated:

$$\begin{aligned}{\mathbf {U}} ={\mathbf {K}} _k {\mathbf {S}} _{\bar{y}_k}, \end{aligned}$$
(30)
$$\begin{aligned} {\mathbf {S}}_k={\rm cholupdate }\, \left\{ {\mathbf {S}}^-_{k},{\mathbf {U}} ,-1\right\} . \end{aligned}$$
(31)

The algorithm of the ISRCDKF method can be summarized in Algorithm 1.

figure a

4.3 Conventional particle filter (PF) method

A particle filter is an implementation of a recursive Bayesian estimator [30, 48]. Bayesian estimation relies on computing the posterior \(p(z_{k}|y_{0:k})\), which is the density function of the unobserved state vector, \(z_{k}\), given the sequence of the observed data \(y_{0:k} \equiv \{y_0,\; y_2,\ldots , y_k\}\). In a Bayesian context, the task of state estimation can be formulated as recursively calculating the predictive distribution \(p(z_{k}|y_{0:k-1})\) and the filtering distribution \(p(z_{k}|y_{0:k})\) as follows [30],

$$\begin{aligned} p(z_{k}|y_{0:k-1})= & {} \int _{{\mathbb {R}}^{n}} p(z_{k}|z_{k-1})p(z_{k-1}|y_{0:k-1})\mathrm{d}z_{k-1},\\ \quad p(z_{k}|y_{0:k}) =\, & {} \frac{p(y_k|z_{k})p(z_{k}|y_{0:k-1})}{p(y_k|y_{0:k-1})}, \nonumber \end{aligned}$$
(32)

where the normalizing constant \(\quad p(y_k|y_{0:k-1}) = \int _{{\mathbb {R}}^z} p(y_k|z_{k})p(z_{k}|y_{0:k-1})\mathrm{d}z_{k}\).

The nonlinear nature of the system model leads to intractable integrals when evaluating the marginal state distribution, \(p(z_k | z_{k-1})\). Therefore, Monte Carlo approximation is utilized, where the joint posterior distribution, \(p(z_{0:k} | y_{0:k})\), is approximated by the point-mass distribution of a set of weighted samples (particles) \(\{z^{(i)}_{0:k},\ell ^{(i)}_k\}_{i=0}^{N}\), i.e., [30, 49]:

$$\begin{aligned} \hat{p}_N(z_{0:k} | y_{0:k}) = \sum _{i=0}^{N}\ell ^{(i)}_k \delta _{z^{(i)}_{0:k}} (\mathrm{d}\,z_{0:k})/\sum _{i=0}^{N}\ell ^{(i)}_k, \end{aligned}$$
(33)

where \(\delta _{z^{(i)}_{0:k}} (\mathrm{d}\,z_{0:k})\) denotes the Dirac function, \(\ell ^{(i)}_k\) are the corresponding importance weights and N is the total number of particles. Based on the same set of particles, the marginal posterior probability of interest, \(p(z_k | y_{0:k})\), can also be approximated as follows [30, 42]:

$$\begin{aligned} \hat{p_N}(z_k | y_{0:k}) = \sum _{i=0}^{N}\ell ^{(i)}_k \delta _{z^{(i)}_{k}} (\mathrm{d}\,z_{k})/\sum _{i=0}^{N}\ell ^{(i)}_k. \end{aligned}$$
(34)

In this Bayesian importance sampling (IS) approach, the particles \(\{z^{(i)}_{0:k}\}_{i=0}^{N}\) are sampled from the following distribution (called also Importance density) [42],

$$\begin{aligned} p(z_{0:k} | y_{0:k})=p(z_k | z_{k-1}) = \int {\mathcal {N}} (z_k | \mu _k, \lambda _k) p(\mu _k, \lambda _k | z_{k-1}) \mathrm{d} \mu _k \mathrm{d} \lambda _k , \end{aligned}$$
(35)

where \(\mu _{k}\) defines the expectation of the state \(z_{k}\) and \(\lambda _{k}\) defines the covariance matrix of the state \(z_{k}\).

Resampling is performed whenever the effective sample size \(N_\mathrm{eff}\) drops below a certain threshold \(N_\mathrm{threshold}\), where a smaller \(N_\mathrm{eff}\) means a larger variance for the weights, hence more degeneracy.

Then, the estimate of the augmented state \(\widehat{z}_{k}\) can be approximated by a Monte Carlo scheme as follows [49]:

$$\begin{aligned} \widehat{z}_{k} = \sum _{i=0}^{N} \ell _{k}^{(i)}z_{k}^{(i)}, \end{aligned}$$
(36)

where \(\ell ^{(i)}_k\) is given by [49]:

$$\begin{aligned} \ell ^{(i)}_k \propto \frac{p(y_{0:k} | z^{(i)}_{0:k}) p(z^{(i)}_{0:k})}{p(z^{(i)}_{0:k} | y_{0:k})}. \end{aligned}$$
(37)

A common problem with the sequential importance sampling-based particle filter is the degeneracy phenomenon. This degeneracy implies that a large computational effort is devoted to updating particles whose contribution to the approximation of \(p(z_{k} | y_{0:k})\) is almost zero. A suitable measure of degeneracy of the algorithm is the estimate effective sample size \(\hat{N}_{eff}\), which is introduced in [48] and [50], and is defined as,

$$\begin{aligned} \hat{N}_\mathrm{eff}= \frac{1}{\sum _{i=0}^N (\ell _k^{(i)})^2}, \end{aligned}$$
(38)

where \(\ell _k^{(i)}\) are the normalized weights obtained using (37).

The conventional PF algorithm for state/parameter estimation is summarized in Algorithm 2.

figure b

Particle filtering suffers from one major drawback. Its efficient implementation requires the ability to sample from \(p(z_{k}|z_{k-1})\), which does not take into account the current observed data, \(y_k\), and thus many particles can be wasted in low likelihood (sparse) areas [30]. This issue is addressed by the iterated square-root central difference Kalman particle filter, which is described in next Section.

4.4 Iterated square-root central difference Kalman particle filter (ISRCDKF-PF) Method

The choice of optimal proposal function is one of the most critical design issues in importance sampling schemes. In [49], the optimal proposal distribution \(\hat{p}(z_{k}|z_{0:k-1}, y_{0:k})\) is obtained by minimizing the variance of the importance weights given the states \(z_{0:k-1}\) and the observations data \(y_{0:k}\). This selection has also been studied by other researchers. However, this optimal choice suffers from one major drawback. The particles are sampled from the prior density \(p(z_{k}|z_{0:k-1})\) and the integral over the new state need to be computed. In the general case, closed-form analytic expression of the posterior distribution of the state is untractable [38]. Therefore, the distribution \(p(z_{k}|z_{0:k-1})\) is the most popular choice of proposal distribution. One of the advantages of PF is its simplicity in sampling from the prior functions \(p(z_{k}|z_{0:k-1})\) and the evaluation of weights \({\ell }^{(i)}_k\) (as presented in the previous section) [38]. However, the latest observation is not considered for the computation of the weights of the particles as the importance density is taken to be equal to the prior density [43]. The transition prior \(p(z_{k}|z_{0:k-1})\) does not take into account the current observation data \(y_{k}\), and many particles can be wasted in low likelihood areas. This choice of importance sampling function simplifies the computational complexity but can cause filtering divergence [30, 45]. In cases where the likelihood density is too narrow as compared to the prior function, very few particles will have considerable weights. Therefore, the distribution \(p(z_{k}|z_{0:k-1})\) is the most popular choice of proposal distribution. Although the ISRCDKF-PF moves the prior toward the likelihood, which due to creating a better proposal distribution, this is done by generating an optimal importance proposal distribution by using ISRCDKF method (as presented in Sect. 4.2). Hence, the ISRCDKF-PF algorithm consists of a PF based on ISRCDKF to generate the optimal importance proposal distribution. The proposal distribution integrates the latest observation into system state transition density.

The developed ISRCDKF-PF algorithm can be summarized in Algorithm 3.

figure c

5 Simulation results

5.1 State and parameter estimations for a scalar nonlinear system

Next, the estimation performance is evaluated and the developed ISRCDKF-PF method is compared to UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF and IUKF-PF methods through a scalar nonlinear system (40). The purpose of this section is to estimate the state variable \(x_{k}\) and the model parameter \(\phi _{1}\). The equations representing the system can be written as follows:

$$\begin{aligned} x_{k}=\phi _{1} x_{k-1} + 1 +\mathrm{sin}(\omega \pi (k-1))+v_{k}, \end{aligned}$$
(39)

where \(v_{k}\) is a \(N(v_{k}; 0, 10^{-4})\) Gaussian distribution modeling the process noise, and \(\omega\) = 0.04 and \(\phi _1\) = 0.5 are scalar parameters. The non-stationary observation model is given by:

$$\begin{aligned} y_{k}= {\left\{ \begin{array}{ll} \phi _{2} x_{k}^{2}+n_{k}& \text{ for } \quad k<=30 \\ \phi _{3} x_{k}-2 +n_{k} & \text{ for } \quad k>30, \end{array}\right. } \end{aligned}$$
(40)

where \(\phi _{2}\) = 0.2 and \(\phi _{3}\) = 0.5. The observation noise, \(n_{k}\), is drawn from a Gaussian distribution \({\mathcal {N}} (n_{k}; 0, 10^{-4})\).

For all simulations, the number of sigma points is fixed to 5 for all the techniques (\(L=2\)) and the particle filters use 500 particles. The initial values of the augmented state vector are fixed to \(z(0)=[x_0 \ \ \theta _0]=[1 \ \ 0.1]\).

5.1.1 Comparative study 1: estimation of state variables from noisy measurements

Here, we compare the estimation accuracy of UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF methods when they are utilized to estimate the state variable of the system model. Hence, it is considered that the state vector to be estimated is \(z_k=x_k\), and the model parameter, \(\phi _{1}\), is assumed to be known.

The simulation results for the estimation of state variable \(x_k\) using UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF methods are shown in Figs. 1 and 2, respectively. Also, the performance comparison of the state estimation techniques in terms of RMSE is presented in Table 1. It is easily observed from Figs. 1 and 2 as well as Table 1 that UKF is outperformed by the alternative techniques. The results also show that the ISRCDKF-PF method achieves a better accuracy than the other methods.

Fig. 1
figure 1

Estimation of state variable using various state estimation techniques (UKF, SRCDKF, IUKF, ISRCDKF)

Fig. 2
figure 2

Estimation of state variable using various state estimation techniques (PF, UKF-PF, IUKF-PF, SRCDKF-PF, ISRCDKF-PF)

Table 1 Comparison of state estimation techniques

5.1.2 Comparative study 2: simultaneous estimation of state variables and model parameters

The estimation of the state variables and parameters were performed using the state estimation techniques, UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF. The estimation results for the model parameters using these estimation techniques are shown in Figs. 3, and 4, respectively.

It can be seen from the results presented in Figs. 3 and 4 that the IUKF-PF, SRCDKF-PF and ISRCDKF-PF methods outperform the UKF-PF method, and that the ISRCDKF-PF shows relative improvement over all other techniques. These results confirm the results obtained in the first comparative study, where only the state variable is estimated. The advantages of the ISRCDKF-PF over the other techniques can also be seen through their abilities to estimate the model parameters. For example, UKF, IUKF, SRCDKF and ISRCDKF could took longer to estimate a model parameters (see Figs. 3, 4).

Fig. 3
figure 3

Estimation of the model parameter (\(\phi _1\)) using UKF, SRCDKF, IUKF and ISRCDKF

Fig. 4
figure 4

Estimation of the model parameter (\(\phi _1\)) using UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF

5.1.3 Effect of number of state and parameter to estimate on the estimation RMSE

To study the effect of the number of states and parameters to be estimated on the estimation performances of UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF, the estimation performance is analyzed for different numbers of estimated states and parameters. The estimation of the state variable and parameter is performed using the techniques, and the simulation results for the state variable and the model parameters are shown in Table 2. Table 2 compares the estimation root mean square errors for the state variable \(x_k\) (with respect to the noise-free data) and the mean of the estimated parameter \(\phi _1\) (true value of \(\phi _1\) ) at steady state (i.e., after convergence of the parameter) using UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF, respectively.

The results also show that the number of states and parameters to estimate affects the estimation accuracy of the state variable. In other words, the estimation RMSE of \(x_k\) increases from the first comparative study where only the state variable \(x_k\) is estimated to the case where the state variable \(x_k\) and one parameter \(\phi _1\) are estimated.

For example, the RMSEs obtained using ISRCDKF-PF for \(x_k\) in the first comparative study and the second comparative study are 0.0288 and 0.0610, respectively, which increase as the number of estimated parameters increases (see Table 2). This observation is valid for the other state estimation techniques.

Table 2 Root mean square errors of estimated state variables and mean of estimated parameter

5.2 State and parameter estimations for three degree of freedom spring–mass–dashpot system

Here, we consider the example of sensor heterogeneity arising from the fact that both acceleration and displacement are measured at various locations of the structural system. The availability of non-collocated data might often arise in the identification of systems where the displacement data may be provided through global positioning systems. The performance of developed ISRCDKF-PF is evaluated through the example of a three degree-of-freedom system, involving a Bouc–Wen hysteretic component, where the availability of displacement and acceleration measurements for different DOFs is assumed. Three degree of freedom spring–mass–dashpot system is illustrated in Fig. 5, which is usually used in SHM literature for the performance evaluation of state estimation techniques. In the considered system, the first degree of freedom is associated with a nonlinear hysteretic component [25].

Fig. 5
figure 5

Three degree of freedom spring–mass–dashpot system. Note that the first degree of freedom is associated with a nonlinear hysteretic component [25]

The system can be modeled by the equation of motion as follows:

$$\begin{aligned} M \ddot{x}(t) + C \dot{x}(t) + R x(t)=F(t), \end{aligned}$$
(41)

where M and C are the matrices of mass and damping, R is the restoring force, x(t) is the vector of displacements, \(\dot{x}(t)\) is the vector of velocities, \(\ddot{x}(t)\) is the vector of accelerometer measurements and F(t) is the vector of excitation force.

The restoring force R can be expressed as:

$$\begin{aligned} R=ak_ix(t)+(1-a)k_ir(t), \end{aligned}$$
(42)

where \(a=\frac{k_f}{k_i}\) is the ratio of post-yield \(k_f\) to pre-yield \(k_i\) stiffness and \(r_1(k)\) is the Bouc–Wen hysteretic component that can be written as follows:

$$\begin{aligned} \dot{r}_1(k)=\dot{x}_1-\beta |\dot{x}_1||r_1|^{n-1}r-\gamma (\dot{x}_1)|r_1|^n, \end{aligned}$$
(43)

where \(\beta\), \(\gamma\) and n are the Bouc–Wen hysteretic parameters.

The system can be represented by the following state equation:

$$\begin{aligned} \begin{bmatrix} {x}_{1} \\ {x}_{2} \\ {x}_{3} \end{bmatrix} -> \begin{bmatrix} m_1&0&0\\ 0&m_2&0\\ 0&0&m_3\\ \end{bmatrix} \begin{bmatrix} \ddot{x}_{1} \\ \ddot{x}_{2} \\ \ddot{x}_{3} \end{bmatrix}+ \begin{bmatrix} c_1+c_2&-c_2&0 \\ -c_2&c_2+c_3&-c_3 \\ 0&-c_3&c_3 \\ \end{bmatrix} \begin{bmatrix} \dot{x}_{1} \\ \dot{x}_{2} \\ \dot{x}_{3} \end{bmatrix} + \begin{bmatrix} k_1&k_2&-k_3&0 \\ 0&-k_2&k_2+k_3&-k_3 \\ 0&0&-k_3&k_3 \end{bmatrix} \begin{bmatrix} r_{1} \\ {x}_{1} \\ {x}_{2} \\ {x}_{3} \end{bmatrix} = \begin{bmatrix} F_{1}(k) \\ F_{2}(k) \\ F_{3}(k)\end{bmatrix} \end{aligned}$$
(44)

where \(r_1(k)\) is the Bouc–Wen hysteretic component with:

$$\begin{aligned} \dot{r}_1(k)=\dot{x}_1-\beta |\dot{x}_1||r_1|^{n-1}r-\gamma (\dot{x}_1)|r_1|^n \end{aligned}$$
(45)

\(\beta\), \(\gamma\) and n are the Bouc–Wen hysteretic parameters.

The state and observation formulations are detailed in our previous work [1].

5.2.1 Generation of dynamic data

For dynamic data generation from the SHM system (46) [25] is used to simulate the responses of the state as functions of time by solving the differential Eq. (46) using fourth-order Runge–Kutta Integration.

$$\begin{aligned} \dot{y}=\left[ \begin{array}{lll} \dot{y}_{1} \\ \dot{y}_{2} \\ \dot{y}_{3} \\ \dot{y}_{4} \\ \dot{y}_{5} \\ \dot{y}_{6} \\ \dot{y}_{7} \\ \end{array} \right] = \left[ \begin{array}{lll} \dot{x}_{1} \\ \dot{x}_{2} \\ \dot{x}_{3} \\ \dot{r}_{1} \\ \ddot{x}_{1} \\ \ddot{x}_{2} \\ \ddot{x}_{3} \\ \end{array} \right] = \left[ \begin{array}{lll} {z}_{5} \\ {z}_{6} \\ {z}_{7} \\ z_{5}-2|z_{5}||z_{4}|^{2-1}z_{4}-1z_{5}|z_{4}|^{2}\\ -9z_{4} -9z_{1} + 9z_{2} -0.5z_{5}+0.25z_{6}+\ddot{v}_g \\ -9z_{1} -18z_{2} + 9z_{3} -0.25z_{5}+0.5z_{6}+0.25z_{z}-\ddot{v}_g \\ 9z_{2} -9z_{3} + 0.25z_{6} -0.25z_{7}-\ddot{v}_g \end{array} \right] , \end{aligned}$$
(46)

where the state variables \(x_1\), \(x_2\), \(x_3\) are displacements and \(r_1\) is the hysteretic Bouc–Wen parameter.

It must be noted that these simulated states are assumed to be noise free. They are contaminated with zero mean Gaussian errors. The SHM parameters as well as other physical properties are shown in Table 3. Figure 6 shows the changes in the state variable (displacement \(x_1\)).

Fig. 6
figure 6

Simulated data used in estimation: state variable (\(x_1\))

Table 3 SHM parameters and physical properties

For all simulations, the following parameters are used. The sampling frequency of the Northridge earthquake acceleration data that was used as ground excitation \(\ddot{v}_g\), is 100 Hz (\(T=0.01\) s). The Northridge earthquake signal was filtered with a low-frequency cutoff of 0.13 Hz and a high-frequency cutoff of 30 Hz. A duration of 20 s of the earthquake record was adopted [25]. The number of sigma points is fixed to 32 for all the techniques (\(L=16\)). The process noise of 1 % RMS noise-to-signal ratio was added. The observation noise level was of 4 % RMS noise-to-signal ratio. The initial values of the augmented state vector \(z(0)=[x_0 \ \ \theta _0]\) are given by:

$$\begin{aligned} x_0 = \begin{bmatrix} 0&0&0&0&0&0&0 \end{bmatrix} \text {and} \ \ \theta _0= \begin{bmatrix} 6&6&6&0.5&0.5&0.5&3&2&1 \end{bmatrix}. \end{aligned}$$

5.2.2 Comparative study 1: estimation of state variables from noisy measurements

The purpose of this study is to compare the estimation accuracy of UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF methods when they are utilized to estimate the seven state variables of the three degree of freedom spring–mass–dashpot system model. Hence, it is considered that the state vector to be estimated, \(z_k=x_k=\left[ x_{1} \; x_2\; x_{3} \; r_1\; v_1\; v_{2} \; v_3 \right] ^T\), and the model parameters, \(k_{1}\), \(k_{2}\), \(k_{3}\), \(c_{1}\), \(c_{2}\), \(c_{3}\), \(\beta\), \(\gamma\), and n are assumed to be known.

The simulation results for state estimations of seven state variables \(x_{1}\), \(x_2\), \(x_{3}\), \(r_1\), \(v_1\), \(v_{2}\) and \(v_3\) using UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF methods are shown in Figs. 7a–d, 8a–d, 9a–d and 10a–d, respectively. Also, the performance comparison of the state estimation techniques in terms of RMSE are presented in Table 4 (mean RMSE (MRMSE) for UKF=0.0328, SRCDKF= 0.0326, IUKF= 0.0323, ISRCDKF= 0.0320,PF = 0.0326, MRMSE (UKF-PF) = 0.0313, MRMSE (SRCDKF-PF) = 0.0312, MRMSE (IUKF-PF) = 0.0307, and MRMSE (ISRCDKF-PF) = 0.0305. It is easily observed from Figs. 78, 9a–d and  10a–d as well as Table 4 that PF is outperformed by the alternative techniques (see Table 4). The results also show that the ISRCDKF-PF method achieves a better accuracy than the IUKF-PF method. Both ISRCDKF-PF and IUKF-PF methods can provide improved accuracy over the UKF-PF and SRCDKF-PF approaches.

Fig. 7
figure 7

Estimation of state variables using various state estimation techniques (UKF, IUKF, SRCDKF and ISRCDKF)

Fig. 8
figure 8

Estimation of state variables using various state estimation techniques (PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF)

Fig. 9
figure 9

Estimation of state variables using various state estimation techniques (UKF, IUKF, SRCDKF and ISRCDKF)

Fig. 10
figure 10

Estimation of state variables using various state estimation techniques (PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF)

Table 4 Comparison of state estimation techniques

5.2.3 Comparative study 2: simultaneous estimation of state variables and model parameters

The estimation of the state variables and parameters were performed using the state estimation techniques, UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF. The estimation results for the model parameters using the estimation techniques (UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF) are shown in Figs. 1112131415 and 16, respectively.

It can be seen from the results presented in Figs. 1112131415 and 16 that the UKF-PF, IUKF-PF and SRCDKF-PF methods outperform the PF, ISRCDKF, IUKF, SRCDKF, and UKF methods, and that the ISRCDKF-PF shows relative improvement over all other techniques. These results confirm the results obtained in the first comparative study, where only the state variables are estimated. The advantages of the ISRCDKF-PF over the IUKF-PF (and the IUKF-PF over the PF, the UKF-PF and the SRCDKF-PF) can also be seen through their abilities to estimate the model parameters. For example, UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, IUKF-PF and SRCDKF-PF could took longer to estimate a model parameters (see Figs. 111213141516). The ISRCDKF-PF, however, could estimate all the model parameters in all four cases.

Fig. 11
figure 11

Estimation of the model parameters (\(k_1\), \(k_2\) and \(k_3\)) using UKF, SRCDKF, IUKF and ISRCDKF

Fig. 12
figure 12

Estimation of the model parameters (\(k_1\), \(k_2\) and \(k_3\)) using PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF

Fig. 13
figure 13

Estimation of the model parameters (\(c_1\), \(c_2\) and \(c_3\)) using UKF, SRCDKF, IUKF and ISRCDKF

Fig. 14
figure 14

Estimation of the model parameters (\(c_1\), \(c_2\) and \(c_3\)) using PF, UKF-PF, SRCDKF-PF, IUKF-PF and ISRCDKF-PF

Fig. 15
figure 15

Estimation of the model parameters (\(\beta\), \(\gamma\) and n) using UKF, SRCDKF, IUKF and ISRCDKF

Fig. 16
figure 16

Estimation of the model parameters (\(\beta\), \(\gamma\) and n) using PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF

5.2.4 Root mean square error analysis

Some practical challenges, however, can affect the accuracy of estimated states and/or parameters. Such challenges include the large number of states and parameters to be estimated, and the presence of measurement noise in the data. The effect of the these challenges on the performances of the UKF, IUKF, SRCDKF, ISRCDKF, PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF for state and parameter estimation are investigated.

  1. (a)

    Effect of number of state and parameter to estimate on the estimation RMSE

To study the effect of the number of states and parameters to be estimated on the estimation performances of PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF, the estimation is analyzed for different numbers of estimated states and parameters. Here, we will consider three cases, which are summarized below. In all cases, it is assumed that the seven states \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) are measured.

  1. 1.

    Case 1: the seven states \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) along with the first parameter \(k_1\) will be estimated.

  2. 2.

    Case 2: the seven states \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) along with the two parameters \(k_1\) and \(k_2\) will be estimated.

  3. 3.

    Case 3: the seven states \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) along with all the model parameters \(k_{1}\), \(k_{2}\), \(k_{3}\), \(c_{1}\), \(c_{2}\), \(c_{3}\), \(\beta\), \(\gamma\), and n will be estimated. In case 3 where all the parameters are estimated, we have added constraint on the parameter n. In fact, in the spring–mass–dashpot system, the nonlinear spring is mathematically ill-defined for negative values of n and physically the values of n cannot be smaller than 1 [37]. So when generating the sigma points, if the value of n is smaller than 1, we shift the mean for the symmetric sigma points by the small amount as follows: \(Z(16,i)=(1+0.5*\mathrm{rand})*\mathrm{constraints.value}\) (where \(i=\) number of sigma points and constraints.value = 1).

The estimation of the state variables and parameter(s) for these three cases is performed using UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF, and the simulation results for the state variables and the model parameters for the three cases are shown in Tables 5, 6 and 7. For example, for case 1, Table 5 compares the estimation root mean square errors for the seven state variables \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) (with respect to the noise-free data) and the mean of the estimated parameter \(k _1\) at steady state (i.e., after convergence of parameter(s)) using UKF, SRCDKF, IUKF, ISRCDKF, PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF, respectively. Tables 6 and  7 also present similar comparisons for cases 2 and 3, respectively.

The results also show that the number of parameters to estimate affects the estimation accuracy of the state variables. In other words, for PF the estimation RMSE of \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) increases from the first comparative study (where only the state variables are estimated) to case 1 (where the states and one parameter, \(k _1\), are estimated), case 2 (where the states and two parameters, \(k _1\) and \(k _2\), are estimated) to case 3 (where the states and nine parameters, \(k_{1}\), \(k_{2}\), \(k_{3}\), \(c_{1}\), \(c_{2}\), \(c_{3}\), \(\beta\), \(\gamma\), and n, are estimated). For example, the RMSEs obtained using PF for \(x_1\) in the first comparative study and cases 1, 2, 3 of the second comparative study are 0.0096, 0.0119, 0.0128 and 0.0180, respectively, which increase as the number of estimated parameters increases (see Tables 56, 7). This observation is valid for the other state variables \(x_2\), \(x_3\), \(z_1\), \(v_1\), \(v_2\) and \(v_3\) and for the other state estimation techniques UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF.

In Fig. 15, the initial value of the parameter n has automatically been changed in all the filters as the UKF, SRCDKF, IUKF et ISRCDKF. This change is due to the constraint applied on that parameter n because it value should not be smaller than 1.

It can also be shown from Tables 56 and 7 that, for all the techniques, estimating more model parameters affects the estimation accuracy. The ISRCDKF-PF method, however, still provides advantages over other methods in terms of the estimation accuracy. These advantages of the ISRCDKF-PF are due to the fact that it uses a better proposal distribution that takes the latest observation into account by using ISRCDKF.

It can also be shown from Tables 56 and 7 that, for all the techniques, estimating more model parameters affects the estimation accuracy. The ISRCDKF-PF method, however, still provides advantages over other methods in terms of the estimation accuracy. These advantages of the ISRCDKF-PF are due to the fact that it uses a better proposal distribution that takes the latest observation into account by using ISRCDKF.

Table 5 Root mean square errors of estimated state variables and mean of estimated parameter
Table 6 Root mean square errors of estimated state variables and mean of estimated parameter
Table 7 Root mean square errors of estimated state variables and mean of estimated parameter
  1. (b)

    Effect of noise content on the estimation RMSE

It is assumed that a noise is added to the state variables. In order to show the performance of PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF estimation algorithms in the presence of noise, three different measurements noise values, \(10^{-1}\), \(10^{-2}\) and \(10^{-3}\) are considered. The simulation results of estimating the seven states \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) using PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF methods when the noise levels vary in {\(10^{-1}\), \(10^{-2}\) and \(10^{-3}\)} are shown in Tables 8, 9, 11, 10 and 12.

In other words, for the PF, UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF estimation techniques, the estimation RMSEs of \(x_1\), \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) increase from the first comparative study (noise variance = \(10^{-1}\)) to case (where the noise value = \(10^{-4}\)). For example, the RMSEs obtained using PF for \(x_1\) where the noise level in \(\{10^{-1}\), \(10^{-2}\) and \(10^{-3}\)} are 0.0125, 0.0096 and 0.0093, which increase as the noise variance increases (refer to Tables 8, 9, 11, 1012). This observation is valid for the other state variables \(x_2\), \(x_3\), \(r_1\), \(v_1\), \(v_2\) and \(v_3\) and for the UKF-PF, IUKF-PF, SRCDKF-PF and ISRCDKF-PF algorithms.

Table 8 Root mean square errors (RMSE) of the estimated states using PF for different noise levels
Table 9 Root mean square errors (RMSE) of the estimated states using UKF-PF for different noise levels
Table 10 Root mean square errors (RMSE) of the estimated states using SRCDKF-PF for different noise levels
Table 11 Root mean square errors (RMSE) of the estimated states using IUKF-PF for different noise levels
Table 12 Root mean square errors (RMSE) of the estimated states using ISRCDKF-PF for different noise levels

The results of the ISRCDKF-PF are consistent for different state and/or parameter initial conditions. This is because, unlike other optimization techniques (which can easily converge at a local minima), the ISRCDKF-PF is a Monte Carlo-based technique which searches for the optimum solution of the state and/or parameter by generating a large number of samples that are used to approximate the posterior density function. These randomly generated samples that are used to approximate the posterior are independent of the initial conditions, making the ISRCDKF-PF robust to the choice of any initial conditions.

6 Conclusions

In this paper, the problem of state and parameter estimations of structural systems were addressed using the developed iterated square-root central difference Kalman particle filter (ISRCDKF-PF). Various conventional and state-of-the-art state estimation methods are compared for the estimation performance (the unscented Kalman filter (UKF), the square-root central difference Kalman filter (SRCDKF), the iterated unscented Kalman filter (IUKF), the iterated square-root central difference Kalman filter (SRCDKF), the conventional particle filter (PF), the unscented Kalman particle filter (UKF-PF), the square-root central difference Kalman particle filter (SRCDKF-PF), the iterated unscented Kalman particle filter (IUKF-PF) and the developed ISRCDKF-PF) in two comparative studies. In the first comparative study, the displacements and the velocities state variables are estimated from noisy measurements of these variables, and the various estimation techniques are compared by computing the estimation root mean square error with respect to the noise-free data. In the second comparative study, the state variables as well as the model parameters are simultaneously estimated. In this case, in addition to comparing the performances of the various state estimation techniques, the effect of the number of estimated model parameters on the accuracy and convergence of these techniques is also assessed. The results of the second comparative study show that, for all the techniques, estimating more model parameters affects the estimation accuracy as well as the convergence of the estimated states and parameters. The developed ISRCDKF-PF method, however, still provides advantages over other methods in terms of the estimation accuracy and convergence.