Abstract
Chirp signal models and their generalizations have been used to model many natural and man-made phenomena in signal processing and time series literature. In recent times, several methods have been proposed for parameter estimation of these models. However, these methods are either statistically sub-optimal or computationally burdensome, especially for two dimensional chirp models. In this paper, we consider the problem of parameter estimation of two dimensional chirp models and propose a computationally efficient estimator and establish asymptotic theoretical properties of the proposed estimators. Moreover, the proposed estimators are observed to have the same rates of convergence as the least squares estimators. Further, the proposed estimators of chirp rate parameters are shown to be asymptotically optimal. Extensive and detailed numerical simulations are conducted, which support the theoretical results of the proposed estimators.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
The parameter estimation problem for two dimensional (2D) chirp models is encountered in many real-life applications such as 2D-homomorphic signal processing, magnetic resonance imaging (MRI), optical imaging, interferometric synthetic aperture radar (INSAR), modeling non-homogeneous patterns in the texture image captured by a camera due to perspective or orientation (see e.g., Francos & Friedlander, 1995, 1998, 1999 and the references cited therein). 2D chirp signals have also been used to model Newton’s rings (Guo & Li, 2018). These rings are predominantly used in testing spherical and flatting optical surface and curvature radius measurement. 2D chirp signals have been used as a spreading function or base for digital watermarking (Stankovic et al., 2001), which is also helpful in data security, medical safety, fingerprinting, and observing content manipulations, see (Zhang et al., 2010). 2D chirps with linear frequency modulation have been extensively used because of their linear feature in time-frequency domain. These models have been employed in many applications due to other important properties also, e.g., spectral shaping of the watermark by choosing appropriately the frequency and frequency rate parameters, which allows minimum overlap of the spreading function with the image data. 2D chirps also allow adaptive watermarking for enhanced robustness to stationary filtering attacks, i.e., filtering whose parameters do not change over an image, see (Stankovic et al., 2001).
Many algorithms based on different approaches have been put forward in the literature to solve such problems. Polynomial phase differencing (PD) operator was introduced in Friedlander and Francos (1995) as an extension of the polynomial phase transform proposed in Peleg and Porat (1991). Several works Friedlander and Francos (1996), (1998) and (1999) utilized PD operator to develop computationally efficient algorithms for estimating similar polynomial phase signals. Cubic phase function (CPF) proposed in O’shea (2002), was extended in Zhang et al. (2008) for similar 2D chirp signal modeling. Further, CPF was utilized to estimate 2D cubic phase signal using genetic algorithm in Djurović et al. (2010). Consistency and asymptotic normality of LSEs for a general 2D polynomial phase signal (PPS) model have been derived in Lahiri and Kundu (2017). A finite step computationally efficient procedure for a similar 2D chirp signal model proposed in Lahiri et al. (2013) was proved asymptotically equivalent to LSEs. Quasi-Maximum Likelihood (QML) algorithm Djurović and Stanković 2014 proposed for 1D PPS, was generalized for 2D PPS in Djurović (2017). Further approximate least squares estimators (ALSEs) proposed in Grover and Kundu (2018) have been proved to be asymptotically equivalent to LSEs. An efficient estimation procedure based on fixed dimension technique, presented in Grover et al. (2021) was shown to be asymptotically equivalent to the optimal LSEs, for a 2D chirp model without the product term \(mn\), e.g., compare model (1) and the one considered in Grover et al. (2021).
Estimators based on phase differencing strategies or high order ambiguity function (HAF) or some of their modifications are computationally easier to obtain. However, the performance of estimation deteriorate below a relatively high signal-to-noise ratio (SNR) threshold and are sub-optimal. Methods that use PD in the steps of estimation, usually estimate coefficients of the highest degree first, and then subsequently estimate the coefficients of a lower degree from the demodulated or de-chirped signal. Therefore, the estimation error of highest degree coefficients accumulates and affects estimation accuracy of lower degree coefficients quite seriously. For more details, one can refer to Barbarossa et al. (1998), Djurović (2017) and Wu et al. (2008). Till date, there is no detailed study of the theoretical properties of the estimators CPF and QML, such as strong consistency and asymptotic normality. Recently, optimal estimators for a simpler 2D chirp model without the interaction term have been developed in Grover et al. (2021). However, the results in Grover et al. (2021) cannot be generalized directly for the underlying model (1). It may be noted that the model considered in this paper is more general, as it takes into account the interaction term \(\mu ^0 mn\). Due to the presence of this interaction term coefficient \(\mu ^0\), the estimation becomes more difficult as the estimators of \(\alpha ^0\) and \(\gamma ^0\) are no longer independent (as in the case for Grover et al., 2021), and hence making their computation as well as the study of theoretical analysis becomes more challenging. The problem becomes more complicated under the assumption of general stationary linear process error assumption.
The main contributions of this paper are; providing a computationally efficient algorithm to estimate the parameters of the model defined in (1), and further establishing theoretical asymptotic properties of the proposed estimators. The proposed algorithm is motivated by the fact that a 2D chirp model with five non-linear parameters can be viewed through a number of 1D chirp models with two non-linear parameters, and hence computational complexity of estimators for the 2D models can be reduced.
The key attributes of the proposed method are that it is computationally faster than the conventional optimal methods such as LSEs, maximum likelihood estimators, or ALSEs and at the same time, having desirable statistical properties such as, attaining the same rates of convergence as the optimal LSEs. In fact the proposed estimators of the chirp rate parameters have the same asymptotic variance as that of the traditional LSEs, and hence are asymptotically optimal.
The rest of the paper is organised as follows: the mathematical model and the methodology to obtain the proposed estimators is presented in Sect. 2. The model assumptions and the asymptotic theoretical results are given in Sect. 3. In Sect. 4, the finite sample performance of the proposed estimators is demonstrated through simulation studies. In this section, a comparison of the performance of the proposed estimators with the state-of-the-art methods such as the least squares method, approximate least squares method, and 2D multilag HAF method is also presented. Finally, Sect. 5 concludes the paper, followed by detailed proofs in appendices.
2 Estimation methodology
This paper addresses the problem of parameter estimation of a 2D chirp signal model defined as follows:
Here, y(m, n) is the observed real valued signal and X(m, n) is the additive noise term. \(A^0,B^0\) are amplitude parameters, \(\alpha ^0,\gamma ^0\) are frequency parameters, \(\beta ^0,\delta ^0\) are frequency rates or chirp rates, and \(\mu ^0\) is the coefficient of product term. \(\mathrm {{\varvec{\xi }}}^0 = (\alpha ^0,\beta ^0, \gamma ^0, \delta ^0, \mu ^0)^\top \) represents vector of non-linear parameters. This model can be used to describe signals having constant amplitude with frequency to be a linear function of spatial co-ordinates. The product term \(mn\) in such chirp models (1), is an important characteristic of numerous measurement interferometric signals, radar signal returns and detecting digital watermarking.
Now we discuss the proposed method of estimation. Let the data matrix for model (1) be denoted as
The proposed algorithm uses the fact that for each fixed column (or row) of \({\varvec{Y}}\), the 2D chirp model breaks down to a cascade of 1D chirp models.
Realise that if we fix one dimension say \(n=n_0\) in (1), then the 2D chirp can be seen as 1-D chirp for \(m=1,2,\ldots ,M\), as follows:
Similarly, for a fixed \(m=m_0\), we have 1-D chirp for \(n=1,2,\ldots ,N,\)
Equation (2) represents 1-D chirp signal model with \(\alpha ^0+n_0\mu ^0\) and \(\beta ^0\) as the frequency and frequency rate parameters respectively. Similarly, Eq. (3) represents 1-D chirp signal model with \(\gamma ^0+m_0\mu ^0\) and \(\delta ^0\) as the frequency and frequency rate parameters respectively.
Hence, our methodology is developed by estimating parameters of these 1D chirps based on a particular column (or row) vector of data matrix, rather than estimating the whole 2D chirp parameters based on the full data matrix. Therefore this procedure reduces computational burden to estimate model parameters drastically. Further suppose column vector \({\varvec{Y}}_{Mn_0}\) denotes the \(n_0^{th}\) column of data matrix \({\varvec{Y}}\) and column vector \({\varvec{Y}}_{m_0N}\) denotes the transpose of \(m_0^{th} \) row of data matrix \({\varvec{Y}}\). Define \(Z_k(\alpha _1,\alpha _2)\) matrix as
We need to estimate the model parameters of (2) which is a 1D chirp model, so we use LSEs to estimate the parameters. We obtain LSEs of non-linear parameters in model (2) for a fixed \(n_0\) by defining following reduced sum of squares, see (Grover et al., 2021):
where, \(P_{Z_M}(\alpha _1,\alpha _2) = Z_M(\alpha _1,\alpha _2)\Big (Z_M(\alpha _1,\alpha _2)^\top Z_M(\alpha _1,\alpha _2)\Big )^{-1}Z_M(\alpha _1,\alpha _2)^\top \) and \(I_{M\times M}\) is the \(M\times M\) identity matrix.
Then,
is the proposed estimator of \((\alpha ^0+n_0\mu ^0,\beta ^0)^\top \) based on minimizing the sum of squares corresponding to \(n_0^{th}\) column of the data matrix \({\varvec{Y}}\). Similarly, we can obtain LSEs of parameters in model (3) by defining
where, \(P_{Z_N}(\alpha _1,\alpha _2) = Z_N(\alpha _1,\alpha _2)\Big (Z_N(\alpha _1,\alpha _2)^\top Z_N(\alpha _1,\alpha _2)\Big )^{-1}Z_N(\alpha _1,\alpha _2)^\top \) and \(I_{N\times N}\) is the \(N\times N\) identity matrix. Then,
will be the proposed estimator of \((\gamma ^0+m_0\mu ^0,\delta ^0)^\top \) based on minimizing the sum of squares corresponding to \(m_0^{th}\) row of data matrix \({\varvec{Y}}\).
We observe that for each fixed column, we get an estimate of the same chirp rate parameter \(\beta ^0\) in (6), and also an estimate of frequency parameter, which is a linear combination of \(\alpha ^0\) and \(\mu ^0\). Similarly, estimates of \(\delta ^0\) and a linear combination of \(\gamma ^0\) and \(\mu ^0\) for a fixed row in (8) has been obtained. It is important to note that the linearity of parameters of 1D-chirp models plays a crucial role in getting proposed estimators of \(\alpha ^0,\gamma ^0\), and \(\mu ^0\) by fitting a linear regression model as follows.
Once the parameters corresponding to each \((M+N)\) 1-D chirp models have been estimated. We apply the following three steps to obtain final estimates of parameters of the model (1):
-
Step-1.
Let \({\varvec{\Gamma }}^\top =\begin{bmatrix} 1&{}1&{}\cdots &{}1&{}0&{}0&{}\cdots &{}0\\ 0&{}0&{}\cdots &{}0&{}1&{}1&{}\cdots &{}1\\ 1&{}2&{}\cdots &{}N&{}1&{}2&{}\cdots &{}M \end{bmatrix}\), and \({\varvec{\Lambda }}^\top =\begin{bmatrix} {\widehat{\alpha }}_1&{\widehat{\alpha }}_2&\cdots&{\widehat{\alpha }}_N&{\widehat{\gamma }}_1&{\widehat{\gamma }}_2&\cdots&{\widehat{\gamma }}_M \end{bmatrix}\). Combine the obtained estimates as follows:
$$\begin{aligned} {\varvec{\Lambda }}={\varvec{\Gamma }}\begin{bmatrix} \alpha \\ \gamma \\ \mu \end{bmatrix}. \end{aligned}$$(9)Then estimate of \((\alpha ^0,\gamma ^0,\mu ^0)^\top \) is \(\Big ({\varvec{\Gamma }}^\top {\varvec{\Gamma }}\Big )^{-1}{\varvec{\Gamma }}^\top {\varvec{\Lambda }}\).
-
Step-2.
The estimates of \(\beta ^0\) and \(\delta ^0\) are simply the averages \(\displaystyle {\widehat{\beta }}= \frac{1}{N}\sum _{n=1}^{N}{\widehat{\beta }}_{n}\) and \(\displaystyle {\widehat{\delta }} =\frac{1}{M}\sum _{m=1}^{M}{\widehat{\delta }}_{m}\), respectively.
-
Step-3.
After getting estimates of non-linear parameters, \(\widehat{\mathrm {{\varvec{\xi }}}} = ({\widehat{\alpha }},{\widehat{\beta }},{\widehat{\gamma }},{\widehat{\delta }}, {\widehat{\mu }})^\top \), the amplitude parameter estimates can be provided as follows:
3 Theoretical results
In this section, we first state the model assumptions required to derive the theoretical asymptotic properties explicitly. These are as follows:
Assumption 1
X(m, n) can be expressed as a linear combination of a double array sequence of independently and identically distributed (i.i.d.) random variables \(\{\epsilon (m,n)\}\) with mean 0, variance \(\sigma ^2\) and finite fourth moment.
such that
Assumption 2
True parameter \( \mathrm {{\varvec{\theta }}}^0 = (A^0,B^0,\alpha ^0,\beta ^0,\gamma ^0,\delta ^0,\mu ^0)^\top \) is an interior point of parameter space \(\Theta \), where \(\Theta = (-M,M)\times (-M,M)\times [0,2\pi ]\times [0,\pi /2]\times [0,2\pi ]\times [0,\pi /2]\times [0,2\pi ]\), and \(A^{0^2}+B^{0^2}>0\), for some \(M>0\).
Assumption 1 puts the model under a very general set-up of noise contamination as it includes the dependent relationship too. Assumption 2 is taken to assure the absence of any identifiability problem and non-zero deterministic part of the signal. Under these general assumptions, we have derived strong consistency and asymptotic normality of the estimators. The obtained results are stated in the following theorems.
Theorem 1
Under assumptions 1 and 2, the proposed estimator of parameter \(\mathrm {{\varvec{\theta }}}^0\) is strongly consistent, i.e.,
Proof
Please see Appendix A for the proof. \(\square \)
Theorem 2
Under assumptions 1 and 2, the proposed estimators of \(\mathrm {{\varvec{\theta }}}^0\) is asymptotically normally distributed.
where \(c = \displaystyle {\sum _{i=-\infty }^{\infty }\sum _{j=-\infty }^{\infty }}a(i,j)^2\), \({\varvec{D}}^{-1} = diag(M^{\frac{1}{2}}N^{\frac{1}{2}}, M^\frac{1}{2}N^{\frac{1}{2}}, M^{\frac{3}{2}}N^{\frac{1}{2}}, M^{\frac{5}{2}}N^{\frac{1}{2}}, M^{\frac{1}{2}}N^{\frac{3}{2}}, M^{\frac{1}{2}}N^{\frac{5}{2}}, M^{\frac{3}{2}}N^{\frac{3}{2}} )\), and
Here \(diag(a_1, a_2,\ldots , a_k)\) represents \(k\times k\) diagonal matrix with elements \(a_1, a_2,\ldots , a_k\) in the principal diagonal and \({\mathcal {N}}_k({\varvec{{\mathcal {M}}}},{\varvec{{\mathcal {S}}}})\) represents the k-variate normal distribution with mean vector \({\varvec{{\mathcal {M}}}}\) and variance-covariance matrix \({\varvec{{\mathcal {S}}}}\).
Proof
Please see Appendix B for the proof. \(\square \)
Although Theorem 2 has been proved for the increasing sample size assuming \(M=N\rightarrow \infty \), however asymptotic normality will still hold even if \(M/N\rightarrow p \) as \(M,N\rightarrow \infty \), for some \(p>0\). It is interesting to note that the asymptotic properties of the proposed estimators of chirp rates \(\beta ^0,\delta ^0 \) will remain the same even if we take \( \min \{M,N\}\xrightarrow {}\infty \). The asymptotic variance-covariance matrix of \((\alpha ^0,\gamma ^0,\mu ^0)\) will however change depending on the value of p among non-linear parameters.
If we further assume that the errors in (1) are i.i.d. Gaussian distributed random variables, then it can be observed that the proposed estimators of chirp rates parameters \(\beta ^0\) and \(\delta ^0 \), asymptotically attain Cramer-Rao lower bound (CRLB). CRLB for estimators of other non-linear parameters \(\alpha ^0\), \(\gamma ^0\), and \(\mu ^0\) are \( \frac{456c\sigma ^2}{(A^{0^2}+B^{0^2})}\), \( \frac{456c\sigma ^2}{(A^{0^2}+B^{0^2})}\) and \( \frac{288c\sigma ^2}{(A^{0^2}+B^{0^2})} \), see (Lahiri & Kundu, 2017).
4 Simulation results
Simulation studies performed in this paper are divided into three parts. The first part demonstrates the evaluation of finite sample size performance of proposed estimators. We compare the performance of the proposed estimators with the asymptotically optimal estimators such as LSEs, and ALSEs, and fast but sub-optimal 2D-multilag-HAF estimators. The second part shows the lower computational cost of the proposed estimators as compared to the LSEs. Finally, the third part exemplifies the ability of the proposed estimators to extract original gray-scale texture from one contaminated with noise and reproduce the original texture. We have performed simulations on the complex counterpart of the model (1), (see Barbarossa, 2014) for comparison purposes.
4.1 Finite sample performance
To provide a detailed assessment of the performance of proposed estimators, we have chosen sample sizes \(M=N=20,40,60,80\) and 100. The fixed values of all parameters to obtain complex-valued chirp data are:
Obtained data from the model is then contaminated with noise X(m, n). We consider two distinct noise structures for our simulations. These are:
-
Independently and identically distributed (i.i.d.) normal errors with mean 0 and variance \(\sigma ^2\);
-
Autoregressive moving average (ARMA) errors with following representation:
$$\begin{aligned} X(m,n) =\;&0.06X(m-1,n-1)-0.054X(m,n-1)+0.087X(m-1,n)\nonumber \\&+\epsilon (m,n)+.01\epsilon (m-1,n-1)+0.035\epsilon (m,n-1)+0.042\epsilon (m-1,n). \end{aligned}$$(14)where \(\epsilon (m,n)\) is a sequence of i.i.d. Gaussian random variables with mean 0 and variance \(\sigma ^2\).
We have obtained the estimates for 1000 replications for a fixed sample size \(M=N\), under a particular error structure with fixed \(\sigma ^2\). The estimators do not have any explicit closed form expression, so we use Nelder-Mead algorithm ( using “optim" function in R software) for optimization of the objective function and to obtain the estimators. Mean square errors (MSEs) obtained under 1000 replications are displayed in the Figs. 1a, b, 2a and b for four different values of \(\sigma \) = 0.1,0.5,0.9,1. The MSEs are plotted on negative logarithmic scale. The findings of these simulation results can be summarized as follows:
-
MSEs of the proposed estimators decrease rapidly as sample size increases, which supports the consistency property of the estimators. Further, as sample size increases, the gap between the MSEs of the proposed estimators and LSEs decreases.
-
The obtained MSEs of proposed estimators of \(\beta ^0\) and \(\delta ^0\), are at par with those of LSEs and ALSEs.
4.2 Time comparison
The computational advantage of the proposed estimators over the conventional LSEs is quite significant. In order to compare the two methods, we measure their computational complexities in terms of the number of points in the grid that are needed to find the initial guesses of these estimators. Once we have the precise initial guesses, applying an iterative algorithm like Nelder-Mead is a matter of seconds. “gridSearch" function from the package “NMOF" is used to calculate the initial guesses. We report observed time to get the estimates for a fixed sample size and the total number of grid points over which cost function evaluations are required, in Table 1. The choice of parameters is same as in (13) along with i.i.d. normal errors with mean 0 and standard deviation \(\sigma =0.9\). For a fixed sample size \(M=N\), the order of computation for LSEs is \(M^4N^4=M^8\), but for the proposed method, the order of computation is \(M^3N+N^3M=2M^4\). The numerical experiments for comparing time efficiency were performed on a system with processor: Intel(R) Core(TM) i3-5005U CPU @ 2.00GHz 2.00GHz; installed memory(RAM): 4.00 GB; and system type: 64-bit Operating System. Codes were written and run in R version 4.0.4 (2021-02-15)—“Lost Library Book", a free software environment for statistical computing and graphics.
For the considered machine, it was not feasible to perform grid-search to get LSEs for more than \(M=N=7\). When we plot logarithm sample size \(M(=N)\) against the logarithm time to compute initial guess for LSEs, then it is observed to be linear. So, to get an idea of the time deviation of LSEs with that of proposed estimates at larger sample sizes, we predict time to obtain LSEs based on grid-search by fitting a simple linear regression model between \(\log \) of sample size and \(\log \) of time to get LSEs. From the results in Table 2, we can clearly observe the massive time difference of getting proposed estimates and LSEs. For example, if we go for sample size, say \(M=N=40\), then it will take more than 20 years to obtain LSEs using grid-search over the same machine (even if we assume a large amount of RAM in a machine), while it took less than 10 min to obtain the proposed estimates.
4.3 Texture pattern estimation
2D chirp signals create interesting gray-scale texture patterns. In order to analyze the effectiveness of the proposed estimators for estimating texture patterns accurately, we generate data from the complex counterpart of the model with same set of parameters as in (13). Then real and imaginary part of the obtained data is contaminated independently with i.i.d. normal errors having mean 0 and variance \(\sigma ^2=0.09\). The data matrix obtained is of size \(100\times 100\). We analyze this data using the proposed estimator, the optimal LSEs, ALSEs, and also with the 2D-multilag-HAF method. Note that we have used true values as initial guess for obtaining optimal estimators, LSEs and ALSEs because of the computational complexity and grid-search method for obtaining 2D-multilag-HAF estimators and proposed estimators.
Plugging these estimators in the deterministic part of the model, we get estimated texture patterns as the real part of the reproduced data. We also present real part of the original dataset to compare the original and estimated textures. It is clear from the figures that texture pattern obtained using the proposed method is visually the same as that obtained using optimal LSEs and ALSEs, while the 2D-multilag-HAF estimator gives a slightly different pattern than the original one (Fig. 3).
5 Conclusion
The paper proposes a computationally efficient estimators with the same convergence rate as the LSEs or ALSEs. The key idea is to develop a strategy by disintegrating the 2D model into several 1D chirp models and then design an optimal estimation method to obtain estimates of the 2D model parameters. The proposed estimators are not only asymptotically unbiased but also have an asymptotic normal distribution and same rate of convergence as that of the LSEs. Furthermore, the estimators converge strongly to the true value of the parameters. Extensive numerical simulations firmly support the theoretical results and also unveil the gigantic gap between time required for obtaining proposed estimates and the LSEs. Synthetic data analysis illustrates the effectiveness of the proposed estimators in recovering 2D gray-scale textures contaminated with noise. For several applications, like enhancing the robustness of the watermarking from frequent attacks e.g., cropping, rotating, compressing, etc., multi-component chirp signals are required for the base watermarking, see Stankovic et al. (2001). The proposed method can effectively be used to address such problems. Further, our proposed methodology can be adapted for parameter estimation of more general signal models by utilizing a sequential procedure similar to Grover et al. (2021). We believe that the estimation technique and the results obtained in this paper can lead to designing computationally efficient algorithms for estimating higher order polynomial phase signals and thus making a significant contribution towards research in this area.
Data Availability
Not applicable.
Code Availability
The authors can share the code used for simulations, on individual requests.
References
Barbarossa, S., Scaglione, A., & Giannakis, G. B. (1998). Product high-order ambiguity function for multicomponent polynomial-phase signal modeling. IEEE Transactions on Signal Processing, 46(3), 691–708.
Barbarossa, S., Di Lorenzo, P., & Vecchiarelli, P. (2014). Parameter estimation of 2D multi-component polynomial phase signals: An application to SAR imaging of moving targets. IEEE Transactions on Signal Processing, 62(17), 4375–4389.
Djurović, I., Wang, P., & Ioana, C. (2010). Parameter estimation of 2-D cubic phase signal using cubic phase function with genetic algorithm. Signal Processing, 90(9), 2698–2707.
Djurović, I., & Stanković, L. (2014). Quasi-maximum-likelihood estimator of polynomial phase signals. IET Signal Processing, 8(4), 347–359.
Djurović, I. (2017). Quasi ML algorithm for 2-D PPS estimation. Multidimensional Systems and Signal Processing, 28(2), 371–387.
Francos, J. M., & Friedlander, B. (1995). The polynomial phase difference operator for modeling of nonhomogeneous images. In Proceedings., International Conference on Image Processing (Vol. 2, pp. 276–279). IEEE.
Friedlander, B., & Francos, J. M. (1996). Model based phase unwrapping of 2-D signals. IEEE Transactions on Signal Processing, 44(12), 2999–3007.
Francos, J. M., & Friedlander, B. (1998). Two-dimensional polynomial phase signals: Parameter estimation and bounds. Multidimensional Systems and Signal Processing, 9(2), 173–205.
Francos, J. M., & Friedlander, B. (1999). Parameter estimation of 2-D random amplitude polynomial-phase signals. IEEE Transactions on Signal Processing, 47(7), 1795–1810.
Fuller, W. A. (1996). Introduction to Statistical Time Series (2nd ed.). John Wiley and Sons.
Grover, R., Kundu, D., & Mitra, A. (2018). Approximate least squares estimators of a two-dimensional Chirp model and their asymptotic properties. Journal of Multivariate Analysis, 168, 211–220.
Grover, R., Kundu, D., & Mitra, A. (2021). An efficient methodology to estimate the parameters of a two-dimensional Chirp signal model. Multidimensional Systems and Signal Processing, 32(1), 49–75.
Guo, Y., & Li, B. Z. (2018). Novel method for parameter estimation of Newton’s rings based on CFRFT and ER-WCA. Signal Processing, 144, 118–126.
Lahiri, A., Kundu, D., & Mitra, A. (2013). Efficient algorithm for estimating the parameters of two dimensional Chirp signal. Sankhya B, 75(1), 65–89.
Lahiri, A., Kundu, D., & Mitra, A. (2015). Estimating the parameters of multiple chirp signals. Journal of Multivariate Analysis, 139, 189–206.
Lahiri, A., & Kundu, D. (2017). On parameter estimation of two-dimensional polynomial phase signal model. Statistica Sinica, 27, 1779–1792.
Montgomery, H. L. (1994). Ten lectures on the interface between analytic number theory and harmonic analysis (No. 84). American Mathematical Soc.
Nandi, S., & Kundu, D. (2004). Asymptotic properties of the least squares estimators of the parameters of the chirp signals. Annals of the Institute of Statistical Mathematics, 56(3), 529–544.
O’shea, P. (2002). A new technique for instantaneous frequency rate estimation. IEEE Signal Processing Letters, 9(8), 251–252.
Peleg, S., & Porat, B. (1991). Estimation and classification of polynomial-phase signals. IEEE Transactions on Information Theory, 37(2), 422–430.
Stankovic, S., Djurovic, I., & Pitas, I. (2001). Watermarking in the space/spatial-frequency domain using two-dimensional Radon-Wigner distribution. IEEE Transactions on Image Processing, 10(4), 650–658.
Wu, Y., So, H. C., & Liu, H. (2008). Subspace-based algorithm for parameter estimation of polynomial phase signals. IEEE Transactions on Signal Processing, 56(10), 4977–4983.
Zhang, K., Wang, S., & Cao, F. (2008). Product cubic phase function algorithm for estimating the instantaneous frequency rate of multicomponent two-dimensional Chirp signals. In 2008 Congress on Image and Signal Processing (Vol. 5, pp. 498–502). IEEE.
Zhang, Y., Mobasseri, B. G., Dogahe, B. M., & Amin, M. G. (2010). Image-adaptive watermarking using 2D Chirps. Signal, Image and Video Processing, 4(1), 105–121.
Acknowledgements
We would like to thank the Editor for his encouragement and continuous support, and anonymous reviewer for providing critical and constructive comments. Part of the work of the third author has been supported by a research grant from the Science and Engineering Research Board, Government of India.
Funding
Part of the work of the third author has been supported by a research grant from the Science and Engineering Research Board, Government of India.
Author information
Authors and Affiliations
Contributions
All authors contributed to study the conceptualization, methodology, review and editing. Formal analysis of proofs and simulations analysis were performed by AS. The first draft of the manuscript was written by AS and all authors commented on revising the previous versions of the manuscript. All authors read and approved the final manuscript. AM and DK had supervised the whole research project.
Corresponding author
Ethics declarations
Conflict of interest
The authors declares that they have no conflict of interest.
Consent for Publication
Not applicable.
Ethical Approval
Not applicable.
Consent to Participate
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A
Proof of Theorem 1:
Given the data matrix \({\varvec{Y}}\), we compute the LSEs of \(\alpha ^0+n_0\mu ^0\) and \(\beta ^0\) corresponding to \(n_0^{th}\) column vector of \({\varvec{Y}}\). We denote the obtained estimators by \({\widehat{\alpha }}_{n_0}\) and \({\widehat{\beta }}_{n_0}\) to emphasize that these depend on \(n_0\). Similarly, for fixed \(m_0^{th}\) row of \({\varvec{Y}}\), we have denoted the LSEs of \(\gamma ^0+m_0\mu ^0\) and \(\delta ^0\) by \({\widehat{\gamma }}_{m_0}\) and \( {\widehat{\delta }}_{m_0}\). Under assumptions that \(X(m_0,n_0)\) are stationary (11) and (12), see Nandi and Kundu (2004), we have
The final estimator of \(\beta ^0\) given by
is strongly consistent estimate of \(\beta ^0\) which is observed by (15) and also that \( {\widehat{\beta }}_{n_0}\) is strongly consistent for \(\beta ^0\) as \(M\xrightarrow {} \infty \). For proof, one may refer to Lahiri et al. (2015).
Similarly \({\widehat{\delta }}=\displaystyle \frac{1}{M}\sum _{m_0=1}^{M}{\widehat{\delta }}_{m_0}\) is strongly consistent estimator of \(\delta ^0\).
Denote
We now prove the consistency of the frequency parameter estimators \( {\widehat{\alpha }}, {\widehat{\gamma }}\), and that of \({\widehat{\mu }}\), estimator of the interaction term parameter. From (15) and (16), we have thefollowing:
where \({\varvec{\Gamma }}^\top =\begin{bmatrix} 1&{}1&{}\cdots &{}1&{}0&{}0&{}\cdots &{}0\\ 0&{}0&{}\cdots &{}0&{}1&{}1&{}\cdots &{}1\\ 1&{}2&{}\cdots &{}N&{}1&{}2&{}\cdots &{}M \end{bmatrix}_{3\times (M+N)}\), and \({\varvec{\Lambda }}^\top =\begin{bmatrix} {\widehat{\alpha }}_1&{\widehat{\alpha }}_2&\cdots&{\widehat{\alpha }}_N&{\widehat{\gamma }}_1&{\widehat{\gamma }}_2&\cdots&{\widehat{\gamma }}_M \end{bmatrix}\).
This implies that
Now we look at the first element \(a_1+a_2-a_3\) of the following matrix
where,
and
Now we look at \(a_1, a_2\) and \(a_3\) individually and compute their limits.
\( \hbox { This implies that } a_1\xrightarrow {a.s.} 0\hbox { as } \min \{M,N\}\rightarrow \infty .\) Here,
\(\hbox { This implies that } a_2 \xrightarrow {a.s.} 0\hbox { as } \min \{M,N\}\rightarrow \infty .\)
\(\hbox { This implies that } a_3 \xrightarrow {a.s.} 0\hbox { as } \min \{M,N\}\rightarrow \infty .\) Hence, \({\widehat{\alpha }}\) is strongly consistent estimate of \(\alpha ^0\). Similarly strong consistency of \({\widehat{\gamma }}\) for \(\gamma ^0\) can be derived.
Now, the third element of \(({\varvec{\Gamma }}^\top {\varvec{\Gamma }})^{-1}\begin{bmatrix} N\times o\bigg (\frac{1}{M}\bigg )\\ M\times o\bigg (\frac{1}{N}\bigg )\\ \frac{N(N+1)}{2}\times o\bigg (\frac{1}{M}\bigg )+\frac{M(M+1)}{2}\times o\bigg (\frac{1}{N}\bigg ) \end{bmatrix}\) can be written as \(-b_1-b_2+b_3\), where
\(\hbox { This implies that } b_1\xrightarrow {a.s.} 0\hbox { as } \min \{M,N\}\rightarrow \infty .\)
\(\hbox { This implies that } b_2\xrightarrow {a.s.} 0\hbox { as } \min \{M,N\}\rightarrow \infty .\)
Hence \( b_3\xrightarrow {a.s.} 0\hbox { as } \min \{M,N\}\rightarrow \infty \). Hence, \({\widehat{\mu }}\) is a strongly consistent estimate of \(\mu ^0\). The strong consistency of estimators of amplitude parameters follows from the continuity mapping theorem. \(\square \)
Appendix B
Proof of Theorem 2:
Suppose \({\varvec{\kappa }}^\top =(A,B,\alpha ,\beta ) \) and
\({\varvec{\kappa }}^{0^\top }= (A^0(n_0),B^0(n_0),\alpha ^0+n_0\mu ^0,\beta ^0)\). We define sum of squares as follows:
Let \(\widehat{{\varvec{\kappa }}}\) be the minimizer of \(Q_{n_0}({\varvec{\kappa }})\), then using Taylor Series expansion on the first derivative vector \(Q_{n_0}^{\prime }({\varvec{\kappa }})\) around the point \({\varvec{\kappa }}^{0}\), we get:
where \(\breve{{\varvec{\kappa }}}\) is a point between \(\widehat{{\varvec{\kappa }}}\) and \({\varvec{\kappa }}^0\). Also \({\varvec{Q}}_{n_0}^{\prime }(\widehat{{\varvec{\kappa }}})=0\) as \(\widehat{{\varvec{\kappa }}}\) is LSE of \({\varvec{\kappa }}^0\).
Let us denote \({\varvec{D}}_1^{-1}=diag(M^{1/2},M^{1/2},M^{3/2},M^{5/2})\). Then on multiplying \({\varvec{D}}_1^{-1}\) both sides of Eq. (19), it gives:
and
The expression of \({\varvec{\Sigma }}_{n_0}^{-1}\) can be obtained by proof of Theorem 2 shown in Lahiri et al. (2015). By using Eq. (20) and above expression of \({\varvec{\Sigma }}_{n_0}^{-1}\), we get following asymptotically equivalent (a.e.) expression of the third element of vector \({\varvec{D_1}}^{-1}(\widehat{{\varvec{\kappa }}}-{\varvec{\kappa }}^0)\) as:
where, \(\eta (m_0,n_0)=X(m_0,n_0)\big (A^0\sin \phi (m_0,n_0,{\varvec{\xi }}^0)-B^0\cos \phi (m_0,n_0,{\varvec{\xi }}^0)\big )\),
\(\phi (m_0,n_0,{\varvec{\xi }}^0)= \alpha ^0m_0+\beta ^0m_0^2+\gamma ^0n_0+\delta ^0n_0^2+\mu ^0 m_0n_0\). We have used these notations for brevity. Similarly expressions of fourth, fifth and sixth element of \({\varvec{D_1}}^{-1}(\widehat{{\varvec{\kappa }}}-{\varvec{\kappa }}^0)\) can be written as in Eqs. (22), (23) and (24) respectively:
From Eqs. (22) and (24) above, we can see that estimators \({\widehat{\beta }}\) and \({\widehat{\delta }}\) of \(\beta ^0\) and \(\delta ^0\) are asymptotically equivalent to the LSEs as they have same asymptotic variances by applying central limit theorem for stationary linear processes, see Fuller (1996). So now, remaining is to show asymptotic properties of estimators \( ( {\widehat{\alpha }}, {\widehat{\gamma }}, {\widehat{\mu }})^\top \).
The expression of proposed estimators of \((\alpha ^0,\gamma ^0, \mu ^0)^\top \) obtained is:
From (25), we get:
For sufficiently large M and N, we have:
Asymptotic normality of the estimators for \(M=N\rightarrow \infty \) with given rates of convergence follows by applying central limit theorem for stationary processes (Fuller, 1996).
We now present some important results which will be used to find the asymptotic variance-covariance matrix of the proposed estimators of non-linear parameters. Using Eqs. (21) and (23) in the paper, we have the following observations:
-
1.
\(Asy Var \bigg (\frac{1}{2\sqrt{N}}\displaystyle \sum _{n_0=1}^{N}M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big )\bigg )=\frac{c\sigma ^2 96}{A^{0^2}+B^{0^2}}\),
-
2.
\(Asy Var \bigg (\frac{1}{2\sqrt{M}}\displaystyle \sum _{m_0=1}^MN^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=\frac{c\sigma ^296}{A^{0^2}+B^{0^2}}\),
-
3.
\(Asy Var \bigg (\frac{1}{2N^{3/2}}\displaystyle \sum _{n_0=1}^{N}n_0M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big )\bigg )=\frac{c\sigma ^232}{A^{0^2}+B^{0^2}}\),
-
4.
\(Asy Var \bigg (\frac{1}{2M^{3/2}}\displaystyle \sum _{m_0=1}^Mm_0N^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=\frac{c\sigma ^232}{A^{0^2}+B^{0^2}}\),
-
5.
\(Asy Covar \bigg (\frac{1}{2\sqrt{N}}\displaystyle \sum _{n_0=1}^{N}M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big ),\frac{1}{2\sqrt{M}}\displaystyle \sum _{m_0=1}^MN^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=0\),
-
6.
\(Asy Covar \bigg (\frac{1}{2\sqrt{N}}\displaystyle \sum _{n_0=1}^{N}M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big ),\frac{1}{2N^{3/2}}\displaystyle \sum _{n_0=1}^{N}n_0M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big )\bigg )=\frac{c\sigma ^248}{A^{0^2}+B^{0^2}}\),
-
7.
\(Asy Covar \bigg (\frac{1}{2\sqrt{N}}\displaystyle \sum _{n_0=1}^{N}M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big ),\frac{1}{2M^{3/2}}\displaystyle \sum _{m_0=1}^Mm_0N^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=0\),
-
8.
\(Asy Covar \bigg (\frac{1}{2\sqrt{M}}\displaystyle \sum _{m_0=1}^MN^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big ),\frac{1}{2N^{3/2}}\displaystyle \sum _{n_0=1}^{N}n_0M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big )\bigg )=0\),
-
9.
\(Asy Covar \bigg (\frac{1}{2\sqrt{M}}\displaystyle \sum _{m_0=1}^MN^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big ),\frac{1}{2M^{3/2}}\displaystyle \sum _{m_0=1}^Mm_0N^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=\frac{c\sigma ^248}{A^{0^2}+B^{0^2}}\),
-
10.
\(Asy Covar \bigg (\frac{1}{2N^{3/2}}\displaystyle \sum _{n_0=1}^{N}n_0M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big ),\frac{1}{2M^{3/2}}\displaystyle \sum _{m_0=1}^Mm_0N^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=\frac{c\sigma ^2}{2(A^{0^2}+B^{0^2})}\).
where \(c=\displaystyle {\sum _{i=-\infty }^{\infty }\sum _{j=-\infty }^{\infty }}a^2(i,j)\).
Using the above results in Eq. (26) from the paper, we get asymptotic variance-covariance matrix of
\(\begin{bmatrix} M^{3/2}N^{1/2}({\widehat{\alpha }}-\alpha ^0)\\ N^{3/2}M^{1/2}({\widehat{\gamma }}-\gamma ^0)\\ M^{3/2}N^{3/2}({\widehat{\mu }}-\mu ^0) \end{bmatrix}\) as: \( \frac{c\sigma ^2}{(A^{0^2}+B^{0^2})}\begin{bmatrix} 996&{}612&{}-1224\\ 612 &{}996&{}-1224\\ -1224 &{}-1224&{}2448 \end{bmatrix}. \)
From (21), (22) and (23) equations of the paper, it is further observed that:
-
1.
\(AsyCovar\bigg ( M^{5/2}N^{1/2}\big ({\widehat{\beta }}-\beta ^0\big ), M^{1/2}N^{5/2}\big ({\widehat{\delta }}-\delta ^0\big )\bigg )=0\),
-
2.
\(AsyCovar\bigg ( M^{5/2}N^{1/2}\big ({\widehat{\beta }}-\beta ^0\big ),\frac{1}{\sqrt{N}}\displaystyle \sum _{n_0=1}^{N}M^{3/2}\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big )\bigg )=\frac{-360c\sigma ^2}{A^{0^2}+B^{0^2}}\),
-
3.
\(AsyCovar\bigg ( M^{5/2}N^{1/2}\big ({\widehat{\beta }}-\beta ^0\big ),\frac{1}{\sqrt{M}}\displaystyle \sum _{m_0=1}^MN^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=0\),
-
4.
\(AsyCovar\bigg ( M^{5/2}N^{1/2}\big ({\widehat{\beta }}-\beta ^0\big ),\frac{1}{N\sqrt{N}}\displaystyle \sum _{n_0=1}^{N}M^{3/2}n_0\big ({\widehat{\alpha }}_{n_0}-(\alpha ^0+n_0\mu ^0)\big )\bigg )=\frac{-180c\sigma ^2}{A^{0^2}+B^{0^2}}\),
-
5.
\(AsyCovar\bigg ( M^{5/2}N^{1/2}\big ({\widehat{\beta }}-\beta ^0\big ),\frac{1}{M\sqrt{M}}\displaystyle \sum _{m_0=1}^Mm_0N^{3/2}\big ({\widehat{\gamma }}_{m_0}-(\gamma ^0+m_0\mu ^0)\big )\bigg )=0.\)
Similar results can be derived for \(M^{1/2}N^{5/2}\big ({\widehat{\delta }}-\delta ^0\big )\).
Asymptotic variance-covariance matrix of the proposed estimators of non-linear parameters is given by:
Next, we derive the asymptotics of amplitude estimators, please recall that by using Taylor series expansion of \(\cos \phi (m_0,n_0,\widehat{{\varvec{\xi }}})\) around the point \({\varvec{\xi }}^0\), we can write:
For brevity, we have denoted \(\cos \phi (m_0,n_0,\widehat{{\varvec{\xi }}})\) by \(\cos {\widehat{\phi }}\), \(\cos \phi (m_0,n_0,\widehat{{\varvec{\xi }}})\) by \(\cos \phi ^0\), and \(\sin \phi (m_0,n_0,\breve{{\varvec{\xi }}})\) by \(\sin \breve{\phi }\), where \(\breve{{\varvec{\xi }}}\) is a point lying between \(\widehat{{\varvec{\xi }}}\) and \({\varvec{\xi }}^0\).
Now consider first element of the following vector,
we get:
where \(R(m_0,n_0)=\big (\widehat{{\varvec{\xi }}}-{\varvec{\xi }}^0\big )^\top \begin{bmatrix} m_0\\ m_0^2\\ n_0\\ n_0^2\\ m_0n_0 \end{bmatrix}\), and second element of the above amplitude vector can be written as:
Now let us look at the first and the last term of Eq. (28) and putting value of \(y(m_0,n_0)\) from the model (1),
The above result has been obtained from a famous number theory conjecture by Montgomery (1994).
In second term of (28), \(R(m_0,n_0)\) is a sum of five terms, so now consider first term of
\(\frac{2}{{MN}}\displaystyle \sum _{m_0=1}^M\sum _{n_0=1}^{N}y(m_0,n_0)\sin {\breve{\phi }}R(m_0,n_0)\),
So, finding the asymptotic distribution of (28) boils down to finding asymptotic distribution of
which is further asymptotically equivalent to:
Asymptotic normality of amplitude estimators is thus proved by Eq. (31). Now we need to derive the expression of their asymptotic variances.
After lengthy calculations, we get the asymptotic variance of \({\widehat{A}}\) as follows:
Similarly by calculating other terms too, we get complete variance co-variance matrix \({\varvec{\Sigma }}\) as mentioned in Theorem 2.
Hence the result. \(\square \)
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Shukla, A., Grover, R., Kundu, D. et al. A computationally efficient algorithm to estimate the parameters of a two-dimensional chirp model with the product term. Multidim Syst Sign Process 34, 633–655 (2023). https://doi.org/10.1007/s11045-023-00879-7
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11045-023-00879-7