1 Introduction

The problem of estimation of two-dimensional (2-D) sinusoidal signals is of importance in a wide range of applications such as analysis of geophysical data [5], image restoration [2], array processing [9], radio astronomy [23], synthetic aperture radar imaging [17], nuclear magnetic resonance spectroscopy [16], medical imaging [18], wireless communications [24], health assessment of living trees [3], source localisation [1], to name a few.

Due to its extensive applicability, numerous techniques have been proposed for the parameter estimation of this signal model. Some of the prominent works are by Bansal et al. [4], Chen et al. [6], Clark and Scharf [7], Cohen and Francos [8], Francos et al. [10], Hua [12], Kundu and Gupta [13], Kundu and Nandi [14], Prasad et al. [21], Rao et al. [22], Zhang and Mandrekar [26] and the list does not end here. A more thorough overview of references in this area can be found in Peng et al. [19].

This paper addresses the problem of parameter estimation of a 2-D model, mathematically expressed as follows:

$$\begin{aligned} \begin{aligned} y(m,n)&= \sum _{k=1}^{p}\{A_k^0 \cos (\mu _k^0 m + \lambda _k^0 n) + B_k^0 \sin (\mu _k^0 m + \lambda _k^0 n) \} \\&\quad + X(m,n);\ m = 1, \ldots , M;\ n = 1, \ldots , N. \end{aligned}\end{aligned}$$
(1)

Here, y(mn) is the observed signal characterised by amplitude parameters, \(A_k^0\)s and \(B_k^0\)s, and frequency parameters \(\mu _k^0\)s and \(\lambda _k^0\)s. The random error component X(mn) is a 2-D sequence of independently and identically distributed random variables (i.i.d.) with mean 0 and variance \(\sigma ^2\). The fundamental problem here is to estimate the nonlinear parameters and the frequencies \(\mu _k^0\)s and \(\lambda _k^0\)s from a finite set of observations of length \(M\times N\) corrupted with additive noise.

The objective is to introduce an algorithm that is computationally feasible to implement in practice as well as one that provides statistically optimal estimators. In this paper, we show that the accuracy of the proposed estimators is as good as that of the usual least squares estimators (LSEs). This effectiveness of the algorithm is demonstrated through extensive numerical simulations along with an in-depth theoretical analysis.

It is important to note that the usual least squares estimator is one of the most conventional methods of parameter estimation of a postulated model embedded with noise. For a 2-D sinusoidal model with multiple components as described in Eq. (1) of the paper, the LSEs are obtained by minimising the error sum of squares:

$$\begin{aligned}\begin{aligned} \sum _{n=1}^{N}\sum _{m=1}^{M} \bigg (y(m,n) - \sum _{k=1}^{p}\{A_k^0 \cos (\mu _k^0 m + \lambda _k^0 n) + B_k^0 \sin (\mu _k^0 m + \lambda _k^0 n)\}\bigg )^2, \end{aligned}\end{aligned}$$

with respect to the unknown parameters. It has been proved that the LSEs of the parameters of this model are strongly consistent and asymptotically normally distributed. Moreover, if the errors are assumed to be normally distributed, the LSEs are asymptotically optimal. For reference, one may look at the paper by Kundu and Gupta [13].

For the purpose of numerical comparison, in Sect. 4, along with the proposed efficient estimators and LSEs, we also compute approximate least squares estimators (ALSEs). The ALSEs are a popular alternative to the LSEs and are obtained by maximising the following 2-D periodogram function defined as follows:

$$\begin{aligned} \frac{1}{MN} \bigg |\sum _{n=1}^{N}\sum _{m=1}^{M} y(m,n) e^{-i(\mu m + \lambda n)}\bigg |^2 \end{aligned}$$

with respect to the nonlinear parameters \(\mu \) and \(\lambda \). Once the nonlinear parameters are obtained, one can use simple linear regression to obtain the linear parameter estimates. Kundu and Nandi [14] show that the performance of the ALSEs is almost identical to that of the LSEs. Moreover, they provide theoretical proofs of their strong consistency and asymptotic equivalence to the LSEs.

The rest of the paper is organised as follows. In the next section, we first explain the proposed methodology for a one-component 2-D sinusoidal model. A sequential algorithm for a multiple sinusoidal model is provided subsequently. In Sect. 3, we provide the model assumptions and derive the large sample properties of the proposed estimators. The numerical results are presented in Sect. 4. The paper is concluded in Sect. 5, and thereafter the proofs of the asymptotic results are provided in the appendices.

2 Proposed Methodology

Here, we describe the proposed algorithm for the estimation of frequencies of the following 2-D sinusoidal model:

$$\begin{aligned} y(m,n) = A^0 \cos (\mu ^0 m + \lambda ^0 n) + B^0 \sin (\mu ^0 m + \lambda ^0 n) + X(m,n). \end{aligned}$$
(2)

Let us first fix \(n = 1\), then the model Eq. (2) reduces to the following:

$$\begin{aligned} y(m,1) = A^0 \cos (\mu ^0 m + \lambda ^0) + B^0 \sin (\mu ^0 m + \lambda ^0) + X(m,1). \end{aligned}$$

Using elementary trigonometric formulae, the above equation can be rewritten as:

$$\begin{aligned} y(m,1) = A^0(1) \cos (\mu ^0 m) + B^0(1) \sin (\mu ^0 m) + X(m,1), \end{aligned}$$

where \(A^0(1)\) and \(B^0(1)\) are functions of \(A^0\), \(B^0\) and \(\lambda ^0\). It is worth noting that the reduced model equation is that of a one-dimensional sinusoidal model with amplitudes \(A^0(1)\) and \(B^0(1)\) and frequency parameter \(\mu ^0\). In general, we fix \(n = n_0\), \(n_0 \in \{1, \ldots , N\}\) and we have the corresponding 1-D model equation:

$$\begin{aligned} y(m,n_0) = A^0(n_0) \cos (\mu ^0 m) + B^0(n_0) \sin (\mu ^0 m) + X(m,n_0). \end{aligned}$$

Thus, for each \(n_0\), we have a 1-D sinusoidal model with different amplitudes but same frequency parameter. We can now estimate this frequency parameter by fitting N 1-D models to the corresponding columns of the data matrix. In the absence of a distributional assumption, the most natural choice of estimation method seems to be the least squares method. Thus, the estimator of \(\mu ^0\) can be obtained by minimising the following function:

$$\begin{aligned}\begin{aligned} Q_M(A(n_0), B(n_0), \mu ) = \sum _{m=1}^{M} \bigg (y(m,n_0) - A(n_0) \cos (\mu m) - B(n_0) \sin (\mu m)\bigg )^2 \end{aligned}\end{aligned}$$

for each \(n_0 \in \{1, \ldots , N\}\). Note that this is a 1-D minimization problem as the linear parameters \(A(n_0)\) and \(B(n_0)\) can first be separated out using linear regression. This brings down the problem to minimising the reduced functions:

$$\begin{aligned}\begin{aligned} R_M(\mu , n_0)&= Q_M({\hat{A}}(n_0), {\hat{B}}(n_0), \mu )\\&= \sum _{m=1}^{M} \bigg (y(m,n_0) - {\hat{A}}(n_0) \cos (\mu m) - {\hat{B}}(n_0) \sin (\mu m)\bigg )^2 \end{aligned}\end{aligned}$$

with respect to \(\mu \). However, this process involves solving N such problems and since the underlying errors are assumed to independently and identically distributed; instead of minimising N different objective functions, we propose to minimise the sum of these objective functions, expressed as follows:

$$\begin{aligned}\begin{aligned} R_{MN}^{(1)}(\mu ) = \sum _{n_0=1}^{N} \bigg \{\sum _{m=1}^{M} \bigg (y(m,n_0) - {\hat{A}}(n_0) \cos (\mu m) - {\hat{B}}(n_0) \sin (\mu m)\bigg )^2\bigg \}, \end{aligned}\end{aligned}$$

with respect to \(\mu \) and get an estimate \({\hat{\mu }}\) of \(\mu ^0\). A similar estimate of \(\lambda ^0\), say \({\hat{\lambda }}\), can be obtained by minimising the following function with respect to \(\lambda \):

$$\begin{aligned}\begin{aligned} R_{MN}^{(2)}(\lambda ) = \sum _{m_0=1}^{M} \bigg \{\sum _{n=1}^{N} \bigg (y(m_0,n) - {\hat{A}}(m_0) \cos (\lambda n) - {\hat{B}}(m_0) \sin (\lambda n)\bigg )^2\bigg \}. \end{aligned}\end{aligned}$$

It is important to note that the optimisation problem is nonlinear and to solve such a problem an iterative algorithm has to be employed. We use Nelder–Mead algorithm to optimise this function for our simulations. The function has the problem of several local minima. Therefore, for the algorithm to converge to global minimum rather than a local minimum, we need precise initial values. To find these initial guesses in practice when the true values are unknown, we minimise this function at the Fourier frequencies, that is at the points \(\frac{\pi j}{M}; j = 1, \ldots , M.\) or \(\frac{\pi k}{N}; k = 1, \ldots , N.\)

The proposed method is not only computationally efficient than the usual least squares method but is also conceptually simple. This method can be easily extended to a more general model with multiple sinusoids as described in (1). The idea is based on modifying the sequential method proposed by Prasad et al. [21] and uses the fact that any two sinusoidal components present in the model are orthogonal to each other. We describe the sequential algorithm below.

figure a

3 Asymptotic Results

In this section, we investigate large-sample properties of the proposed estimators under the following assumption on the structure of the error component, X(mn):

Assumption 1

X(m,n) is a double array sequence of i.i.d. random variables with mean 0 and finite variance \(\sigma ^2 > 0.\)

Also, the true parameter vector \((A^0, B^0, \mu ^0, \lambda ^0)\) satisfies the following assumption:

Assumption 2

The true parameter vector \((A^0, B^0, \mu ^0, \lambda ^0)\) is an interior point of the parameter space \(\Theta = [-K,K] \times [-K,K] \times [0, \pi ] \times [0, \pi ]\) and \({A^0}^2 + {B^0}^2 > 0\). Here \(K >0\) is any real number.

The results are stated in the following theorems.

Theorem 1

The proposed frequencies’ estimators are strongly consistent if the assumptions 1 and 2 hold true, that is,

  1. (a)

    \({\hat{\mu }} \xrightarrow {a.s.} \mu ^0\) as \(M \rightarrow \infty \),

  2. (b)

    \({\hat{\lambda }} \xrightarrow {a.s.} \lambda ^0\) as \(N \rightarrow \infty \).

Proof

See Appendix A. \(\square \)

Theorem 2

The proposed frequencies’ estimators are asymptotically normally distributed with mean 0 and variance \(\frac{6 \sigma ^2}{{A^0}^2 + {B^0}^2}\) if the assumptions 1 and 2 hold true, that is,

  1. (a)

    \(M^{3/2}N^{1/2}({\hat{\mu }} - \mu ^0) \xrightarrow {d} {\mathcal {N}}(0,\frac{6 \sigma ^2}{{A^0}^2 + {B^0}^2}),\)

  2. (b)

    \(M^{1/2}N^{3/2}({\hat{\lambda }} - \lambda ^0) \xrightarrow {d} {\mathcal {N}}(0,\frac{6 \sigma ^2}{{A^0}^2 + {B^0}^2})\),

as \(min\{M,N\}\) \(\rightarrow \infty \).

Proof

See Appendix B. \(\square \)

It is evident that the proposed method yields frequency estimators with cubic convergence rates which is extremely fast. Also, the accuracy of these estimators is inversely proportional to the sum of squares of the amplitudes, that is, \({A^0}^2 + {B^0}^2\), which is not very surprising. As a matter of fact, the asymptotic distribution of the proposed estimators coincides with that of the usual LSEs. Moreover, if the errors are assumed to be normally distributed, the derived asymptotic variances are same as the Cramer–Rao lower bounds (CRLBs).

We now prove the strong consistency and derive the asymptotic distribution of the proposed sequential estimators of the frequencies of multiple component sinusoidal model. These asymptotic properties are derived under Assumption 1 on the error component and the following assumptions on the parameters of the model:

Assumption 3

The true parameter vector for each component, that is, \((A_k^0, B_k^0, \mu _k^0, \lambda _k^0)\) is an interior point of the parameter space \(\Theta \) \(\forall \ k = 1, \ldots , p\). Also the frequencies are such that \(\mu _i^0 \ne \mu _j^0\) and \(\lambda _i^0 \ne \lambda _j^0\) \(\forall \ i \ne j, i, j = 1, \ldots , p\).

Assumption 4

The linear parameters \(A_k^0\)s and \(B_k^0\)s satisfy the following relationship:

$$\begin{aligned} K^2> {A_1^0}^2 + {B_1^0}^2> {A_2^0}^2 + {B_2^0}^2> \ldots > {A_p^0}^2 + {B_p^0}^2. \end{aligned}$$

The results are stated in the subsequent theorems.

Theorem 3

Under the assumptions 1, 3 and 4, the frequency estimates \({\hat{\mu }}_k\) and \({\hat{\lambda }}_k\) are strongly consistent estimators of \(\mu _k^0\) and \(\lambda _k^0\), respectively, that is \(\forall \ k = 1, \ldots , p,\)

  1. (a)

    \({\hat{\mu }}_k \xrightarrow {a.s.} \mu _k^0\) as \(M \rightarrow \infty \),

  2. (b)

    \({\hat{\lambda }}_k \xrightarrow {a.s.} \lambda _k^0\) as \(N \rightarrow \infty \).

Proof

See Appendix C. \(\square \)

Theorem 4

Under the assumptions 1, 3 and 4, the following results hold true:

\(\forall \) \(k = 1, \ldots , p\)

  1. (a)

    \({\hat{A}}_k \xrightarrow {a.s.} A_k^0 \) as \(\min \{M, N\} \rightarrow \infty \),

  2. (b)

    \({\hat{B}}_k \xrightarrow {a.s.} B_k^0 \) as \(\min \{M, N\} \rightarrow \infty \),

\(\forall \) \(k > p\)

  1. (a)

    \({\hat{A}}_k \xrightarrow {a.s.} 0 \) as \(\min \{M, N\} \rightarrow \infty \),

  2. (b)

    \({\hat{B}}_k \xrightarrow {a.s.} 0 \) as \(\min \{M, N\} \rightarrow \infty \).

Proof

See Appendix C. \(\square \)

Theorem 5

The proposed sequential estimators of the frequencies are asymptotically normally distributed if the assumptions 1, 3 and 4 hold true, that is,

  1. (a)

    \(M^{3/2}N^{1/2}({\hat{\mu }}_k - \mu _k^0) \xrightarrow {d} {\mathcal {N}}(0,\frac{6 \sigma ^2}{{A_k^0}^2 + {B_k^0}^2}),\)

  2. (b)

    \(M^{1/2}N^{3/2}({\hat{\lambda }}_k - \lambda _k^0) \xrightarrow {d} {\mathcal {N}}(0,\frac{6 \sigma ^2}{{A_k^0}^2 + {B_k^0}^2})\),

as \(\min \{M,N\}\) \(\rightarrow \infty \).

Proof

See Appendix D. \(\square \)

It is evident that like the estimators for the one component model, the sequential estimators are strongly consistent and asymptotically normally distributed as well. The algorithm produces estimators of \(\mu _k^0\) and \(\lambda _k^0\) with convergence rates \(O_p(M^{-3/2}N^{-1/2})\) and \(O_p(N^{-3/2}M^{-1/2})\), respectively, same as those of the usual LSEs. Another interesting property of the proposed algorithm is that depicted in Theorem 4, when the algorithm is continued beyond p number of components, the corresponding amplitudes converge to zero. This feature can help one to estimate the number of components in practice. It is important to note that these asymptotic results can be further exploited to construct confidence intervals as well as to devise testing of hypotheses and problems of practical significance.

4 Numerical Experiments

In this section, we present the simulation results. These simulation experiments were designed to assess the performance of the proposed methodology in comparison with the conventional least squares estimation method. The data matrix is generated using the following one-component 2-D sinusoidal model:

$$\begin{aligned}\begin{aligned} y(m,n) =&2 \cos (1.5 m + 0.75 n) + 3 \sin (1.5 m + 0.75 n) \\&+ X(m,n);\ m = 1, \ldots , M, n = 1, \ldots , N. \end{aligned}\end{aligned}$$

The error random variables are generated from Gaussian distribution with mean 0 and variance \(\sigma ^2\). For the first set of experiments, we fix the error variance at \(\sigma ^2 = 0.01\) and vary the sample size from 20 to 200. We compute the estimates of \(\lambda ^0\) and \(\mu ^0\) for 1000 independent replications using different error sequences for all cases. Figure 1 illustrates the MSEs of the resulting parameter estimates along with the derived asymptotic variances.Footnote 1 In the next set of experiments, we fix the sample size. For a 100 \(\times \) 100 data matrix, we compute the MSEs of the nonlinear parameter estimates for varying signal-to-noise ratio. The results are shown in Fig. 2. In these experiments, the initial values are taken as the true values. From the figure, it is apparent that the MSEs of the proposed estimates are well-matched with those of LSEs and lower than that of ALSEs. These almost coincide with the asymptotic variances for all sample sizes and for SNR above 10 dB. These results validate the theoretical claim that the accuracy of the efficient estimators is as good as that of LSEs. The results in all the figures are reported in the log scale.

Next, we evaluate the performance of the proposed sequential estimators for a multiple component model. We generate the data from a 2-D sinusoidal model with two components using the following model equation:

Fig. 1
figure 1

MSEs of the efficient estimators along with those of LSEs and ALSEs and the corresponding theoretical asymptotic variances versus the data size (\(M = N\))

Fig. 2
figure 2

MSEs of the efficient estimators along with those of LSEs and ALSEs and the corresponding theoretical asymptotic variances versus SNR

Fig. 3
figure 3

MSEs of the sequential efficient estimators along with those of sequential LSEs and sequential ALSEs and the corresponding theoretical asymptotic variances versus the data size (\(M = N\))

Fig. 4
figure 4

MSEs of the sequential efficient estimators along with those of sequential LSEs and sequential ALSEs and the corresponding theoretical asymptotic variances versus the data size (\(M = N\))

Fig. 5
figure 5

MSEs of the sequential efficient estimators along with those of sequential LSEs and sequential ALSEs and the corresponding theoretical asymptotic variances versus SNR

Fig. 6
figure 6

MSEs of the sequential efficient estimators along with those of sequential LSEs and sequential ALSEs and the corresponding theoretical asymptotic variances versus SNR

Fig. 7
figure 7

Comparison of computational complexity involved in the proposed efficient method and the least squares estimation method

Fig. 8
figure 8

Comparison of time taken for computation of proposed efficient estimators and the least squares estimators

$$\begin{aligned}\begin{aligned} y(m,n)&= 5 \cos (1.5 m + 0.75 n) + 4 \sin (1.5 m + 0.75 n) + 2 \cos (1.0 m \\&\quad + 0.25 n) + 3 \sin (1.0 m + 0.25 n) \\&\quad + X(m,n);\ m = 1, \ldots , M, n = 1, \ldots , N. \end{aligned}\end{aligned}$$

The same error structure and simulation set-up as described above for the one component model were used. Moreover, the accuracy of the proposed estimators is compared with that of the sequential LSEs and sequential ALSEs. Figure 3 shows the MSEs of the proposed estimators, sequential LSEs and sequential ALSEs of the first component parameters with respect to varying sample sizes. We also plot the CRLBs to benchmark the performance of these estimators. The MSEs of the second component parameter estimates are shown in Fig. 4. In Figs. 5 and 6 these results are investigated for varying SNRs. It can be seen that the performance of the proposed sequential estimators is at par with that of sequential LSEs and sequential ALSEs. Moreover, for the second component parameters, the MSEs of all the three estimators coincide with the corresponding CRLBs for increasing M(N). For varying SNR, the results of all the estimators concur with CRLBs for both the component parameters.

In order to exemplify the advantage of the proposed estimators, we compare the computational complexity of both methods, the proposed and the least squares method. The complexity is measured in terms of number of function evaluations needed to find the initial values of the parameter estimates of a one component model.Footnote 2 Figure 7 demonstrates the complexity of both the estimators as the sample size, \(M = N\) varies. It is evident that there is a significant difference between the computational involvement for the two types of estimates under consideration. This implies that calculating the proposed estimators is much faster than calculating the LSEs. Figure 8 corroborates the faster implementation of the proposed algorithm as compared to the traditional least squares estimation method. Note that the number of function evaluations needed to compute ALSEs is same as that required for LSEs and hence the ALSEs are omitted from this comparison.

5 Conclusion

In this paper, we have proposed a novel method of estimation of the frequencies of a 2-D sinusoidal model. Although the LSEs have optimal statistical properties, due to considerable computation burden finding them is practically infeasible. The proposed method reduces this burden of computability to a great extent and provides efficient estimators with accuracy on an equal footing with that of the LSEs. Statistical analysis of the proposed estimators shows that they are strongly consistent and asymptotically equivalent to the corresponding LSEs. Simulation experiments demonstrate the ability of the approach to estimate the frequencies accurately. These results are presented in comparison with the LSEs as well as in comparison to the CRLBs and the performance is at par with both.