Abstract
We study maximum-likelihood-type estimation for diffusion processes when the coefficients are nonrandom and observations occur in nonsynchronous manner. The problem of nonsynchronous observations is important when we consider the analysis of high-frequency data in a financial market. Constructing a quasi-likelihood function to define the estimator, we adaptively estimate the parameter for the diffusion part and the drift part. We consider the asymptotic theory when the terminal time point \(T_n\) and the observation frequency goes to infinity, and show the consistency and the asymptotic normality of the estimator. Moreover, we show local asymptotic normality for the statistical model, and asymptotic efficiency of the estimator as a consequence. To show the asymptotic properties of the maximum-likelihood-type estimator, we need to control the asymptotic behaviors of some functionals of the sampling scheme. Though it is difficult to directly control those in general, we study tractable sufficient conditions when the sampling scheme is generated by mixing processes.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Given a probability space \((\Omega , {\mathcal {F}}, P)\) with a right-continuous filtration \(\textbf{F}=\{{\mathcal {F}}_t\}_{t\ge 0}\), let \(X^{(\alpha )}=\{X_t^{(\alpha )}\}_{t\ge 0}=\{(X^{(\alpha ),1}_t,X^{(\alpha ),2}_t)\}_{t\ge 0}\) be a two-dimensional \(\textbf{F}\)-adapted process satisfying the following stochastic differential equation:
where \(x_0\in \mathbb {R}^2\), \(\{W_t\}_{0\le t\le T}\) is a two-dimensional standard \(\textbf{F}\)-Wiener process, \(\{\mu _t(\theta )\}_{t\ge 0}\) and \(\{b_t(\sigma )\}_{t\ge 0}\) are deterministic functions with values in \({\mathbb {R}}^{2}\) and \({\mathbb {R}}^{2\times 2}\), respectively, \(\alpha =(\sigma ,\theta )\), \(\sigma \in \Theta _1\), \(\theta \in \Theta _2\), and \(\Theta _1\) and \(\Theta _2\) are bounded open subsets of \({\mathbb {R}}^{d_1}\) and \(\mathbb {R}^{d_2}\), respectively. Let \(\alpha _0=(\sigma _0,\theta _0)\in \Theta _1\times \Theta _2\) be the true value, and let \(X_t=(X_t^1,X_t^2)=X_t^{(\alpha _0)}\). We consider estimation of \(\alpha _0\) when X is observed with nonsynchronous manner, that is, observation times of \(X^1\) and \(X^2\) are different to each other.
The problem of nonsynchronous observations appears in the analysis of high-frequency financial data. If we analyze the intra-day stock price data, we observe stock prices when a new transaction or a new order arrives. Then, the observation times are different for different stocks, and hence, we cannot avoid the problem of nonsynchronous observations. Statistical analysis with such data is much more complicated compared to the analysis with synchronous data. Parametric estimation for diffusion processes with synchronous and equidistant observations has been analyzed through quasi-maximum-likelihood methods in Florens-Zmirou (1989), Yoshida (1992, 2011), Kessler (1997), and Uchida and Yoshida (2012). Related to the estimation problem for nonsynchronously observed diffusion processes, estimators for the quadratic covariation have been actively studied. Hayashi and Yoshida (2005, 2008, 2011) and Malliavin and Mancino (2002, 2009) have independently constructed consistent estimators under nonsynchronous observations. There are also studies of covariation estimation under the simultaneous presence of microstructure noise and nonsynchronous observations (Barndorff-Nielsen et al., 2011; Bibinger et al., 2014; Christensen et al., 2010, and so on). For parametric estimation with nonsynchronous observations, Ogihara and Yoshida (2014) have constructed maximum-likelihood-type and Bayes-type estimators and have shown the consistency and the asymptotic mixed normality of the estimators when the terminal time point \(T_n\) is fixed and the observation frequency goes to infinity. Ogihara (2015) have shown local asymptotic mixed normality for the model in Ogihara and Yoshida (2014), and the maximum-likelihood-type and Bayes-type estimators have been shown to be asymptotically efficient. On the other hand, we need to consider asymptotic theory that the terminal time point \(T_n\) goes to infinity to consistently estimate the parameter \(\theta \) in the drift term. To the best of the author’s knowledge, there are no studies of the asymptotic theory of parametric estimation for nonsynchronously observed diffusion processes when \(T_n\rightarrow \infty \).
In this work, we consider the asymptotic theory for nonsynchronously observed diffusion processes when \(T_n\rightarrow \infty \), and construct maximum-likelihood-type estimators for the parameter \(\sigma \) in the diffusion part and the parameter \(\theta \) in the drift part. We show the consistency and the asymptotic normality of the estimators. Moreover, we show local asymptotic normality of the statistical model, and we obtain asymptotic efficiency of our estimator as a consequence. Our estimator is constructed based on the quasi-likelihood function that is similarly defined to the one in Ogihara and Yoshida (2014), though we need some modification to deal with the drift part. To investigate asymptotic theory for the maximum-likelihood-type estimator, we need to specify the limit of the quasi-likelihood function. Then, we need to assume some conditions for the asymptotic behavior of the sampling scheme. In Ogihara and Yoshida (2014), for a matrix
generated by the sampling scheme, the existence of the probability limit of \(n^{-1}\textrm{tr}((GG^\top )^p) \ (p\in \mathbb {Z}_+)\) is required, where \((S_i^{n,l})_i\) are observation times of \(X^l\) and \(\top \) denotes transpose of a vector or a matrix. Since we consider the different asymptotics, the asymptotic behavior of the quasi-likelihood function is different from that in Ogihara and Yoshida (2014). We also need to consider estimation for the drift parameter \(\theta \). Then, we need other assumptions for the asymptotic behavior of the sampling scheme [Assumption (A5)]. Though these conditions for the sampling scheme are difficult to check directly, we study tractable sufficient conditions in Sect. 2.4.
The rest of this paper is organized as follows. In Sect. 2, we introduce our model settings and the assumptions for main results. Our estimator is constructed in Sect. 2.1, and the asymptotic normality of the estimator is given in Sect. 2.2. Section 2.3 deals with local asymptotic normality of our model and asymptotic efficiency of the estimator. Tractable sufficient conditions for the assumptions of the sampling scheme are given in Sect. 2.4. Section 3 contains the proofs of main results. Preliminary results are collected in Sect. 3.1. Section 3.2 is for the consistency of the estimator for \(\sigma \), Sect. 3.3 is for the asymptotic normality of the estimator for \(\sigma \), Sect. 3.4 is for the consistency of the estimator for \(\theta \), and Sect. 3.5 is for the asymptotic normality of the estimator for \(\theta \). Other proofs are collected in Sect. 3.6.
2 Main results
2.1 Setting and parameter estimation
Let \(\mathbb {N}\) be the set of all positive integers. For \(l\in \{1,2\}\), let the observation times \(\{S_i^{n,l}\}_{i=0}^{M_l}\) be strictly increasing random times with respect to i, and satisfy \(S_0^{n,l}=0\) and \(S_{M_l}^{n,l}=nh_n\), where \(M_l\) is a random positive integer depending on n and \((h_n)_{n=1}^\infty \) is a sequence of positive numbers satisfying
as \(n\rightarrow \infty \) for some \(\epsilon _0>0\). Intuitively, n is of the order of the number of observations and \(h_n\) is of the order of the length of the observation intervals. More precise assumptions of observation times are given in (A2), (A4), and (A5) later. We assume that \(\{S_i^{n,l}\}_{0\le i\le M_l, l=1,2}\) is independent of \({\mathcal {F}}_T\), and its distribution does not depend on \(\alpha \). We consider nonsynchronous observations of X, that is, we observe \(\{S_i^{n,l}\}_{0\le i\le M_l, l=1,2}\) and \(\{X^l_{S^{n,l}_i}\}_{0\le i\le M_l, l=1,2}\). In particular, we consider the nonendogenous observation times.
We denote by \(\Vert \cdot \Vert \) the operator norm with respect to the Euclidean norm for a matrix. We often regard a p-dimensional vector v as a \(p\times 1\) matrix. For \(j\in \mathbb {N}\), we denote \(\partial _z=\frac{\partial }{\partial z}\) for a variable \(z\in \mathbb {R}^j\), and denote \(\partial _z^l=(\partial _{z_{i_1}}\cdots \partial _{z_{i_l}})_{i_1,\ldots , i_l=1}^j\) for \(l\in \mathbb {N}\). For functions f and g, we often use shorthand notation \(\partial _zf\partial _zg=(\partial _zf(\partial _zg)^\top +\partial _zg (\partial _zf)^\top )/2\). For a set A in a topological space, let \(\textrm{clos}(A)\) denote the closure of A. For a matrix A, \([A]_{ij}\) denotes its (i, j) element. For a vector \(v=(v_j)_{j=1}^K\), we denote \([v]_j=v_j\), and \(\textrm{diag}(v)\) denotes a \(K\times K\) diagonal matrix with elements \([{\textrm{diag}}(v)]_{jj}=v_j\).
Let \(M=M_1+M_2\). For \(1\le i\le M\), let
For a two-dimensional stochastic process \((U_t)_{t\ge 0}=((U_t^1,U_t^2))_{t\ge 0}\), let \(\Delta _i^l U=U^l_{S^{n,l}_i}-U^l_{S^{n,l}_{i-1}}\), and let \(\Delta ^l U=(\Delta _i^l U)_{1\le i\le M_l}\) and \(\Delta _i U=\Delta _{\varphi (i)}^{\psi (i)} U\) for \(1\le i\le M\). Let \(\Delta U=((\Delta ^1 U)^\top , (\Delta ^2 U)^\top )^\top \). Let \(|K|=b-a\) for an interval \(K=(a,b]\). Let \(I_i^l=(S_{i-1}^{n,l},S_i^{n,l}]\) for \(1\le i\le M_l\), and let \(I_i=I_{\varphi (i)}^{\psi (i)}\) for \(1\le i\le M\). We denote a unit matrix of size k by \({\mathcal {E}}_k\).
Let \(\tilde{\Sigma }_i^l(\sigma )=\int _{I_i^l}[b_tb_t^\top (\sigma )]_{ll}{\textrm{d}}t\) and \(\tilde{\Sigma }_{i,j}^{1,2}(\sigma )=\int _{I_i^1\cap I_j^2}[b_tb_t^\top (\sigma )]_{12}{\textrm{d}}t\), and let \(\tilde{\Sigma }_i=\tilde{\Sigma }_{\varphi (i)}^{\psi (i)}\) for \(1\le i\le M\). By setting \(\tilde{\mathcal {D}}=\textrm{diag}((\tilde{\Sigma }_i)_{1\le i\le M})\)
we can calculate the covariance matrix of \(\Delta X\) as
As we will see later, we can ignore the term related to \(\mu _t(\theta )\) (drift term) when we consider estimation of \(\sigma \), because this term converges to zero very fast. Therefore, we first construct an estimator for \(\sigma \), and then construct an estimator for \(\theta \). Such adaptive estimation can speed up the calculation.
We define the quasi-likelihood function \(H_n^1(\sigma )\) for \(\sigma \) as follows:
Then, the maximum-likelihood-type estimator for \(\sigma \) is defined by
We consider estimation for \(\theta \) next. Let \(V(\theta )=(V_t(\theta ))_{t\ge 0}\) be a two-dimensional stochastic process defined by \(V_t(\theta )=(\int _0^t\mu ^1_s(\theta )^\top {\textrm{d}}s,\int _0^t\mu ^2_s(\theta )^\top {\textrm{d}}s)^\top \). Let \(\bar{X}(\theta )=\Delta X-\Delta V(\theta )\). We define the quasi-likelihood function \(H_n^2(\theta )\) for \(\theta \) as follows:
Then, the maximum-likelihood-type estimator for \(\theta \) is defined by
The quasi-(log-)likelihood function \(H_n^1\) is defined in the same way as that in Ogihara and Yoshida (2014). Since \(\Delta X\) follows normal distribution, we can construct such a Gaussian quasi-likelihood function even for the nonsynchronous data. When the coefficients are random, though the distribution of \(\Delta X\) is not Gaussian, such Gaussian-type quasi-likelihood function is still valid due to the local Gaussian property of diffusion processes. The Gaussian mean that comes from the drift part is ignored when we construct the quasi-likelihood \(H_n^1\). When we estimate the parameter \(\theta \) for the drift part, we subtract the mean in \(\bar{X}(\theta )\) to construct the quasi-likelihood function \(H_n^2\). Since the effect of the drift term on the estimation of \(\sigma \) is small, it works well to estimate \(\sigma \) in this way and then plug in \(\hat{\sigma }_n\) to \(S_n\) to construct the estimator for \(\theta \). Thus, we can speed up the calculation by separating the estimation for \(\sigma \) and \(\theta \).
Remark 2.1
\(H_n^1(\sigma )\) and \(H_n^2(\theta )\) are well defined only if \(\det S_n(\sigma )>0\) and \(\det S_n(\hat{\sigma }_n)>0\), respectively. For the covariance matrix \(S_n\) of nonsynchronous observations \(\Delta X\), it is not trivial to check these conditions. Proposition 1 in Section 2 of Ogihara and Yoshida (2014) shows that these conditions are satisfied if \(b_t(\sigma )\) is continuous on \([0,\infty )\times \textrm{clos}(\Theta _1)\) and \(\inf _{t,\sigma } \det (b_tb_t^\top (\sigma ))>0\). We assume such conditions in our setting (Assumption (A1) in Sect. 2.2).
Remark 2.2
As seen in Ogihara and Yoshida (2014), the quasi-likelihood analysis for nonsynchronously observed diffusion processes becomes much more complicated compared to synchronous observations. In this work, estimation for the drift parameter \(\theta \) is added, and hence, we consider nonrandom drift and diffusion coefficients to avoid overcomplication. For general diffusion processes with the random drift and diffusion coefficients, we need to set predictable coefficients to use the martingale theory. However, the quasi-likelihood function loses a Markov property with nonsynchronous observations and the coefficients in the quasi-likelihood function contain randomness of future time. Then, we need to approximate the coefficients by predictable functions. This operation is particularly complicated. Moreover, approximating the true likelihood function by the quasi-likelihood function is much more difficult problem when we show local asymptotic normality and asymptotic efficiency of the estimators. Therefore, we left asymptotic theory under general random drift and diffusion coefficients as a future work.
2.2 Asymptotic normality of the estimator
In this section, we state the assumptions of our main results, and state the asymptotic normality of the estimator.
For \(m\in \mathbb {N}\), an open subset \(U\subset \mathbb {R}^m\) is said to admit Sobolev’s inequality if, for any \(p>m\), there exists a positive constant C depending on U and p, such that \(\sup _{x\in U}|u(x)|\le C\sum _{k=0,1}(\int |\partial _x^k u(x)|^p{{\textrm{d}}x})^{1/p}\) for any \(u\in C^1(U)\). This is the case when U has a Lipschitz boundary. We assume that \(\Theta \), \(\Theta _1\), and \(\Theta _2\) admit Sobolev’s inequality.
Let \(\Sigma _t(\sigma )=b_tb_t^\top (\sigma )\), and let
Let \(\rho _{t,0}=\rho _t(\sigma _0)\).
Assumption (A1). There exists a positive constant \(c_1\), such that \(c_1\mathcal {E}_2 \le \Sigma _t(\sigma )\) for any \(t\in [0,\infty )\) and \(\sigma \in \Theta _1\). For \(k\in \{0,1,2,3,4\}\), \(\partial _\theta ^k\mu _t(\theta )\) and \(\partial _\sigma ^kb_t(\sigma )\) exist and are continuous with respect to \((t,\sigma ,\theta )\) on \([0,\infty )\times \textrm{clos}(\Theta _1)\times \textrm{clos}(\Theta _2)\). For any \(\epsilon >0\), there exist \(\delta >0\) and \(K>0\), such that
for any \(k\in \{0,1,2,3,4\}\), \(\sigma \in \Theta _1\), \(\theta \in \Theta _2\), and \(t,s\ge 0\) satisfying \(|t-s|<\delta \). Let \(r_n=\max _{i,l}|I_i^l|\).
Assumption (A2). \(r_n\overset{P}{\rightarrow }0\) as \(n\rightarrow \infty \).
Assumption (A3). For any \(l\in \{1,2\}\), \(i_1\in \mathbb {Z}_+\), \(i_2\in \{0,1\}\), \(i_3\in \{0,1,2,3,4\}\), \(k_1,k_2\in \{0,1,2\}\) satisfying \(k_1+k_2=2\), and any polynomial function \(F(x_1,\ldots , x_{14})\) of degree equal to or less than 6, there exist continuous functions \(\Phi _{i_1,i_2}^{1,F}(\sigma )\), \(\Phi _{l,i_3}^2(\sigma )\) and \(\Phi ^{3,k_1,k_2}_{i_1,i_3}(\theta )\) on \(\textrm{clos}(\Theta _1)\) and \(\textrm{clos}(\Theta _2)\), such that
as \(T\rightarrow \infty \) for \(\sigma \in \textrm{clos}(\Theta _1)\), \(\theta \in \textrm{clos}(\Theta _2)\), where \(\phi _{l,t}(\theta )=[\Sigma _t(\sigma _0)]_{ll}^{-1/2}(\mu _t^l(\theta )-\mu _t^l(\theta _0))\).
Assumption (A1) and the Ascoli–Arzelà theorem yield that the convergences in (A3) can be replaced by uniform convergence with respect to \(\sigma \) and \(\theta \) (the left-hand sides of the above equations become relatively compact, and then, any uniformly convergent subsequence converges to the right-hand sides due to the pointwise convergence assumptions). Assumption (A3) is satisfied if \(\mu _t(\theta )\) and \(b_t(\sigma )\) are independent of t, or are periodic functions with respect to t having a common period (when the period does not depend on \(\sigma \) nor \(\theta \)). Let \(\mathfrak {S}\) be the set of all partitions \((s_k)_{k=0}^\infty \) of \([0,\infty )\) satisfying \(\sup _{k\ge 1}|s_k-s_{k-1}|\le 1\) and \(\inf _{k\ge 1}|s_k-s_{k-1}|>0\). For \((s_k)_{k=0}^\infty \in \mathfrak {S}\), let \(M_{l,k}=\#\{i;\sup I_i^l\in (s_{k-1},s_k]\}\) and \(q_n=\max \{k;s_k\le nh_n\}\), and let \(\mathcal {E}_{(k)}^l\) be an \(M_l\times M_l\) matrix satisfying \([\mathcal {E}_{(k)}^l]_{ij}=1\) if \(i=j\) and \(\sup I_i^l\in (s_{k-1},s_k]\), and otherwise, \([\mathcal {E}_{(k)}^l]_{ij}=0\). Let
Assumption (A4). There exist positive constants \(a_0^1\) and \(a_0^2\), such that \(\{h_nM_{l,q_n+1}\}_{n=1}^\infty \) is P-tight and
for \(l\in \{1,2\}\) and any partition \((s_k)_{k=0}^\infty \in \mathfrak {S}\). Moreover, for any \(p\in \mathbb {N}\), there exists a nonnegative constant \(a_p^1\), such that
as \(n\rightarrow \infty \) for any partition \((s_k)_{k=0}^\infty \in \mathfrak {S}\). Let \(\mathfrak {I}_l=(|I_i^l|^{1/2})_{i=1}^{M_l}\).
Assumption (A5). For \(p\in \mathbb {Z}_+\), there exist nonnegative constants \(f_p^{1,1}\), \(f_p^{1,2}\), and \(f_p^{2,2}\), such that \(\{|\mathcal {E}_{(q_n+1)}^l\mathfrak {I}_l|\}_{n=1}^\infty \) is P-tight for \(l\in \{1,2\}\), and
as \(n\rightarrow \infty \) for any partition \((s_k)_{k=0}^\infty \in \mathfrak {S}\).
Assumption (A4) corresponds to [A3\('\)] in Ogihara and Yoshida (2014). The functionals in (A4) and (A5) appear in \(H_n^1\) and \(H_n^2\), and hence, we cannot specify the limits of \(H_n^1\) and \(H_n^2\) unless we assume existence of the limits of these functionals. It is difficult to directly check (A4) and (A5) for concrete statistical experiments with general sampling schemes. We study sufficient conditions for these conditions in Sect. 2.4.
Assumption (A6). The constant \(a_1^1\) in (A4) is positive, and there exist positive constants \(c_2\) and \(c_3\), such that
for any \(\sigma \in \textrm{clos}(\Theta _1)\) and \(\theta \in \textrm{clos}(\Theta _2)\).
Assumption (A6) is necessary to identify the parameter \(\sigma \) and \(\theta \) from the data. For \(p<q\)
by Lemma 3.3 later and Lemma A.1 in Ogihara (2018). Then, \(a_p^1\) is monotone non-increasing with respect to p. This implies that \(a_p^1=0\) for any \(p\in \mathbb {N}\) if \(a_1^1=0\). In this case, the non-diagonal components of the covariance matrix \(S_n\) are negligible in the limit. Then, we cannot consistently estimate the parameter in \(\rho _t(\sigma )\). This is why, we need the assumption \(a_1^1>0\) (see Proposition 3.9 and the following discussion to obtain the consistency).
Let \(\mathcal {A}(\rho )=\sum _{p=1}^\infty a_p^1\rho ^{2p}\) for \(\rho \in (-1,1)\). Then, (2.3) implies that \(\mathcal {A}(\rho )\) is finite. Moreover, (A5) yields
which implies \(f_p^{1,1}\le 1\). Similarly, we have \(f_p^{1,2}\le 1\) and \(f_p^{2,2}\le 1\). Let \(\partial _\sigma ^kB_{l,t,0}=\partial _\sigma ^kB_{l,t}(\sigma _0)\), and let
and let \(\Gamma _1=\lim _{T\rightarrow \infty }T^{-1}\int _0^T\gamma _{1,t}{\textrm{d}}t\), which exists under (A1), (A3), and (A4). Let
which exists under (A1), (A3), and (A5). Let \(T_n=nh_n\) and
Theorem 2.3
Assume (A1)–(A6). Then, \(\Gamma \) is positive definite, and
as \(n\rightarrow \infty \).
2.3 Local asymptotic normality
Let \(\alpha _0\in \Theta \), \(\Theta \subset \mathbb {R}^d\), and \(\{P_{\alpha ,n}\}_{\alpha \in \Theta }\) be a family of probability measures defined on a measurable space \(({\mathcal {X}}_n,\mathcal {A}_n)\) for \(n\in \mathbb {N}\), where \(\Theta \) is an open subset of \({\mathbb {R}}^d\). As usual, we shall refer to \({\textrm{d}}P_{\alpha _2,n}/{\textrm{d}}P_{\alpha _1,n}\) the derivative of the absolutely continuous component of the measure \(P_{\alpha _2,n}\) with respect to measure \(P_{\alpha _1,n}\) at the observation x as the likelihood ratio. The following definition of local asymptotic normality is Definition 2.1 in Chapter II of Ibragimov and Has’minskiĭ (1981).
Definition 2.4
A family \(P_{\alpha ,n}\) is called locally asymptotically normal (LAN) at point \(\alpha _0\in \Theta \) as \(n\rightarrow \infty \) if for some nondegenerate \(d \times d\) matrix \(\epsilon _n\) and any \(u\in \mathbb {R}^d\), the representation
in \(P_{\alpha _0,n}\)-probability as \(n\rightarrow \infty \), where
as \(n\rightarrow \infty \), and \(\mathcal {L}(\cdot |P_{\alpha ,n})\) denotes the distribution with respect to \(P_{\alpha ,n}\).
Let \(\Theta =\Theta _1\times \Theta _2\). For \(\alpha \in \Theta \), let \(P_{\alpha ,n}\) be the probability measure generated by the observations \(\{S_i^{n,l}\}_{i,l}\) and \(\{X_{S_i^{n,l}}^{(\alpha ),l}\}_{i,l}\).
Theorem 2.5
Assume (A1)–(A6). Then, \(\{P_{\alpha ,n}\}_{\alpha ,n}\) satisfies the LAN property at \(\alpha =\alpha _0\) with
The proof is left to Sect. 3.6. Theorem 11.2 in Chapter II of Ibragimov and Has’minskiĭ (1981) gives lower bounds of estimation errors for any regular estimator of parameters under the LAN property. Then, the optimal asymptotic variance of \(\epsilon _n^{-1}({U_n}-\alpha _0)\) for regular estimator \(U_n\) is \(\mathcal {E}_d\). We will show that \((\hat{\sigma }_n, \hat{\theta }_n)\) is regular in Remark 3.18. Therefore, Theorem 2.5 ensures that our estimator \((\hat{\sigma }_n,\hat{\theta }_n)\) is asymptotically efficient in this sense under the assumptions of the theorem.
2.4 Sufficient conditions for the assumptions
It is not easy to directly check Assumptions (A4) and (A5) for general random sampling schemes (even for a sampling scheme generated by simple Poisson processes given in Example 2.6). In this section, we study tractable sufficient conditions for these assumptions. The proofs of the results in this section are left to Sect. 3.6.
Let \(q>0\) and \(\mathcal {N}^{n,l}_t=\sum _{i=1}^{M_l}1_{\{S_i^{n,l}\le t\}}\). We consider the following conditions for the point process \(\mathcal {N}_t^{n,l}\).
- Assumption (B1-q).:
-
$$\begin{aligned} \sup _{n\ge 1}\max _{l\in \{1,2\}}\sup _{0\le t\le (n-1)h_n}E[(\mathcal {N}^{n,l}_{t+h_n}-\mathcal {N}^{n,l}_t)^q]<\infty . \end{aligned}$$
- Assumption (B2-q).:
-
$$\begin{aligned} \limsup _{u\rightarrow \infty } \sup _{n\ge 1}\max _{l\in \{1,2\}}\sup _{0\le t\le nh_n-uh_n} u^qP(\mathcal {N}^{n,l}_{t+uh_n}-\mathcal {N}^{n,l}_t=0)<\infty . \end{aligned}$$
Example 2.6
Let \((\bar{\mathcal {N}}_t^1, \bar{\mathcal {N}}_t^2)\) be two independent homogeneous Poisson processes with positive intensities \(\lambda _1\) and \(\lambda _2\), respectively, and \(\mathcal {N}^{n,l}_t=\bar{\mathcal {N}}_{h_n^{-1}t}^l\), that is, \(S_i^{n,l}=\inf \{t\ge 0; \bar{\mathcal {N}}_{h_n^{-1}t}^l\ge i\}\). Even in this simple case, it is not trivial to directly check (A4) and (A5). On the other hand, (B1-q) obviously holds for any \(q > 0\). Moreover, (B2-q) holds for any \(q > 0\), since
Then, by Corollary 2.12, we can check Assumptions (A2), (A4), and (A5) for this sampling scheme.
To give sufficient conditions for (A4) and (A5), we consider mixing properties of \(\mathcal {N}^{n,l}\). That is, we assume conditions for the following mixing coefficient \(\alpha _k^n\). Let
and let
Proposition 2.7
Assume that (B1-q) and (B2-q) hold and that
for any \(q>0\). Moreover, assume that there exist positive constants \(a_0^1\) and \(a_0^2,\) and a nonnegative constant \(a_p^1\) for \(p\in \mathbb {N},\) such that \(\{E[h_nM_{l,q_n+1}]\}_{n=1}^\infty \) is bounded and
as \(n\rightarrow \infty \) for \(p\in \mathbb {Z}_+,\) \(l\in \{1,2\}\) and any partition \((s_k)_{k=0}^\infty \in \mathfrak {S}\). Then, (A4) holds.
Proposition 2.8
Assume that (B1-q) and (B2-q) hold and that (2.4) is satisfied for any \(q>0\). Moreover, assume that there exist nonnegative constants \(f_p^{1,1},\) \(f_p^{1,2},\) and \(f_p^{2,2}\) for \(p\in \mathbb {Z}_+,\) such that \(\{E[|\mathcal {E}_{(q_n+1)}^l\mathfrak {I}_l|]\}_{n=1}^\infty \) is bounded and
as \(n\rightarrow \infty \) for \(l\in \{1,2\},\) \(p\in \mathbb {Z}_+\) and any partition \((s_k)_{k=0}^\infty \in \mathfrak {S}\). Then, (A5) holds.
Proposition 2.9
Assume that there exists \(q>0,\) such that (A4) and (B2-q) hold, \(\{\mathcal {N}_{t+h_n}^{n,l}-\mathcal {N}_t^{n,l}\}_{0\le t\le T_n-h_n, l\in \{1,2\}, n\in \mathbb {N}}\) is P-tight, and \(\sum _{k=1}^\infty k\alpha _k^n<\infty \). Then, \(a_1^1>0\).
In the following, let \((\bar{\mathcal {N}}_t^l)_{t\ge 0}\) be an exponential \(\alpha \)-mixing point process for \(l\in \{1,2\}\). Assume that the distribution of \((\bar{\mathcal {N}}_{t+t_k}^l-\bar{\mathcal {N}}_{t+t_{k-1}}^l)_{1\le k\le K, l=1,2}\) does not depend on \(t\ge 0\) for any \(K\in \mathbb {N}\) and \(0\le t_0<t_1<\cdots < t_K\).
Lemma 2.10
Let \(\mathcal {N}^{n,l}_t=\bar{\mathcal {N}}_{h_n^{-1}t}^l\) for \(0\le t\le nh_n\) and \(l\in \{1,2\}\). Then, (2.4) is satisfied for any \(q>2,\) and there exist constants \(a_0^1,\) \(a_0^2,\) and \(a_p^1=a_p^2\) for \(p\in \mathbb {N},\) such that (2.5) holds and \(\{E[h_nM_{l,q_n+1}]\}_{n=1}^\infty \) is bounded for any \((s_k)_{k=0}^\infty \in \mathfrak {S}\). Moreover, there exist nonnegative constants \(f_p^{1,1},\) \(f_p^{1,2},\) and \(f_p^{2,2}\) for \(p\in \mathbb {Z}_+,\) such that (2.6) holds and \(\{E[|\mathcal {E}_{(q_n+1)}^l\mathfrak {I}_l|\}_{n=1}^\infty \) is bounded for \(l\in \{1,2\}\) and any \((s_k)_{k=0}^\infty \in \mathfrak {S}\).
Proposition 2.11
(Proposition 8 in Ogihara & Yoshida, 2014) Let \(q\in \mathbb {N}\). Assume (B2-\((q+1)).\) Then, \(\sup _nE[h_n^{-q+1}r_n^q]<\infty \). In particular, (A2) holds under (B2-2).
By the above results, we obtain simple tractable sufficient conditions for the assumptions of the sampling scheme when the observation times are generated by the exponential \(\alpha \)-mixing point process \(\bar{\mathcal {N}}_t^l\) defined above.
Corollary 2.12
Let \(\mathcal {N}_t^{n,l}=\bar{\mathcal {N}}_{h_n^{-1}t}^l\) for \(0\le t\le T_n\) and \(l\in \{1,2\}\). Assume that (B1-q) and (B2-q) hold for any \(q>0\). Then, (A2), (A4), and (A5) hold, and \(a_1^1>0\).
3 Proofs
3.1 Preliminary results
For a real number a, [a] denotes the maximum integer which is not greater than a. Let \(\Pi =\Pi _n=\{S_i^{n,l}\}_{1\le i\le M_l, l\in \{1,2\}}\). We denote \(|x|^2=\sum _{i_1,\ldots , i_k}|x_{i_1,\ldots , i_k}|^2\) for \(x=\{x_{i_1,\ldots , i_k}\}_{i_1,\ldots , i_k}\) with \(k\in \mathbb {N}\) and \(x_{i_1,\ldots ,i_k}\in \mathbb {R}\). For a matrix \(A=(A_{ij})_{ij}\), \(\mathrm{Abs(A)}\) denotes the matrix \((|A_{ij}|)_{ij}\). C denotes generic positive constant whose value may vary depending on context. We often omit the parameters \(\sigma \) and \(\theta \) in general functions \(f(\sigma )\) and \(g(\theta )\).
For a sequence \(p_n\) of positive numbers, let us denote by \(\{\bar{R}_n(p_n)\}_{n\in {\mathbb {N}}}\) a sequence of random variables (which may also depend on \(1\le i\le M\) and \(\alpha \in \Theta \)) satisfying that \(\{\sup _{\alpha ,i}E_\Pi [|p_n^{-1}\bar{R}_n(p_n)|^q]\}_{n\in \mathbb {N}}\) is P-tight for any \(q\ge 1\), where \(E_\Pi [\textbf{X}]=E[\textbf{X}|\sigma (\Pi _n)]\) for a random variable \(\textbf{X}\).
For a matrix A and vectors v, w with suitable sizes, we repeatedly use the following inequality:
Lemma 3.1
(A special case of Lemma 3.1 in Ogihara and Uehara, 2022) Let \((Z_n)_{n\in \mathbb {N}}\) be nonnegative-valued random variables. Then
-
1.
\(E_\Pi [Z_n]\overset{P}{\rightarrow }0\) as \(n\rightarrow \infty \) implies that \(Z_n\overset{P}{\rightarrow }0\) as \(n\rightarrow \infty \).
-
2.
P-tightness of \((E_\Pi [Z_n])_{n\in \mathbb {N}}\) implies P-tightness of \((Z_n)_{n\in \mathbb {N}}\).
Let \(\bar{V}=V(\theta _0)\), and let
Let \(\bar{\rho }_n=\sup _\sigma (\max _{i,j}|\rho _{i,j}(\sigma )| \vee \sup _t|\rho _t(\sigma )|)\), and let
Let \(\Delta _{i,t}^l U=U^l_{t\wedge S^{n,l}_i}-U^l_{t\wedge S^{n,l}_{i-1}}\), and let \(\Delta _{i,t} U=\Delta _{\varphi (i),t}^{\psi (i)} U\) for \(t\ge 0\) and a two-dimensional stochastic process \((U_t)_{t\ge 0}=((U_t^1,U_t^2))_{t\ge 0}\).
Under (A4), we have
Then, we obtain
Lemma 3.2
Assume (A1). Then, for any \(p\ge 1,\) there exist positive constants \(C_p\) (depending on p) and C, such that
for \(l\in \{1,2\}\) and \(1\le i\le M_l\).
Proof
Since \(\mu _t^l(\theta )\) and \([b_tb_t(\sigma _0)]_{ll}\) are bounded by (A1), the Burkholder–Davis–Gundy inequality yields
\(\square \)
Lemma 3.3
(Lemma 2 in Ogihara & Yoshida, 2014) \(\Vert G\Vert \vee \Vert G^\top \Vert \le 1\).
Lemma 3.4
\(\Vert \tilde{G}\Vert \vee \Vert \tilde{G}^\top \Vert \le \bar{\rho }_n\).
Proof
Since all the elements of G are nonnegative, we have
Since \(\Vert \tilde{G}^\top \Vert =\Vert \tilde{G}\Vert \), we obtain the conclusion. \(\square \)
Let \(\mathcal {D}=\textrm{diag}(\{|I_i|\}_{i=1}^M)\). It is difficult to deduce the orders of upper bounds of the operator norms \(\Vert S_n(\sigma )\Vert \) and \(\Vert S_n^{-1}\Vert \), because they depend on the maximum and minimum lengths of observation intervals. However, we can deduce the orders of upper bounds for \(\tilde{\mathcal {D}}^{-1/2}S_n(\sigma )\tilde{\mathcal {D}}^{-1/2}\) and its inverse. Indeed, we obtain the following estimates, which are repeatedly used in the following sections (we use \(\mathcal {D}\) instead of \(\tilde{\mathcal {D}}\) to avoid parameter dependence).
Lemma 3.5
Assume (A1). Then, there exists a positive constant C, such that \(\Vert \mathcal {D}^{1/2} \partial _\sigma ^kS_n^{-1}(\sigma )\mathcal {D}^{1/2} \Vert \le C(1-\bar{\rho }_n)^{-k-1}\) and \(|[S_n^{-1}(\sigma )]_{ij}|\le C[\mathcal {D}^{-1/2}{\dot{S}}_n^{-1}\mathcal {D}^{-1/2}]_{ij}\) if \(\bar{\rho }_n<1,\) and \(\Vert \mathcal {D}^{-1/2} \partial _\sigma ^kS_n(\sigma )\mathcal {D}^{-1/2}\Vert \le C\) for any \(\sigma \in \Theta _1,\) \(1\le i,j\le M,\) and \(k\in \{0,1,2,3,4\}\).
Proof
By (A1) and Lemma 3.3, we have
Moreover, by (A1) and Lemma 3.4, we have
if \(\bar{\rho }_n<1\).
Using the equation \(\partial _\sigma S_n^{-1}=-S_n^{-1}\partial _\sigma S_n S_n^{-1}\), we obtain
if \(\bar{\rho }_n<1\). Similarly, we obtain
if \(\bar{\rho }_n<1\) for \(k\in \{0,1,2,3,4\}\).
If \(\bar{\rho }_n<1\), since Lemma 3.4 yields
we obtain
\(\square \)
Under (A1), we have \(\Sigma _t(\sigma )\ge c_1\mathcal {E}_2\), which implies that \(\sup _{t,\sigma }|\rho _t(\sigma )|<1\). Then, by (A2) and uniform continuity of \(b_t\), for some fixed \(\delta >0\) and any \(\epsilon >0\), there exists \(N\in \mathbb {N}\), such that \(P(1-\bar{\rho }_n<\delta )<\epsilon \) for \(n\ge N\). Therefore, we have
as \(n\rightarrow \infty \), and we have
for any \(q>0\) and \(n\ge N\), which implies that
Moreover, Lemma 3.4 yields
if \(\bar{\rho }_n<1\).
3.2 Consistency of \(\hat{\sigma }_n\)
In this section, we show consistency: \(\hat{\sigma }_n\overset{P}{\rightarrow }\sigma _0\) as \(n\rightarrow \infty \). For this purpose, we specify the limit of \(H_n^1(\sigma )-H_n^1(\sigma _0)\).
Lemma 3.6
Assume (A1) and (A2). Then
as \(n\rightarrow \infty \) for \(k\in \{0,1,2,3\}\).
Proof
Let \(X_t^c=\int _0^tb_s(\sigma _0){\textrm{d}}W_s\). By the definition of \(H_n^1\), we have
We first show that
where \(({\dot{e}}_n(\sigma ))_{n=1}^\infty \) denotes a general sequence of random variables, such that \(\sup _\sigma |{\dot{e}}_n(\sigma )|\overset{P}{\rightarrow }0\) as \(n\rightarrow \infty \).
Since
it suffices to show that \(\Psi _i=\sqrt{n}{\dot{e}}_n\) for \(i\in \{1,2\}\).
Moreover, Lemma 3.2 yields
Furthermore, Lemma 3.5, (3.4), (3.10), and the equation \(E_\Pi [\Delta X^c(\Delta X^c)^\top ]=S_n(\sigma _0)\) yield
Then, we obtain (3.7) by (3.8)–(3.11) and Lemma 3.1.
Next, we show that
Itô’s formula yields
where \(X^{c,l}_t\) is the l-th component of \(X_t^c\).
Since \(\langle \Delta _i X^c, \Delta _j X^c\rangle _t=\int _{[0,t)\cap I_i\cap I_j}[\Sigma _t]_{\psi (i),\psi (j)}{\textrm{d}}t\), together with the Burkholder–Davis–Gundy inequality, we have
Together with Lemmas 3.3 and 3.5, the triangle inequality for \(L^{q/2}\) that
and that
by Lemma 3.5, we have
on \(\{\bar{\rho }_n<1\}\) for \(q\ge 1\). Then, thanks to (3.2), (3.4), (3.13) and Lemma 3.1, we obtain (3.12).
(3.12), (3.7), Sobolev’s inequality, and similar estimates for \(\partial _\sigma ^k(H_n^1(\sigma )-H_n^1(\sigma _0))\) yield
for \(k\in \{0,1,2,3\}\). \(\square \)
For \((s_k)_{k=0}^\infty \in \mathfrak {S}\), let \(\dot{\mathcal {A}}_{k,p}^1=\mathcal {E}_{(k)}^1(GG^\top )^p\) and \(\dot{\mathcal {A}}_{k,p}^2=\mathcal {E}_{(k)}^2(G^\top G)^p\) for \(p\in \mathbb {Z}_+\) and \(1\le k\le q_n\). The following lemma is used when we specify the limit of \(n^{-1}(H_n^1(\sigma )-H_n^1(\sigma _0))\) in the next proposition.
Lemma 3.7
Assume (A2) and (A4). Then, for any \(p\ge 1\)
as \(n\rightarrow \infty \).
Proof
By the definition of \(\dot{\mathcal {A}}_{k,p}^l\), we obtain
Two summands in the right-hand side coincide when both \(\sup I_i^1\) and \(\sup I_j^2\) are included or not included in \((s_{k-1},s_k]\). In other cases, we have \(\min _{u=0,1}|\sup I_i^1-s_{k-u}|\le r_n\) if \([G^\top ]_{ji}>0\). Therefore, we obtain
Thanks to (A2) and (A4), the right-hand side of the above inequality is equal to \(O_p(h_n^{-1})=o_p(n)\). \(\square \)
Let \(\mathcal {Y}_1(\sigma )=\lim _{T\rightarrow \infty }(T^{-1}\int _0^T y_{1,t}(\sigma ){\textrm{d}}t)\), where
The limit \(\mathcal {Y}_1(\sigma )\) exists under (A1), (A3), and (A4).
Proposition 3.8
Assume (A1)–(A4). Then
as \(n\rightarrow \infty \) for \(k\in \{0,1,2,3\}\).
Proof
Let \(\mathcal {A}_p^1=(\tilde{G}\tilde{G}^\top )^p\), \(\mathcal {A}_p^2=(\tilde{G}^\top \tilde{G})^p\), \(\tilde{\Sigma }_{i,0}^l=\tilde{\Sigma }_i^l(\sigma _0)\), and \(\tilde{\Sigma }_{i,j,0}^{1,2}=\tilde{\Sigma }_{i,j}^{1,2}(\sigma _0)\). Thanks to (A1), for any \(\epsilon >0\), there exists \(\delta >0\), such that \(|t-s|<\delta \) implies
for any \(\sigma \) and \(\theta \). We fix such \(\delta >0\), and fix a partition \(s_k= k\delta /2\). Then, (3.5) and (A4) yield
if \(\bar{\rho }_n<1\).
Let \(\dot{\rho }_k=\rho _{s_{k-1}}\) and \({\dot{B}}_{k,l}=([\Sigma _{s_{k-1}}(\sigma _0)]_{ll}/[\Sigma _{s_{k-1}}(\sigma )]_{ll})^{1/2}\). Then, (3.15) yields that for any \(p\in \mathbb {Z}_+\), we have
on \(\{2pr_n<\delta /2\}\). Here, the factor p in the right-hand side appears, because we consider the difference between 2p products of \(\rho _{i'j'}\) and \(\dot{\rho }_k^{2p}\). Moreover, Lemma 3.4 and (3.4) yield
almost surely.
Then, together with (A2) and Lemma 3.7, we obtain
where \(\dot{\rho }_{k,0}=\rho _{s_{k-1}}(\sigma _0)\), and \((e_n)_{n=1}^\infty \) denotes a general sequence of random variables such that \(\limsup _{n\rightarrow \infty }|e_n|\rightarrow 0\) as \(\delta \rightarrow 0\).
Moreover, by (3.3) and Lemma 3.4, we can apply Lemma A.3 in Ogihara (2018) to \(S_n\). Then, we have
if \(\bar{\rho }_n<1\). Therefore, thanks to (3.2) and (3.17), we obtain
Lemma 3.7, (3.6), (3.19), and (3.20) yield
Here, we used that \(\textrm{tr}(\dot{\mathcal {A}}_{k,0}^l)=\textrm{tr}(\mathcal {E}_{(k)}^l)=M_{l,k}\).
Moreover, (A4) and (3.15) yield
for \(p\ge 1\) and any choice of \(f(t)=\rho _t^{2p}B_{l,t}^2\), \(\rho _t^{2p-1}\rho _{t,0}B_{1,t}B_{2,t}\) and \((\rho _t^{2p}-\rho _{t,0}^{2p})/(2p)\).
Here, we used that \(q_n=O(nh_n)\) by the definition of \((s_k)_{k=0}^\infty \in \mathfrak {S}\). Similarly, we obtain
Together with (3.21), (A3) and the equation
we obtain
The above arguments show that the supremum with respect to \(\sigma \) of the residual term in the above equation is also equal to \(ne_n\), and consequently, we obtain (3.14) with \(k=0\). Similarly, we obtain (3.14) with \(k\in \{1,2,3\}\). \(\square \)
Proposition 3.9
Assume (A1)–(A4). Then, there exists a positive constant \(\chi ,\) such that
Proof
The proof is based on the ideas of proof of Lemma 5 in Ogihara and Yoshida (2014). Let
and let \(\tilde{\mathcal {A}}_{k,p}^1=(G_kG_k^\top )^p\) and \(\tilde{\mathcal {A}}_{k,p}^2=(G_k^\top G_k)^p\). Let \(\tilde{\mathcal {A}}_k=\sum _{p=1}^\infty \dot{\rho }_k^{2p}\textrm{tr}(\tilde{\mathcal {A}}_{k,p}^1)\) and \(\tilde{\mathcal {B}}_k=\sum _{p=1}^\infty (2p)^{-1}(\dot{\rho }_k^{2p}-\dot{\rho }_{k,0}^{2p})\textrm{tr}(\tilde{\mathcal {A}}_{k,p}^1)\). Similarly to the proof of Lemma 3.7, the difference between \(\textrm{tr}(\dot{\mathcal {A}}_{k,p}^l)\) and \(\textrm{tr}(\tilde{\mathcal {A}}_{k,p}^l)\) comes from terms with \(\sup I_i^1\) close to \(s_{k-1}\) or \(s_k\), and hence, we obtain
Therefore, (3.21) yields
Then, since
and a similar estimate holds by switching the roles of \(M_{1,k}\) and \(M_{2,k}\), we have
For \(l\in \{1,2\}\), let
then since \(1-x+\log x \le 0\) for \(x>0\), we obtain
and therefore, we have
Let \((\lambda _i^k)_{i=1}^{M_{1,k}}\) be all the eigenvalues of \(G_kG_k^\top \). Similarly to Lemma 3.3, we have \(0\le \lambda _i^k\le 1\). Then, we have
Moreover, by setting \(g_i^k=\sqrt{1-\lambda _i^k\dot{\rho }_k^2}\), \(g_{i,0}^k=\sqrt{1-\lambda _i^k\dot{\rho }_{k,0}^2}\), and \(F(x)=1-x+\log x\), we have
Here, we also used the expansion formulas \((1-x)^{-1}=\sum _{p=0}^\infty x^p\) and \(-\log (1-x)=\sum _{p=1}^\infty x^p/p\) for \(|x|<1\).
Let
Since \(g_i^k\le 1\), \(0\le \lambda _i^k\le 1\), and \(|\dot{\rho }_k|< 1\), we have
Together with Lemma 11 in Ogihara and Yoshida (2014) and
we have
Moreover, the inequality \(a^2\ge (a+b)^2/2-b^2\) with \(a={\dot{B}}_{k,1}{\dot{B}}_{k,2}g_{i,0}^k-g_i^k\) and \(b=g_i^k-g_{i,0}^k\) yields
and hence, we have
By a similar argument for \(F_{2,k}\), there exists a positive constant \(\tilde{\chi }\) which does not depend on k nor n, such that
Together with (3.23), we have
By letting \(n\rightarrow \infty \), (A4) and (3.3) yield the conclusion. \(\square \)
(A6) and Remark 4 in Ogihara and Yoshida (2014) yield that
when \(\sigma \ne \sigma _0\).
Then, by Proposition 3.9, we have \(\mathcal {Y}_1(\sigma )<0\) (note that \(a_0^1\wedge a_0^2\ge a_1^1\) by (2.3) and a similar argument). Therefore, for any \(\delta >0\), there exists \(\eta >0\), such that
Then, since \(H_n^1(\hat{\sigma }_n)-H_n^1(\sigma _0)\ge 0\) by the definition, for any \(\epsilon >0\), we have
for sufficiently large n by Proposition 3.8, which implies \(\hat{\sigma }_n\overset{P}{\rightarrow }\sigma _0\) as \(n\rightarrow \infty \).
3.3 Asymptotic normality of \(\hat{\sigma }_n\)
Let \(S_{n,0}=S_n(\sigma _0)\) and \(\Sigma _{t,0}=\Sigma _t(\sigma _0)\). (3.7) and the equation \(\partial _\sigma S_{n,0}^{-1}=-S_{n,0}^{-1}\partial _\sigma S_{n,0} S_{n,0}^{-1}\) imply
Let \((L_n)_{n\in \mathbb {N}}\) be a sequence of positive integers such that \(L_n\rightarrow \infty \) and \(L_nn^\eta (nh_n)^{-1}\rightarrow 0\) as \(n\rightarrow \infty \) for some \(\eta >0\). Let \(\check{s}_k=kT_n/L_n\) for \(0\le k\le L_n\), let \(J^k=(\check{s}_{k-1},\check{s}_k]\), and let \(S_{n,0}^{(k)}\) be an \(M\times M\) matrix satisfying
For a two-dimensional stochastic process \((U_t)_{t\ge 0}=((U_t^1,U_t^2))_{t\ge 0}\), let \(\Delta _{i,t}^{l,(k)} U=U^l_{(S^{n,l}_i\vee \check{s}_{k-1})\wedge \check{s}_k\wedge t}-U^l_{(S^{n,l}_{i-1}\vee \check{s}_{k-1})\wedge \check{s}_k\wedge t}\), and let \(\Delta _{i,t}^{(k)} U=\Delta _{\varphi (i),t}^{\psi (i),(k)} U\) for \(1\le i\le M\). Let \(\Delta _i^{(k)} U=\Delta _{i,T_n}^{(k)} U\), and let \(\Delta ^{(k)} U=(\Delta _i^{(k)} U)_{1\le i\le M}\).
Let
Then, since \(\Delta X^c=\sum _{k=1}^{L_n}\Delta ^{(k)} X^c\) and \(S_{n,0}=\sum _{k=1}^{L_n}S_{n,0}^{(k)}\), (3.25) yields
Moreover, Itô’s formula yields
Let \(\mathcal {G}_t=\mathcal {F}_t\bigvee \sigma (\{\Pi _n\}_n)\) for \(t\ge 0\). We will show
using Corollary 3.1 and the remark after that in Hall and Heyde (1980). For this purpose, it is sufficient to show
and
by (3.26), where \(E_k\) denotes the conditional expectation with respect to \(\mathcal {G}_{\check{s}_{k-1}}\).
We first show four auxiliary lemmas. Let \(\tilde{M}_k=\#\{i; 1\le i\le M, \sup I_i\in J^k\}\).
Lemma 3.10
Assume (A1). Then, there exists a positive constant C, such that \(\Vert \mathcal {D}^{-1/2}S_{n,0}^{(k)}\mathcal {D}^{-1/2}\Vert \le C\) and \(\textrm{tr}(\mathcal {D}^{-1/2}S_{n,0}^{(k)}\mathcal {D}^{-1/2})\le C(\tilde{M}_k+1)\) for any \(1\le k\le L_n\).
Proof
Since
Lemma 3.3 yields
Moreover, we have
\(\square \)
Lemma 3.11
Assume (A4) and that \(nh_nL_n^{-1}\rightarrow \infty \) as \(n\rightarrow \infty \). Then, \(\{L_nn^{-1}\max _{1\le k\le L_n}\tilde{M}_k\}_{n=1}^\infty \) is P-tight.
Proof
Let \(\mathcal {M}_n=[nh_nL_n^{-1}]\). We define a partition of \([0,\infty )\) by
Then, \((s_j)_{j=0}^\infty \in \mathfrak {S}\) when \(nh_nL_n^{-1}\ge 1\), and \((s_j)_{j=0}^{2L_n\mathcal {M}_n}\) is a subpartition of \((\check{s}_k)_{k=0}^{L_n}\).
For \(M_{l,j}\) which corresponds to this partition (\(\tilde{M}_k\) remains to be defined using \(\check{s}_k\)), we have
since \({\check{s}_k=nh_nkL_n^{-1}=s_{2\mathcal {M}_nk}}\). Therefore, (A4) yields
\(\square \)
Lemma 3.12
Assume (A1). Then
on \(\{\bar{\rho }_n<1\}\) for \(|k-k'|>1,\) where \(\mathcal {Q}_n=[r_n^{-1}(T_n/L_n-2r_n)]\).
Proof
Using the expansion formula (3.5), we have
if \(\bar{\rho }_n<1\), where
We consider a necessary condition for
to be zero for any \(i'\) and \(j'\). We first observe that the element \([{\mathfrak {C}}^n_{p,q}]_{ij}\) is equal to zero if \([\bar{S}^{p+q+1}]_{ij}=0\), where
Moreover, \([S_{n,0}^{(k)}]_{i'i}\ne 0\) only if \(I_i\cap J^k \ne \emptyset \), and \([S_{n,0}^{(k')}]_{jj'}\ne 0\) only if \(I_j\cap J^{k'}\ne \emptyset \). Since \(\inf _{x\in I_i, y\in I_j}|x-y|> T_n/L_n-2r_n\) if \(I_i\cap J^k \ne \emptyset \) and \(I_j\cap J^{k'}\ne \emptyset \), we have \([\bar{S}^r]_{ij}=0\) for \(r\le \mathcal {Q}_n\) when \([S_{n,0}^{(k)}]_{i'i}\ne 0\) and \([S_{n,0}^{(k')}]_{jj'}\ne 0\). Therefore, \([S_{n,0}^{(k)}{\mathfrak {C}}_{p,q}^n S_{n,0}^{(k')}]_{i',j'}=0\) for any \(i'\) and \(j'\) if \(p+q+1\le \mathcal {Q}_n\).
Then, (3.31) and Lemmas 3.4, 3.5 and 3.10 yield
on \(\{\bar{\rho }_n<1\}\). \(\square \)
Lemma 3.13
Let \(m\in \mathbb {N}\). Let V be an \(m\times m\) symmetric, positive definite matrix and A be a \(m\times m\) matrix. Let X be a random variable following N(0, V). Then
Proof
We only show the result for \(E[(X^\top AX)^4]\). Let U be an orthogonal matrix and \(\Lambda \) be a diagonal matrix satisfying \(UVU^\top =\Lambda \). Then, we have \(UX\sim N(0,\Lambda )\), and
where the summation of \((l_{2q-1},l_{2q})_{q=1}^4\) is taken over all disjoint pairs of \(\{j_1,\ldots j_8\}\). Then, by setting \(B=UAU^\top \), we have
Let \(_nC_k=\frac{n!}{k!(n-k)!}\). Out of \(j_1,\ldots , j_8\), we connect \(j_{2p-1}\) to \(j_{2p}\) and \(l_{2q-1}\) to \(l_{2q}\) (\(1\le p,q\le 4\)). Then, the pattern of the connected components gives five different cases.
-
1.
Four connected components (four components of size 2): only one case of the pairs \((l_{2q-1},l_{2q})_{q=1}^4\) appears, which corresponds to \(\textrm{tr}((B\Lambda )^4)\).
-
2.
Three connected components (a component of size 4 and two components of size 2): The choice of elements for a components of size 4 gives \(_4C_2\) ways, and the choice of the pair \((l_{2q-1},l_{2q})\) for this component gives two ways, and hence, \(_4C_2\times 2=12\) ways in total. This case corresponds to \(\textrm{tr}(B\Lambda )^2\textrm{tr}((B\Lambda )^2)\).
-
3.
Two connected components (two components of size 4): The choice of elements for each component gives \(\frac{_4C_2}{2}\) ways, excluding duplicates, and the choice of the pair \((l_{2q-1},l_{2q})\) for each component gives two ways, and hence, \(\frac{_4C_2}{2}\times 2\times 2=12\) ways in total. This case corresponds to \(\textrm{tr}((B\Lambda )^2)^2\).
-
4.
Two connected components (a component of size 6 and a component of size 2): The choice of elements for a components of size 6 gives \(_4C_1\) ways, and the choice of the pair \((l_{2q-1},l_{2q})\) for this component gives \(4\times 2=8\) ways, and hence \(_4C_1\times 8=32\) ways in total. This case corresponds to \(\textrm{tr}(B\Lambda )\textrm{tr}((B\Lambda )^3)\).
-
5.
One connected component (a component of size 8): The choice of the pair \((l_{2q-1},l_{2q})\) gives \(6\times 4\times 2=48\) ways. This case corresponds to \(\textrm{tr}((B\Lambda )^4)\).
Then, we obtain the conclusion. \(\square \)
Proposition 3.14
Assume (A1)–(A4) and (A6). Then
as \(n\rightarrow \infty \).
Proof
It is sufficient to show (3.29) and (3.30). Let \(\mathfrak {A}_k=(\Delta ^{(k)} X^c)^\top \partial _\sigma S_{n,0}^{-1}\Delta ^{(k)}X^c\) and \(\mathfrak {B}_k=\partial _\sigma S_{n,0}^{-1}S_{n,0}^{(k)}\). By the definition of \(\mathcal {X}_k\), we have
Thanks to Lemmas 3.13, 3.10, 3.11 and 3.5, (3.4), and Lemma A.1 in Ogihara (2018), the first term in the right-hand side is calculated as
Moreover, Lemma 3.13 yields
If \(k'_1<k-1\), Lemmas 3.5 and 3.12, Lemma A.1 in Ogihara (2018) and the equation \(\partial _\sigma S_{n,0}^{-1}=-S_{n,0}^{-1}\partial _\sigma S_{n,0} S_{n,0}^{-1}\) yield
on \(\{\bar{\rho }_n<1\}\). Here, we used that \(\textrm{tr}(\tilde{\mathcal {D}}^{1/2}S_{n,0}^{-1}\tilde{\mathcal {D}}^{1/2})\le M\cdot \Vert \tilde{\mathcal {D}}^{1/2}S_{n,0}^{-1}\tilde{\mathcal {D}}^{1/2}\Vert \le CM(1-\bar{\rho }_n)^{-1}\). Similarly, we obtain
Since \(\bar{\rho }_n^{\mathcal {Q}_n}\) converges to zero very fast if \(\bar{\rho }_n<1\) and \(r_n\le 1\), together with (A2) and (3.3), the summation for of the terms with \(k'_1<k-1\) or \(k'_2<k-1\) in the right-hand side of (3.35) is equal to \(o_p(1)\).
Then, together with Lemmas 3.5, 3.10, and 3.11, and Lemma A.1 in Ogihara (2018), we obtain
as \(n\rightarrow \infty \). Then, (3.33), (3.34), and (3.36) yield (3.30).
Next, we show (3.29). Let \(\mathcal {I}_{i,j}^k=I_i\cap I_j\cap J^k\). Then, (3.27) yields
We can decompose
where \(F_{ij}^k(t)=[\Sigma _{t,0}]_{\psi (i),\psi (j)}1_{\mathcal {I}_{i,j}^k}(t)\), and \(\mathcal {F}_{i,j}^k=\int _0^{T_n}F_{i,j}^k(t){\textrm{d}}t\). Moreover, switching the roles of \(i_1,i_2\) and \(j_1,j_2\), we obtain
Therefore, we have
\(\partial _\sigma S_{n,0}^{-1}S_{n,0}\) corresponds to \(\hat{\mathcal {D}}(t)\) in the proof (p. 2993) of Proposition 10 of Ogihara and Yoshida (2014). Then, by a similar step to the proof of Proposition 10 in Ogihara and Yoshida (2014), we have (3.29). \(\square \)
Proposition 3.15
Assume (A1)–(A4) and (A6). Then, \(\Gamma _1\) is positive definite and
as \(n\rightarrow \infty \).
Proof
Proposition 3.9, (A6), and Remark 4 in Ogihara and Yoshida (2014) yield
for some positive constant c. Moreover, \(\mathcal {Y}_1(\sigma _0)=0\) by \(B_{l,t,0}=1\), and \(\partial _\sigma \mathcal {Y}_1(\sigma _0)=0\) by
Then, Taylor’s formula yields
Therefore, considering \(\sigma \) sufficiently close to \(\sigma _0\), \(\Gamma _1=-\partial _\sigma ^2 \mathcal {Y}_1(\sigma _0)\) should be positive definite by (3.39).
By Taylor’s formula and the equation \(\partial _\sigma H_n^1(\hat{\sigma }_n)=0\), we have
where \(\sigma _t=t\hat{\sigma }_n+(1-t)\sigma _0\).
Therefore, we obtain
Since Proposition 3.8 yields
and
is P-tight, together with Proposition 3.14, we conclude
\(\square \)
3.4 Consistency of \(\hat{\theta }_n\)
Let
which exists under (A1), (A3), and (A5).
Proposition 3.16
Assume (A1)–(A6). Then
as \(n\rightarrow \infty \) for \(k\in \{0,1,2,3\}\).
Proof
We first show that
where \(({\dot{e}}_n(\theta ))_{n=1}^\infty \) denotes a general sequence of random variables, such that \(\sup _\theta |{\dot{e}}_n(\theta )|\overset{P}{\rightarrow }0\) as \(n\rightarrow \infty \).
on \(\{\bar{\rho }_n<1\}\).
Since (3.2), Lemma 3.2 and Taylor’s formula yield
and
(3.4), (3.10), (3.41), (3.45), and Lemma 3.5 simply
Thanks to (3.4), (3.10), (3.41), and Lemma 3.5, we have
For \(k\in \{0,1\}\) and \(q\ge 1\), the Burkholder–Davis–Gundy inequality, Lemma 3.5 and a similar estimate to (3.10) yield
Together with (3.4), (3.41), (3.48), and Sobolev’s inequality, we have
Then, (3.47) and (3.50) yield (3.43).
Applying (3.43) to \(\theta \) and \(\theta =\theta _0\), we have
and hence, by similar estimates to (3.49), we have
Then, (3.5), (3.17), and a similar argument to (3.16) yield
where \(\dot{\mathfrak {I}}_{k,l}=\mathcal {E}_{(k)}^l\dot{\mathfrak {I}}_l\). Together with (A3), (A5), (3.52), and a similar argument to (3.22), we obtain
as \(n\rightarrow \infty \). Similar estimates for \((nh_n)^{-1}\partial _\theta ^k(H_n^2(\theta )-H_n^2(\theta _0))\) \((k\in \{0,1,2,3,4\})\) yield the conclusion. \(\square \)
Proposition 3.17
Assume (A1)–(A6). Then, \(\hat{\theta }_n\overset{P}{\rightarrow }\theta _0\) as \(n\rightarrow \infty .\)
Proof
By Lemma 3.5, we have
Therefore, together with (3.15) and (3.16), we obtain
Hence, we have
Assumption (A6) yields that for any \(\theta \in \Theta \)
(3.42), (3.57) together with a similar estimate to (3.24), we have the conclusion. \(\square \)
3.5 Asymptotic normality of \(\hat{\theta }_n\)
Proof of Theorem 2.3
By the definition of \(H_n^2(\theta )\), we obtain
By a similar argument to the derivation of (3.43), we can replace \(S_n^{-1}(\hat{\sigma }_n)\) in the right-hand side of the above equation by \(S_{n,0}^{-1}\) with approximation error equal to \(o_p(\sqrt{nh_n})\). Then, we have
Let
for \(1\le k\le L_n\). Then, we have
Lemma 3.5 and a similar argument to (3.10) yield
Moreover, (3.5), (A5), and a similar argument to the proof of Proposition 3.8 yield
Therefore, (3.58) and the martingale central limit theorem (Corollary 3.1 and the remark after that in Hall & Heyde, 1980) yield
By (3.56) and (A6), there exists a positive constant c, such that \(\mathcal {Y}_2(\theta )\le -c|\theta -\theta _0|^2\). Then, \(\Gamma _2=-\partial _\theta ^2\mathcal {Y}_2(\theta _0)\) is positive definite, since \(\mathcal {Y}_2(\theta _0)=0\) and \(\partial _\theta \mathcal {Y}_2(\theta _0)=0\).
Therefore, a similar estimate to Sect. 3.3, P-tightness of \(\{(nh_n)^{-1}\sup _\theta |\partial _\theta ^3H_n^2(\theta )|\}_n\), and the equation \(-(nh_n)^{-1}\partial _\theta ^2H_n^2(\theta _0)\overset{P}{\rightarrow }\Gamma _2\) yield
(3.40) and a similar equation for \(\sqrt{nh_n}(\hat{\theta }_n-\theta _0)\) yield
Then, since \(\sum _{k=1}^{L_n}E_k[\mathcal {X}_k\dot{\mathcal {X}}_k]=0\), we obtain
\(\square \)
3.6 Proofs of the results in Sects. 2.3 and 2.4
Proof of Theorem 2.5
Let \(\sigma _{tu}=\sigma _0+t\epsilon _n u\) for \(u\in \mathbb {R}^d\) and \(t\in [0,1]\), and let
Then, we have
By similar arguments to Propositions 3.8 and 3.14, and Sects. 3.4 and 3.5, we obtain
Therefore, we have the desired conclusion. \(\square \)
Remark 3.18
We can show that \((\hat{\sigma }_n,\hat{\theta }_n)\) is a regular estimator by the proof of Theorem 2.5, (3.60), and Theorem 2 in Jeganathan (1982).
Outline of the proof of Proposition 2.7
The proof is similar to the proof of Proposition 6 in Ogihara and Yoshida (2014). P-tightness of \(\{h_nM_{l,q_n+1}\}_{n=1}^\infty \) immediately follows from (B1-1). Fix \(1\le j\le q_n\). Then, using the mixing property (2.4) for \(\mathcal {N}_t^{n,l}\), we obtain the following result; there exists \(\eta >0\), such that for any \(q\ge 4\), there exists \(C_q>0\) that does not depend on j, such that
(The above inequality corresponds to (31) in Ogihara and Yoshida (2014). This is obtained by defining \(b_n=h_n^{-1}\), \(t_k=s_{j-1}+k[h_n^{-1}]^{-1}(s_j-s_{j-1})\) for \(0\le k\le [h_n^{-1}]\), and \(X'_k=\textrm{tr}(\mathcal {E}_{(j,k)}^1(GG^\top )^p)1_{A_{k,b_n^{\delta '}}^p}-E[\textrm{tr}(\mathcal {E}_{(j,k)}^1(GG^\top )^p)1_{A_{k,b_n^{\delta '}}^p}]\) in the proof of Proposition 6 in Ogihara and Yoshida (2014), where \(\mathcal {E}_{(j,k)}^l\) is an \(M_l\times M_l\) matrix satisfying \([\mathcal {E}_{(j,k)}^l]_{ii'}=1\) if \(i=i'\) and \(\sup I_i^l\in (t_{k-1},t_k]\), and otherwise, \([\mathcal {E}_{(j,k)}^l]_{ii'}=0\).)
Therefore, by setting sufficiently large q, so that \(nh_n^{1+q\eta }\rightarrow 0\), we have
Here, we used that for any partition \((s_k)_{k=0}^\infty \in \mathfrak {S}\), we have \(q_n\le nh_n/\epsilon +1\) with \(\epsilon =\inf _{k\ge 1}|s_k-s_{k-1}|>0\), which implies \(q_n=O(nh_n)\). Together with the assumptions, we obtain the conclusion.
Outline of the proof of Proposition 2.8
Similarly to the previous proposition, using the idea of Proposition 6 in Ogihara and Yoshida (2014) and the mixing property (2.4) for \(\mathcal {N}_t^{n,l}\), we have that there exists \(\eta >0\), such that for any \(q\ge 4\), there exists \(C_q>0\), such that
for \(1\le j\le q_n\). (We define \(b_n\) and \(t_k\) the same as the previous proposition, and define
Together with the assumptions and similar estimates for \(\mathfrak {I}_1\mathcal {E}_{(j)}^1(GG^\top )^pG\mathfrak {I}_2\) and \(\mathfrak {I}_2\mathcal {E}_{(j)}^2(G^\top G)^p\mathfrak {I}_2\), we obtain the conclusion.
Outline of the proof of Proposition 2.9
We can show the results by a similar approach to the proof of Proposition 9 in Ogihara and Yoshida (2014). Roughly speaking, under (B2-q), the probability \(P(\mathcal {N}_{t+Nh_n}^{n,l}-\mathcal {N}_t^{n,l}=0)\) is small enough to estimate the denominator of
for sufficiently large N. Then, we obtain estimates for the numerator using an inequality \(x_1^2+\cdots +x_n^2\ge R^2/n\) when \(x_1+\cdots +x_n=R\).
Proof of Lemma 2.10
We only show
The other results are similarly obtained.
(2.4) is satisfied, because \(\alpha _k^n\le c_1e^{-c_2k}\) for some positive constants \(c_1\) and \(c_2\).
Let \(\bar{\tau }_i^l\) be i-th jump time of \(\bar{\mathcal {N}}^l\). Then, we have \(S_i^{n,l}=h_n\bar{\tau }_i^l\). Let \(\bar{G}\) be a matrix with infinity size defined by
for \(i,j\ge 1\).
For \(k\in \mathbb {N}\), let
The following idea is based on Section 7.5 of Ogihara and Yoshida (2014). Roughly speaking, if there are sufficient observations around the interval \([k-1,k)\), we can apply mixing property of \(\bar{\mathcal {N}}_t^{n,l}\) to \(\mathfrak {G}_k^p\). On the following sets \(A_{k,r}^p\) and \(\bar{A}_{k,r}^p\), we have sufficient observations of \(\mathcal {N}^{n,l}\) and \(\bar{\mathcal {N}}^l\). Let \(\bar{\Delta }_{j,t}^r U=U_{t+rj}-U_{t+r(j-1)}\) for a stochastic process \((U_t)_{t\ge 0}\), and let
Then, we obtain
We also have \(P((\bar{A}_{k,r}^p)^c)\le C(p+1)r^{-q}\) by (B2-q). For any \(\epsilon >0\), there exists \(r>0\), such that
Therefore, \(\{E[\mathfrak {G}_k^p]\}_k\) is a Cauchy sequence, and hence, the limit \(a_p^1=\lim _{k\rightarrow \infty }E[\mathfrak {G}_k^p]\) exists for \(p\in \mathbb {N}\). Moreover, we see existence of
for \(l\in \{1,2\}\).
Furthermore, for any \(\epsilon >0\), there exists \(r>0\), such that
for \(k\ge [rp]\). We also have
for \(rp+1\le k\le n-rp\), since
Let \(r_j=[h_n^{-1}s_j]\). Then, since \(|\mathfrak {G}_k^{n,p}|\le \sum _{i;S_{i-1}^{n,l}\in ((k-1)h_n,kh_n]}1\le E[\bar{\mathcal {N}}_1^1]\), (3.63), (3.64), and the Cauchy–Schwarz inequality yield
for \(1<j<q_n\). To get the corresponding inequality for \(j=1, q_n\), we replace the summation range of k in the above inequality with the range from \(r_{j-1}+[rp]+2\) to \(r_j\) when \(j=1\), and with the range from \(r_{j-1}\) to \(r_j-[rp]-1\) when \(j=q_n\). Boundedness of \(\{E[h_nM_{l,q_n+1}]\}_{n\in \mathbb {N}}\) is shown using the same techniques. Then, we have the conclusion. \(\square \)
Data Availability
Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.
References
Barndorff-Nielsen, O. E., Hansen, P. R., Lunde, A., & Shephard, N. (2011). Multivariate realised kernels: Consistent positive semi-definite estimators of the covariation of equity prices with noise and non-synchronous trading. Journal of Econometrics, 162(2), 149–169.
Bibinger, M., Hautsch, N., Malec, P., & Reiss, M. (2014). Estimating the quadratic covariation matrix from noisy observations: Local method of moments and efficiency. Annals of Statistics, 42(4), 80–114.
Christensen, K., Kinnebrock, S., & Podolskij, M. (2010). Pre-averaging estimators of the ex-post covariance matrix in noisy diffusion models with non-synchronous data. Journal of Econometrics, 159(1), 116–133.
Florens-Zmirou, D. (1989). Approximate discrete-time schemes for statistics of diffusion processes. Statistics, 20(4), 547–557.
Hall, P., & Heyde C. C. (1980). Martingale limit theory and its application. Academic Press, Inc. [Harcourt Brace Jovanovich, Publishers]. Probability and Mathematical Statistics.
Hayashi, T., & Yoshida, N. (2005). On covariance estimation of non-synchronously observed diffusion processes. Bernoulli, 11(2), 359–379.
Hayashi, T., & Yoshida, N. (2008). Asymptotic normality of a covariance estimator for nonsynchronously observed diffusion processes. Annals of the Institute of Statistical Mathematics, 60(2), 367–406.
Hayashi, T., & Yoshida, N. (2011). Nonsynchronous covariation process and limit theorems. Stochastic Processes and Their Applications, 121(10), 2416–2454.
Ibragimov, I. A., & Has’minskiĭ, R. Z. (1981). Statistical estimation. Applications of Mathematics (Vol. 16). Springer. Asymptotic theory, Translated from the Russian by Samuel Kotz.
Jeganathan, P. (1982). On the asymptotic theory of estimation when the limit of the log-likelihood ratios is mixed normal. Sankhyā Series A, 44(2), 173–212.
Kessler, M. (1997). Estimation of an ergodic diffusion from discrete observations. Scandinavian Journal of Statistics, 24(2), 211–229.
Malliavin, P., & Mancino, M. E. (2002). Fourier series method for measurement of multivariate volatilities. Finance and Stochastics, 6(1), 49–61.
Malliavin, P., & Mancino, M. E. (2009). A Fourier transform method for nonparametric estimation of multivariate volatility. Annals of Statistics, 37(4), 1983–2010.
Ogihara, T. (2015). Local asymptotic mixed normality property for nonsynchronously observed diffusion processes. Bernoulli, 21(4), 2024–2072.
Ogihara, T. (2018). Parametric inference for nonsynchronously observed diffusion processes in the presence of market microstructure noise. Bernoulli, 24(4B), 3318–3383.
Ogihara, T., & Uehara, Y. (2022). Local asymptotic normality for ergodic jump-diffusion processes via transition density approximation. Bernoulli (to appear).
Ogihara, T., & Yoshida, N. (2014). Quasi-likelihood analysis for nonsynchronously observed diffusion processes. Stochastic Processes and Their Applications, 124(9), 2954–3008.
Uchida, M., & Yoshida, N. (2012). Adaptive estimation of an ergodic diffusion process based on sampled data. Stochastic Processes and Their Applications, 122(8), 2885–2924.
Yoshida, N. (1992). Estimation for diffusion processes from discrete observation. Journal of Multivariate Analysis, 41(2), 220–242.
Yoshida, N. (2011). Polynomial type large deviation inequalities and quasi-likelihood analysis for stochastic differential equations. Annals of the Institute of Statistical Mathematics, 63(3), 431–479.
Acknowledgements
On behalf of all authors, the corresponding author states that there is no conflict of interest. The author is grateful to the journal editor and two anonymous referees for careful reading and constructive comments, which led to improvements of this paper.
Funding
Open access funding provided by The University of Tokyo. This work was supported by Japan Society for the Promotion of Science (Grant no. 21H00997).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ogihara, T. Asymptotically efficient estimation for diffusion processes with nonsynchronous observations. Jpn J Stat Data Sci 6, 505–550 (2023). https://doi.org/10.1007/s42081-023-00196-0
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s42081-023-00196-0