Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

Ergodic diffusion processes like the Cox-Ingersoll & Ross [3], the geometric Ornstein-Uhlenbeck or Dixit & Pindyck [4] are widely used in the mathematical finance context, see [2] or [4].

Many times, for ergodic diffusions, the transition density is not known and the parameter estimation is made using approximations of the likelihood function based in some kind of discretization or using martingale estimating functions, see, for instance, [1, 5,6,7, 12]. In [10], a new parameter estimation technique was presented and applied to the stochastic processes satisfying the following stochastic differential equation,

$$\displaystyle \begin{aligned} dX_{t}=b(a-X_{t})X_{t}^{\gamma}dt+\sigma\sqrt{X_{t}^{\gamma+1}} dB_{t},\quad a,b,\sigma>0, \gamma\geq 0, \end{aligned} $$
(1)

and for the combination of parameters that makes this processes ergodic.

The idea was that if \(\left (X_t\right )_{t\geq 0}\) is an ergodic process then as t → the density function of X t converges to the invariant density and then the process parameters can be estimated from the invariant density as if the observations, X 1, …, X n, of the process were independent and identically distributed (i.i.d.) random variables, all of them with the same invariant distribution.

Also in [10], pseudo maximum likelihood estimators were deduced from the invariant density and in the present work we will compute pseudo moments estimators and their asymptotic properties will be studied. In the final section a comparative study, through simulation, will be implemented to compare the pseudo moments estimators with the pseudo maximum likelihood estimators already mentioned and also with the usual estimators obtained from discrete approximations of the transition density.

2 Ergodicity

A continuous time diffusion process

$$\displaystyle \begin{aligned} dX_{t}=\mu(X_{t};\boldsymbol{\theta})dt+\sigma(X_{t};\boldsymbol{\theta})dB_{t}, \end{aligned}$$

with state space \(\mathbb {R}\), is said to be ergodic (see, for instance, [9]), if

$$\displaystyle \begin{aligned} S(x;\boldsymbol{\theta})=\int_{x_{0}}^{x}\exp\left(-2\int_{x_{0}}^{y}\frac{\mu(v;\boldsymbol{\theta})}{\sigma^{2}(v;\boldsymbol{\theta})} dv\right)dy\to \pm\infty, \text{ as }x\to\pm\infty, \end{aligned}$$

and

$$\displaystyle \begin{aligned} M(\boldsymbol{\theta})=\int_{-\infty}^{+\infty}\frac{1}{\sigma^{2}(x;\boldsymbol{\theta})}\exp\left(2\int_{x_{0}}^{x} \frac{\mu(v;\boldsymbol{\theta})}{\sigma^{2}(v; \boldsymbol{\theta})}dv\right)dx<\infty, \end{aligned}$$

with x 0 an interior point of the state space.

The invariant density is then given by

$$\displaystyle \begin{aligned} f_{\boldsymbol{\theta}}(x)=\frac{1}{M(\boldsymbol{\theta})\sigma^{2}(x;\boldsymbol{\theta})}\exp\left(2\int_{x_{0}}^{x}\frac{\mu(v;\boldsymbol{\theta})}{\sigma^{2}(v;\boldsymbol{\theta})}dv\right). \end{aligned}$$

Theorem 1

The processes satisfying the stochastic differential equation (1) are ergodic, when 2ab > σ 2(γ + 1), with invariant density,

$$\displaystyle \begin{aligned} f_{(\alpha,\beta)}(x)=\frac{x^{\alpha-1}e^{-\beta x}\beta^{\alpha}}{\varGamma(\alpha)}\sim Gamma(\alpha,\beta),\mathit{\text{ with }} \alpha=\frac{2ab}{\sigma^{2}}-\gamma,\quad \beta=\frac{2b}{\sigma^{2}}. \end{aligned}$$

Proof

The processes have state space ]0, [ and with, θ = (a, b, γ, σ), we have

$$\displaystyle \begin{aligned}\begin{aligned} S(x;\boldsymbol{\theta})&=\int_{x_{0}}^{x} s(y,\boldsymbol{\theta})dy=\int_{x_{0}}^{x}\exp\left(-2\int_{x_{0}}^{y} \frac{b(a-v)v^{\gamma}}{\sigma^{2}v^{\gamma+1}}dv\right)dy\\ &=x_{0}^{\frac{2ab}{\sigma^{2}}}e^{-\frac{2b}{\sigma^{2}}x_{0}}\int_{x_{0}}^{x}y^{-\frac{2ab}{\sigma^{2}}}e^{\frac{2b}{\sigma^{2}}y} dy\rightarrow \pm\infty, x\rightarrow+\infty, x\rightarrow0, \end{aligned}\end{aligned}$$

and

$$\displaystyle \begin{aligned}\begin{aligned} M(\boldsymbol{\theta})&=\int_{0}^{+\infty}\frac{1}{\sigma^{2}x^{\gamma+1}}\exp\left(2\int_{x_{0}}^{x}\frac{b(a-v)v^{\gamma}}{\sigma^{2} v^{\gamma+1}}dv\right)dx\\ &=\frac{x_{0}^{-\frac{2ab}{\sigma^{2}}}e^{\frac{2b}{\sigma^{2}}x_{0}}}{\sigma^{2}}\int_{0}^{\infty}x^{\frac{2ab}{\sigma^{2}}-\gamma-1} e^{-\frac{2b}{\sigma^{2}}x}dx<\infty, \text{ if }2ab>\sigma^{2}(\gamma+1). \end{aligned}\end{aligned}$$

The invariant density is then given by

$$\displaystyle \begin{aligned}\begin{aligned} f_{\boldsymbol{\theta}}(x)&=\frac{x_{0}^{-\frac{2ab}{\sigma^{2}}}e^{\frac{2b}{\sigma^{2}}x_{0}}}{\sigma^{2}}x^{\frac{2ab}{\sigma^{2}}- \gamma-1}e^{-\frac{2b}{\sigma^{2}}x}\left(\frac{x_{0}^{-\frac{2ab}{\sigma^{2}}}e^{\frac{2b}{\sigma^{2}}x_{0}}}{\sigma^{2}}\int_{0}^{\infty} x^{\frac{2ab}{\sigma^{2}}-\gamma-1}e^{-\frac{2b}{\sigma^{2}}x}dx\right)^{-1}\\ &=\frac{x^{\alpha-1}e^{-\beta x}\beta^{\alpha}}{\varGamma(\alpha)}\sim Gamma(\alpha,\beta),\text{ with } \alpha=\frac{2ab}{\sigma^{2}}-\gamma, \; \beta=\frac{2b}{\sigma^{2}}, \end{aligned} \end{aligned}$$

completing the proof.

3 Estimators and Consistency

If we are working with a strictly stationary ergodic process (for instance, if X 0 have already the invariant distribution), then for any t > 0 the random variable X t will have the invariant distribution. In this framework we propose to deal with the observations of the process like if they were identically distributed with the invariant distribution and then use the invariant density for estimation purposes. From the previous section we know that the processes satisfying the stochastic differential equation (1) with 2ab > σ 2(γ + 1) are ergodic with the invariant density Gamma(α, β), where \(\alpha =\frac {2ab}{\sigma ^{2}}-\gamma , \beta =\frac {2b}{\sigma ^{2}}\).

In the following, let us suppose that we have observations X 1, …, X n of the process, collected at equally spaced times t 1 < … < t n and that γ and σ are known parameters, that is, the only parameters of interest for estimation purposes are a and b.

3.1 Pseudo Maximum Likelihood Estimators

We can compute pseudo maximum likelihood estimators, that is, defining the likelihood function

$$\displaystyle \begin{aligned} f_{X_1,\ldots,X_n}(\alpha,\beta;x_1,\ldots,x_n):=\prod_{i=1}^n f_{X_i}(\alpha,\beta;x_i) \end{aligned}$$

just like in the case of i.i.d. observations and where \(f_{X_i}\) is the Gamma(α, β) density of Eq. (1).

Since

$$\displaystyle \begin{aligned} \forall i=1,\ldots,n,\quad f_{X_i}(\alpha,\beta;x_i)=\frac{x_i^{\alpha-1}e^{-\beta x_i}\beta^{\alpha}}{\varGamma(\alpha)} \end{aligned}$$

we get the likelihood function,

$$\displaystyle \begin{aligned} L(\alpha,\beta;X_1,\ldots,X_n)=\prod_{i=1}^{n}\frac{X_i^{\alpha-1}e^{-\beta X_i}\beta^{\alpha}}{\varGamma(\alpha)} \end{aligned}$$

and the log-likelihood,

$$\displaystyle \begin{aligned} \log(L(\alpha,\beta;X_1,\ldots,X_n))=(\alpha-1)\sum_{i=1}^n\log(X_i)-\beta\sum_{i=1}^{n}X_i+n\alpha\log(\beta)-n\log(\varGamma(\alpha)). \end{aligned}$$

From differentiating the log-likelihood function and equating to zero, we get (with ψ(.) the digamma function),

$$\displaystyle \begin{aligned} \frac{1}{n}\sum_{i=1}^n\log(X_i)+\log\left(\frac{2b}{\sigma^2}\right)-\psi\left(\frac{2\bar{X}_nb}{\sigma^2}\right)=0 \end{aligned} $$
(2)

with \(\bar {X}_n=\frac {1}{n}\sum _{i=1}^{n}X_i\) and

$$\displaystyle \begin{aligned} a=\bar{X}_n+\frac{\sigma^2\gamma}{2b} \end{aligned}$$

getting the estimator \(\hat {b}_n\) of b as the solution of the Eq. (2) and the estimator of a, as

$$\displaystyle \begin{aligned} \hat{a}_n=\bar{X}_n+\frac{\sigma^2\gamma}{2\hat{b}_n}. \end{aligned}$$

Theorem 2

If 2ab > σ 2(γ + 1), the pseudo maximum likelihood estimators \(\hat {a}_n\) and \(\hat {b}_n\) are almost sure (a.s.) consistent estimators for a and b.

Proof

The proof of the theorem can be found in [10].

3.2 Pseudo Moments Estimators

We can obtain the moments estimators for a and b, using the invariant gamma density and by solving the equations,

$$\displaystyle \begin{aligned} \left\{ \begin{array}{ll} \bar{X}_n&=\frac{\alpha}{\beta} \\ M_{2,n}&=\frac{\alpha+\alpha^2}{\beta^2} \end{array}\right. , \end{aligned}$$

with \(\bar {X}_n=\frac {1}{n}\sum _{i=1}^{n}X_{i}\) the sample mean and \(M_{2,n}=\frac {1}{n}\sum _{i=1}^{n}X_{i}^2\) the empirical second moment.

Solving these equations we get the moments estimators for the parameters a and b,

$$\displaystyle \begin{aligned} \tilde{a}_{n}=\bar{X}_{n}+\frac{M_{2,n}-\bar{X}_n^2}{\bar{X}_{n}}\gamma \quad \wedge\quad \tilde{b}_{n}= \frac{\sigma^2\bar{X}_{n}}{2(M_{2,n}-\bar{X}_n^2)} \end{aligned}$$

or using the (non-central) sample variance \(S_n^2=\frac {1}{n}\sum _{i=1}^{n}(X_{i}-\bar {X}_{n})^2\),

$$\displaystyle \begin{aligned} \tilde{a}_{n}=\bar{X}_{n}+\frac{S_n^2}{\bar{X}_{n}}\gamma \quad \wedge\quad \tilde{b}_{n}=\frac{\sigma^2\bar{X}_{n}}{2S_n^2}. \end{aligned}$$

We have the following result about the consistency of the pseudo moments estimators.

Theorem 3

If 2ab > σ 2(γ + 1), the pseudo moments estimators \(\tilde {a}_n\) and \(\tilde {b}_n\) are a.s. consistent estimators for a and b.

Proof

Suppose that ξ is a random variable with the invariant gamma density Gamma(α, β), where \(\alpha =\frac {2ab}{\sigma ^{2}}-\gamma , \beta =\frac {2b}{\sigma ^{2}}\). It is straightforward to prove the consistency of both estimators, since, using the ergodic theorem, we have that

$$\displaystyle \begin{aligned} \lim_{n\to\infty}\bar{X}_{n}=\mathbb{E}[\xi]=\frac{\alpha_0}{\beta_0},\quad a.s. \end{aligned}$$

and

$$\displaystyle \begin{aligned} \lim_{n\to\infty}S_n^2=\lim_{n\to\infty}\frac{1}{n}\sum_{i=1}^{n}(X_{i}-\bar{X}_{n})^2= \mathbb{V}[\xi]=\frac{\alpha_0}{\beta_0^2},\quad a.s. \end{aligned}$$

Then,

$$\displaystyle \begin{aligned} \lim_{n\to\infty}\left(\bar{X}_{n}+\frac{S_n^2}{\bar{X}_{n}}\gamma\right)=\frac{\alpha_0}{\beta_0}+\frac{\gamma}{\beta_0}=a_0\quad a.s. \end{aligned}$$

and

$$\displaystyle \begin{aligned} \lim_{n\to\infty}\frac{\sigma^2\bar{X}_{n}}{2S_n^2}=\frac{\beta_0\sigma^2}{2}=b_0\quad a.s. \end{aligned}$$

proving the consistency of the estimators.

Remark 1

We have assumed that σ is known, if σ is unknown the problem of estimating σ can be solved using the quadratic variation of the process, and following [11] we get, a estimator for σ 2,

$$\displaystyle \begin{aligned} \hat{\sigma}_{1,n}^{2}=\frac{\sum_{i=1}^{n-1}(X_{i+1}-X_{i})^{2}}{\sum_{i=1}^{n-1}X_{i}^{\gamma+1}\varDelta_n}, \end{aligned}$$

or following [12]

$$\displaystyle \begin{aligned} \hat{\sigma}_{2,n}^{2}=\frac{1}{T}\sum_{i=1}^{n-1}\frac{(X_{i+1}-X_{i})^{2}}{X_{i}^{\gamma+1}}. \end{aligned}$$

4 Simulation and Data Analysis

In this section, we will compare through simulation the moments estimators with the approximate maximum likelihood estimators presented in [10] and the estimators based in discrete approximations of the log-likelihood function. We will suppose that the observations are equally spaced, that is, t i+1 − t i = Δ, i = 1, …, n.

The estimators for a and b based on the discretized continuous-time likelihood function, see [1] or [7], \(\check {a}_n\) and \(\check {b}_n\), are given by:

$$\displaystyle \begin{aligned} \check{a}_n=\frac{\sum_{i=1}^{n-1}\frac{X_{i+1}-X_{i}}{X_{i}}\sum_{i=1}^{n-1}X_{i}^{\gamma+1}-(X_{n}-X_{1})\sum_{i=1}^{n-1}X_{i}^{\gamma}} {\sum_{i=1}^{n-1}\frac{X_{i+1}-X_{i}}{X_{i}}\sum_{i=1}^{n-1}X_{i}^{\gamma}-(X_{n}-X_{1})\sum_{i=1}^{n-1}X_{i}^{\gamma-1}} \end{aligned}$$

and

$$\displaystyle \begin{aligned} \check{b}_n=\frac{1}{\varDelta}\frac{\sum_{i=1}^{n-1}\frac{X_{i+1}-X_{i}}{X_{i}}\sum_{i=1}^{n-1}X_{i}^{\gamma}-(X_{n}-X_{1})\sum_{i=1}^{n-1} X_{i}^{\gamma-1}}{\sum_{i=1}^{n-1}X_{i}^{\gamma-1}\sum_{i=1}^{n-1}X_{i}^{\gamma+1}-\left(\sum_{i=1}^{n-1}X_{i}^{\gamma}\right)^2}. \end{aligned}$$

For simulation purposes, we will perform the generation of the trajectories of the processes using the approximation strong Taylor scheme of order 1.5, see [8].

The iterative scheme used is the following:

$$\displaystyle \begin{aligned} \begin{aligned} Y_{i+1}=\,&Y_i+b(a-Y_i)Y_i^\gamma\varDelta+\sigma Y_i^{\frac{\gamma+1}{2}}\varDelta B\\ &+\frac{\sigma^2(\gamma+1)}{4}Y_i^\gamma((\varDelta B)^2-\varDelta)+\sigma b(\gamma(a-Y_i)-Y_i)Y_i^{\frac{3\gamma-1}{2}}\varDelta Z\\ &+\frac{1}{2}\left(b^2(a-Y_i)(\gamma(a-Y_i)-Y_i)+\frac{1}{2}b\gamma\sigma^2(\gamma(a-Y_i)-a-Y_i)\right)Y_i^{2\gamma-1}\varDelta^2\\ &+\left(\frac{\sigma b}{2}(a-Y_i)+\frac{\sigma^3(\gamma-1)}{8}\right)(\gamma+1)Y_i^{\frac{3\gamma-1}{2}}(\varDelta B\varDelta-\varDelta Z)\\ &+\frac{\sigma^3\gamma(\gamma+1)}{4}Y_i^{\frac{3\gamma-1}{2}}\left(\frac{1}{3}(\varDelta B)^3-\varDelta\right)\varDelta B, \end{aligned} \end{aligned}$$

where \(\varDelta B=\sqrt {\varDelta }U_1\), \(\varDelta Z=\frac {1}{2}\varDelta ^{3/2}(U_1+U_2/\sqrt {3})\) and U 1 and U 2 are independent N(0, 1) random variables.

We simulated 500 trajectories and for the estimation of the parameter a we present the results when n = 500 in each trajectory, for the parameter b we considered n = 250, 500, and 1000 observations in each trajectory. We estimated a and b using the pseudo moments estimators \(\tilde {a}_n\) and \(\tilde {b}_n\) and we compared them with the pseudo maximum likelihood estimators \(\hat {a}_n\) and \(\hat {b}_n\) and the estimators \(\check {a}_n\) and \(\check {b}_n\) obtained from the discretized likelihood function.

We considered σ = 0.1, we present Table 1 for γ = 0 and Table 2 for γ = 1 (for other values of γ we get very similar results), the true value for a is always 1 and for b we considered the values 0.1, 0.5, 1, and 2.

Table 1 Mean and S.D. (standard deviation) for the estimators of a and b when γ = 0
Table 2 Mean and S.D. (standard deviation) for the estimators of a and b when γ = 1

In all the outputs, we can see that the proposed estimators for a and b give good results, very close to the approximate maximum likelihood estimators and we can also see that the estimator for b, \(\check {b}_n\) based in the score function only produce good results when the true value of b is 0.1 (small).

5 Conclusion

In this paper we proposed moments estimators for some ergodic processes. The consistency proof of the proposed estimators and a simulation study to show the applicability of the estimators were provided. For future research, we have the open problem of proving the normality of the asymptotic distribution of the estimators.