Abstract
We prove a complete class theorem that characterizes all stationary time reversible Markov processes whose finite dimensional marginal distributions (of all orders) are infinitely divisible. Aside from two degenerate cases (iid and constant), in both discrete and continuous time every such process with full support is a branching process with Poisson or Negative Binomial marginal univariate distributions and a specific bivariate distribution at pairs of times. As a corollary, we prove that every nondegenerate stationary integer valued process constructed by the Markov thinning process fails to have infinitely divisible multivariate marginal distributions, except for the Poisson. These results offer guidance to anyone modeling integer-valued Markov data exhibiting autocorrelation.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Many applications feature autocorrelated count data \(X_t\) at discrete times t. A number of authors have constructed and studied stationary stochastic processes \(X_t\) whose one-dimensional marginal distributions come from an arbitrary infinitely-divisible distribution family \(\{\mu ^{\theta }\}\), such as the Poisson \(\textsf {Po}(\theta )\) or negative binomial \(\textsf {NB}(\theta ,p)\), and that are “\(\textsf {AR}(1)\)-like” in the sense that their autocorrelation function is \({\textsf {Corr}}[X_s,X_t] =\rho ^{|s-t|}\) for some \(\rho \in (0,1)\) (Lewis, 1983; Lewis et al., 1989; McKenzie, 1988; Al-Osh and Alzaid, 1987; Joe, 1996). The most common approach is to build a time-reversible Markov process using thinning, in which the process at any two consecutive times may be written in the form
with \(\xi _t\), \(\eta _t\), and \(\zeta _t\) all independent and from the same infinitely-divisible family (see Sec. () below for details). A second construction of a stationary time-reversible process with the same one-dimensional marginal distributions and autocorrelation function, with the feature that its finite-dimensional marginal distributions of all orders are infinitely-divisible, is to set \(X_t:={\mathcal {N}}(G_t)\) for a random measure \({\mathcal {N}}\) on some measure space \((E,{\mathcal {E}},m)\) that assigns independent infinitely-divisible random variables \({\mathcal {N}}(A_i)\sim \mu ^{\theta _i}\) to disjoint sets \(A_i\in {\mathcal {E}}\) of measure \(\theta _i=m(A_i)\), and a family of sets \(\{G_t\}\subset {\mathcal {E}}\) whose intersections have measure \(m\big (G_s\cap G_t\big )=\theta \rho ^{|s-t|}\) (see Sec. ()).
For the normal distribution \(X_t\sim \textsf {No}(\mu ,\sigma ^2)\), these two constructions both yield the usual Gaussian \(\textsf {AR}(1)\) process. The two constructions also yield identical processes for the Poisson \(X_t\sim \textsf {Po}(\theta )\) distribution, but they differ for all other nonnegative integer-valued infinitely-divisible distributions. For each nonnegative integer-valued infinitely-divisible marginal distribution except the Poisson, the process constructed by thinning does not have infinitely-divisible marginal distributions of all orders (Theorem , Sec. ()), and the process constructed using random measures does not have the Markov property (Theorem , Sec. 3.5). Thus none of these is completely satisfactory for modeling autocorrelated count data with heavier tails than the Poisson distribution.
In the present manuscript we construct and characterize every process that is Markov, infinitely-divisible, stationary, and time-reversible with non-negative integer values. The formal characterization is contained in the statement of Theorem in Sec. (), which follows necessary definitions and the investigation of special cases needed to establish the general result.
1.1 Thinning Process
Any univariate infinitely-divisible (ID) distribution \(\mu (dx)\) on \({\mathbb {R}}^1\) is \(\mu ^1\) the for a convolution semigroup \(\{\mu ^\theta :~\theta \ge 0\}\) and, for \(0<\theta <\infty \) and \(0<\rho <1\), determines uniquely a “thinning distribution” \(\mu ^\theta _\rho (dy\mid x)\) of Y conditional on the sum \(X=Y+Z\) of independent \(Y\sim \mu ^{\rho \theta }\) and \(Z\sim \mu ^{(1{-}\rho )\theta }\). This thinning distribution determines a unique stationary time-reversible Markov process with one-step transition probability distribution given by the convolution
for Borel sets \(A\subset {\mathbb {R}}\), where \({\mathcal {F}}_t=\sigma \{X_s:~s\le t\}\) is the minimal filtration. By induction the autocorrelation is \({\textsf {Corr}}(X_s,X_t) =\rho ^{|s-t|}\) for square-integrable \(X_t\). The process can be constructed beginning at any \(t_0\in {\mathbb {Z}}\) by setting
Time-reversibility and hence the lack of dependence of this definition on the choice of \(t_0\) follows from the argument presented in the proof of Theorem in Sec. () below.
1.1.1 Thinning Example 1: Poisson
For Poisson-distributed \(X_t\sim \mu ^\theta =\textsf {Po}(\theta )\) with mean \(\theta >0\), for example, the thinning recursion step for \(0<\rho <1\) and \(t>t_0\) can be written
and hence the joint generating function at two consecutive times is
This was called the “Poisson \(\textrm{AR}(1)\) Process” by McKenzie (McKenzie, 1985) and has been studied by many other authors since its introduction.
1.1.2 Thinning Example 2: Negative Binomial
In the thinning process applied to the Negative Binomial \(X_t\sim \mu ^\theta =\textsf {NB}(\theta ,p)\) distribution with mean \(\theta (1{-}p)/p\), recursion for \(t>t_0\) takes the form
for beta-binomial distributed \(\xi _t \sim {\textsf{BB}}(n;\alpha ,\beta )\) (Johnson et al., 2005, §2.2) with \(n=X_{t-1}\), \(\alpha =\theta \rho \), and \(\beta = \theta (1{-}\rho )\), and negative binomial \(\zeta _t \sim \textsf {NB}\big (\theta (1{-}\rho ),p\big )\). Thus the joint generating function is
From this one can compute the conditional generating function
where \(_2F_1(a,b;c;z)\) denotes Gauss’ hypergeometric function (Abramowitz and Stegun, 1964, §15) and, from this (for comparison below),
This process, as we will see below in Theorem , is Markov, stationary, and time-reversible, with infinitely-divisible one-dimensional marginal distributions \(X_t\sim \textsf {NB}(\theta ,p)\), but the joint marginal distributions at three or more consecutive times are not ID. It appears to have been introduced by Joe (Joe, 1996, p. 665).
1.2 Random Measure Process
Another approach to the construction of processes with specified univariate marginal stationary distribution \(\mu ^\theta (dx)\) is to set \(X_t := {\mathcal {N}}(G_t)\) for a random measure \({\mathcal {N}}\) and a class of sets \(\left\{ G_t\right\} \), as in (Wolpert and Taqqu, 2005, §3.3, 4.4). We begin with a countably additive random measure \({\mathcal {N}}(dx\,dy)\) that assigns independent random variables \({\mathcal {N}}(A_i)\sim \mu ^{|A_i|}\) to disjoint Borel sets \(A_i\in {\mathcal {B}} ({\mathbb {R}}^2)\) of finite area \(|A_i|\) (this is possible by the Kolmogorov consistency conditions), and a collection of sets
(shown in Fig. ()) whose intersections satisfy \(|G_s\cap G_t| = \theta e^{-\lambda |s-t|}\). For \(t\in {\mathbb {Z}}\), set
For any n times \(t_1<t_2<\dots <t_n\) the sets \(\left\{ G_{t_i}\right\} \) partition \({\mathbb {R}}^2\) into \(n(n+1)/2\) sets of finite area (and one with infinite area, \((\cup G_{t_i})^c\)), so each \(X_{t_i}\) can be written as the sum of some subset of \(n(n+1)/2\) independent random variables. In particular, any \(n=2\) variables \(X_s\) and \(X_t\) can be written as
just as in the thinning approach, so both 1-dimensional and 2-dimensional marginal distributions for the random measure process coincide with those for the thinning process of Sec. ().
Evidently the process \(X_t\) constructed from this random measure is stationary, time-reversible and infinitely divisible in the strong sense that all finite-dimensional marginal distributions are ID. Although the 1- and 2-dimensional marginal distributions of this process coincide with those of the thinning process, the k-dimensional marginals may differ for \(k\ge 3\), so this process cannot be Markov. We will see in Theorem below that the only nonnegative integer-valued distribution for which it is Markov is the Poisson.
1.2.1 Random Measure Example 1: Poisson
The conditional distribution of \(X_{t_n}={\mathcal {N}}(G_{t_n})\) given \(\{X_{t_j}: ~j<n\}\) can be written as the sum of n independent terms, \((n-1)\) of them with binomial distributions (all with the same probability parameter \(p=\rho ^{|t_n-t_{n-1}|}\), and with size parameters that sum to \(X_{t_{n-1}}\)) and one with a Poisson distribution (with mean \(\theta (1-\rho ^ {|t_n-t_{n-1}|}\)). It follows by induction that the random-measure Poisson process is identical in distribution to the thinning Poisson process of Sec. ().
1.2.2 Random Measure Example 2: Negative Binomial
The random variables \(X_1\), \(X_2\), \(X_3\) for the random measure process built on the Negative Binomial distribution \(X_t\sim \textsf {NB}(\theta ,p)\) with autocorrelation \(\rho \in (0,1)\) can be written as sums
of six independent negative binomial random variables \(\zeta _s\sim \textsf {NB}(\theta _s, p)\) with shape parameters
(each \(\zeta _s={\mathcal {N}}\big (\cap _{t\in s} G_t\big )\) and \(\theta _s=|\cap _{t\in s} G_t|\) in Fig. ()). It follows that the conditional probability
differs from that of the thinning negative binomial process in Eqn. () for all \(\theta >0\) and \(\rho >0\). Thus this process is stationary, time-reversible, and has infinitely-divisible marginal distributions of all orders, but it cannot be Markov since its 2-dimensional marginal distributions coincide with those of the Markov thinning process but its 3-dimensional marginal distributions do not.
In Sec. () of this paper we characterize every discrete-time process that is Markov, Infinitely-divisible, Stationary, and Time-reversible with non-negative Integer values (MISTI for short). In Sec. () we first present the necessary definitions and preliminary results; in Sec. () we extend the results to continuous time, with discussion in Sec. ().
2 MISTI Processes
A real-valued stochastic process \(X_t\) indexed by \(t\in {\mathbb {Z}}\) is stationary if each finite-dimensional marginal distribution
satisfies
for each set \(T\subset {\mathbb {Z}}\) of finite cardinality \(|T|<\infty \), Borel set \(B\in {\mathcal {B}} ({\mathbb {R}}^{|T|})\), and \(s\in {\mathbb {Z}}\), where as usual “\(s+T\)” denotes \(\{(s+t):~t\in T\}\). A stationary process is time-reversible if also
(where “\(-T\)” is \(\{-t:~t\in T\}\)) and Markov if for every \(t\in {\mathbb {Z}}\) and finite \(T\subset \{s\in {\mathbb {Z}}:~s\ge t\}\),
for all \(B\in {\mathcal {B}} ({\mathbb {R}}^{|T|})\), where \({\mathcal {F}}_t:=\sigma \{X_s:~s\le t\}\). The process \(X_t\) is Infinitely Divisible (ID) or, more specifically, multivariate infinitely divisible (MVID) if each \({\mu _{\scriptscriptstyle T}}\) is the n-fold convolution of some other distribution \(\mu _{\scriptscriptstyle T}^{(1/n)}\) for each \(n\in {\mathbb {N}}\). This is more restrictive than requiring only that the one-dimensional marginal distributions be ID and, for integer-valued processes that satisfy
it is equivalent by the Lévy-Khinchine formula (Rogers and Williams, 2000, p. 74) to the condition that each \({\mu _{\scriptscriptstyle T}}\) have characteristic function of the form
for some finite measure \(\nu _{\scriptscriptstyle T}\) on \({\mathcal {B}} ({\mathbb {Z}}^{|T|})\). Call a process \(X_t\) or its distributions \({\mu _{\scriptscriptstyle T}}(du)\) MISTI if it is Markov, nonnegative Integer-valued, Stationary, Time-reversible, and Infinitely divisible, i.e., satisfies Eqns. (–). We now turn to the problem of characterizing all MISTI distributions.
2.1 Three-dimensional Marginals
By stationarity and the Markov property all MISTI finite-dimensional distributions \({\mu _{\scriptscriptstyle T}}(du)\) are determined completely by the marginal distribution for \(X_t\) at two consecutive times; to exploit the MVID property we will study the three-dimensional marginal distribution for \(X_t\) at any set T of \({|T|}=3\) consecutive times— say, \(T=\{1,2,3\}\). By Eqn. () we can represent \(X_{\{1,2,3\}}\) in the form
for independent Poisson-distributed random variables
with means \({\lambda _{\,ijk}} := \nu (\{(i,j,k)\})\); here and hereafter, a subscript “\(+\)” indicates summation over the entire range of that index— \({\mathbb {N}}_0=\left\{ 0,2\dots \right\} \) for \(\left\{ {N_{ijk}}\right\} \) and \(\left\{ {\lambda _{\,ijk}}\right\} \), \({\mathbb {N}}=\left\{ 1,2,\dots \right\} \) for \(\left\{ {\theta _{j}}\right\} \). The sums \({\theta _{j}}:={\lambda _{\,+j+}}\) for \(j\ge 1\) characterize the univariate marginal distribution of each \(X_t\)— for example, through the probability generating function (pgf)
To avoid trivial technicalities we will assume that \(0 < {\textsf {P}}[X_t=1]=\varphi '(0) ={\theta _{1}}e^{-{\theta _{+}}}\), i.e., \({\theta _{1}}>0\). Now set \({r_{i}}:={\lambda _{\,i1+}}/{\theta _{1}}\), and for later use define functions:
Since \({r_{i}}\) and \({\theta _{j}}\) are nonnegative and summable (by Eqns. (, )), \(p({s})\) and \(P({z})\) are analytic on the open unit ball \({\mathbb {U}}\subset {\mathbb {C}}\) and continuous on its closure. Similarly, since \({\lambda _{\,ijk}}\) is summable, each \(\psi _j({s},{t})\) is analytic on \({\mathbb {U}}^2\) and continuous on its closure. Note \(\psi _j(1,1)={\theta _{j}}\), \(p(0)={r_{0}}\) and \(p(1)=1\), while \(P(0)=0\) and \(P(1)={\theta _{+}}\); also \(\varphi ({z}) = \exp \left\{ P({z})-{\theta _{+}}\right\} \). Each \(\psi _j(s,t)=\psi _j(t,s)\) is symmetric by Eqn. (), as are the conditional probability generating functions:
2.1.1 Conditioning on \(X_2=0\)
By the Markov property Eqn. (), \(X_1\) and \(X_3\) must be conditionally independent given \(X_2\), so the conditional probability generating function must factor:
Taking logarithms,
or, for all \({s}\) and \({t}\) in the unit ball in \({\mathbb {C}}\),
Thus \({\lambda _{\,i0k}}=0\) whenever both \(i>0\) and \(k>0\) and, by symmetry,
2.1.2 Conditioning on \(X_2=1\)
Similarly
since \(\left\{ {N_{i1k}}\right\} \) is conditionally multinomial given \({N_{+1+}}\) and independent of \(\left\{ {N_{i0k}}\right\} \). By the Markov property this too must factor, as \(\varphi ({s},{t}\mid 1)=\varphi ({s},1\mid 1) \,\varphi (1,{t}\mid 1)\), so by Eqn. ()
or, since \({\lambda _{\,i1k}}={\lambda _{\,k1i}}\) by Eqns. (, ),
2.1.3 Conditioning on \(X_2=2\)
The event \(\left\{ X_2=2\right\} \) for \(X_2:=\sum _{j\ge 1}j{N_{+j+}}\) can happen in two ways: either \({N_{+1+}} =2\) and each \({N_{+j+}}=0\) for \(j\ge 2\), or \({N_{+2+}}=1\) and \({N_{+j+}}=0\) for \(j=1\) and \(j\ge 3\), with \({N_{+0+}}\) unrestricted in each case. These two events have probabilities \(({\theta _{1}}^2/2)e^{-{\theta _{+}}}\) and \(({\theta _{2}})e^{-{\theta _{+}}}\), respectively, so the joint generating function for \(\{X_1,X_3\}\) given \(X_2=2\) is
In view of Eqn. (), this will factor in the form \(\varphi ({s},{t}\mid 2)= \varphi ({s},1\mid 2)\, \varphi (1,{t}\mid 2)\) as required by Markov property Eqn. () if and only if for all s, t in the unit ball:
or
To satisfy the ID requirement of Eqn. (), this must hold with each \({\theta _{j}}\) replaced by \({\theta _{j}}/n\) for each integer \(n\in {\mathbb {N}}\). Since the left and right sides are homogeneous in \(\theta \) of degrees 3 and 2 respectively, this will only happen if each square-bracketed term vanishes identically, i.e., if
and
so
2.1.4 Conditioning on \(X_2=j\)
The same argument applied recursively, using the Markov property for each \(j\ge 1\) in succession, leads to:
so
and consequently
Conditionally on \(\left\{ X_2=j\right\} \), \(X_1\) and \(X_3\) are distributed independently, each as the sum of j independent random variables with generating function p(s), plus one with generating function \(\varphi (s,1\mid 0)\)— so \(X_t\) is a branching process (Harris, 1963) whose unconditional three-dimensional marginal distributions have generating function:
See Secs. and for further development of this branching process representation.
2.2 Stationarity
Without loss of generality we may take \({\lambda _{\,000}} = 0\). By Eqn. () with \({s}=0\) and \({t}=1\) we have \({\lambda _{\,0j+}}={\theta _{j}}{r_{0}}^j\); by Eqn. () we have \({\lambda _{\,i00}} = {\lambda _{\,i0+}}\). By time-reversibility we conclude that \({\lambda _{\,i00}} = 0\) for \(i=0\) and, for \(i\ge 1\),
Now we can evaluate
and, from this and Eqn. (), evaluate the joint generating function for \(X_{\{1,2,3\}}\) as:
and so that for \(X_{\{1,2\}}\) as:
Now consider Eqn. () with \({t}=1\),
It follows first for \(j=1\) and then for \(i=1\) that
so again by time reversibility with \(i=j\), since \({\theta _{1}}>0\), we have
Thus \({r_{0}}\), \({r_{1}}\), and \(\{{\theta _{j}}\}\) determine all the \(\{{r_{j}}\}\) and so all the \(\{{\lambda _{\,ijk}}\}\) by Eqns. (, ) and hence the joint distribution of \(\{X_t\}\).
Now consider Eqn. () first for \(j=2\) and then \(i=2\):
Equating these for \(i=j\ge 2\) (by time-reversibility) and applying Eqn. () for \(0< k<i\) (the cases \(k=0\) and \(k=i\) need to be handled separately),
3 The Solutions
Eqn. () holds for all \(i\ge 2\) if \({r_{0}}=0\) or \({r_{1}}=0\), leaving \({r_{j}}=0\) by Eqn. () for all \(j\ge 2\), hence \({r_{0}}+{r_{1}}=1\) and \(\left\{ {\theta _{j}}\right\} \) is restricted only by the conditions \({\theta _{1}}>0\) and \({\theta _{+}}<\infty \).
3.1 The Constant Case
The case \({r_{0}}=0\) leads to \({r_{1}}=1\) and \({r_{j}}=0\) for all \(j\ne 1\), so \(p(z)\equiv z\). By Eqn. () the joint pgf is
so \(X_1=X_2=X_3\) and all \(\{X_t\}\) are identical, with an arbitrary ID distribution.
3.2 The IID Case
The case \({r_{1}}=0\) leads to \({r_{0}}=1\) and \({r_{j}}=0\) for all \(j\ne 0\) so \(p(z)\equiv 1\) and
by Eqn. (), making all \(\{X_t\}\) independent, with identical but arbitrary ID distributions.
3.3 The Poisson Case
Aside from these two degenerate cases, we may assume \({r_{0}}>0\) and \({r_{1}}>0\), and (by Eqn. ()) rewrite Eqn. () in the form:
whose unique solution for all integers \(i\ge 1\) (by induction) is
If \({r_{2}}=0\), then again \({r_{i}}=0\) for all \(i\ge 2\) but, by Eqn. (), \({\theta _{j}}=0\) for all \(j\ge 2\); thus \(P(z)={\theta _{1}} z\) so each \(X_t\sim \textsf {Po}({\theta _{1}})\) has a Poisson marginal distribution with mean \({\theta _{1}}={\theta _{+}}\). In this case \({r_{0}}+{r_{1}}=1\), \(p(z)={r_{0}}+{r_{1}}z\), and the two-dimensional marginals (by Eqn. ()) of \(X_1\), \(X_2\) have joint pgf
the bivariate Poisson distribution (Johnson et al., 1997, § 37.2), so \(X_t\) is the familiar “Poisson \(\textrm{AR}(1)\) Process” of McKenzie (McKenzie, 1985; McKenzie, 1988) (with autocorrelation \(\rho ={r_{1}}\)) considered in Sec. (). Its connection with Markov branching processes was recognized earlier (Steutel et al., 1983). By Eqn. () the conditional distribution of \(X_{t+1}\), given \({\mathcal {F}}_t :=\sigma \left\{ X_s:~s\le t\right\} \), is that of the sum of \(X_t\) independent Bernoulli random variables with pgf p(s) and a Poisson innovation term with pgf \(\exp \{P({r_{0}} s)-P({r_{0}})\}\), so the Markov process \(X_t\) may be written recursively starting at any \(t_0\) as
(all independent) for \(t> t_0\), the thinning construction of Sec. ().
3.4 The Negative Binomial case
Finally if \({r_{0}}>0\), \({r_{1}}>0\), and \({r_{2}}>0\), then (by Eqn. ()) \({r_{i}}= {r_{1}}(q{r_{0}})^{i-1}\) for \(i\ge 1\) and hence (by Eqn. ()) \({\theta _{j}}=\alpha q^j/j\) for \(j\ge 1\) with \(q:=(1-{r_{0}}-{r_{1}})/{r_{0}}(1-{r_{0}})\) and \(\alpha := {\theta _{1}}/q\). The condition \({\theta _{+}}<\infty \) entails \(q<1\) and \({\theta _{+}}=-\alpha \log (1{-}q)\). The 1-marginal distribution is \(X_t\sim \textsf {NB}(\alpha ,p)\) with \(p:=(1{-}q)\), and the functions \(P(\cdot )\) and \(p(\cdot )\) are \(P(z)=-\alpha \log (1-qz)\), \(p(s)={r_{0}}+{r_{1}} s/(1-q{r_{0}} s)\), so the joint pgf for the 2-marginal distribution of \(X_1,X_2\) is
with one-step autocorrelation \(\rho :=(1{-}{r_{0}})^2/{r_{1}}\). This bivariate distribution was introduced as the “compound correlated bivariate Poisson”(Edwards and Gurland, 1961), but we prefer to call it the Branching Negative Binomial distribution. In the branching formulation \(X_{t}\) may be viewed as the sum of \(X_{t-1}\) iid random variables with pgf \(p(s)={r_{0}} +{r_{1}} s/(1-q{r_{0}} s)\) and one with pgf \(\exp \left\{ P(s{r_{0}})-P({r_{0}})\right\} =(1-q{r_{0}})^\alpha (1-q{r_{0}}\,s)^{-\alpha }\). The first of these may be viewed as \(Y_t\) plus a random variable with the \(\textsf {NB}(Y_t,1{-}q{r_{0}})\) distribution, for \(Y_t\sim \textsf {Bi}(X_{t-1},1-{r_{0}})\), and the second has the \(\textsf {NB}(\alpha ,1{-}q{r_{0}})\) distribution, so a recursive updating scheme beginning with \(X_{t_0} \sim \textsf {NB}(\alpha ,p)\) is:
In the special case of \(\rho =q\) the joint pgf simplifies to \(\varphi ({s},{z}) = p^\alpha [1+q(1-{s}-{z})]^{-\alpha }\) and the joint distribution of \(X_1,X_2\) reduces to the negative trinomial distribution (Johnson et al., 1997, Ch. 36) with pmf
and simple recursion \(X_t\mid X_{t-1}\sim \textsf {NB}\big (\alpha +X_{t-1}, ~\frac{1}{1+q}\big )\).
3.5 Results
We have just proved:
Theorem 1
Let \(\left\{ X_t\right\} \) be a Markov process indexed by \(t\in {\mathbb {Z}}\) taking values in the non-negative integers \({\mathbb {N}}_0\) that is stationary, time-reversible, has infinitely-divisible marginal distributions of all finite orders, and satisfies \({\textsf {P}}[X_t=1]>0\). Then \(\left\{ X_t\right\} \) is one of four processes:
-
1.
\(X_t\equiv X_0\sim \mu _0(dx)\) for an arbitrary ID distribution \(\mu _0\) on \({\mathbb {N}}_0\) with \(\mu _0(\{1\})>0\);
-
2.
\(X_t\mathrel {\mathop {\sim }\limits ^{\textrm{iid}}}\mu _0(dx)\) for an arbitrary ID distribution \(\mu _0\) on \({\mathbb {N}}_0\) with \(\mu _0(\{1\})>0\);
-
3.
For some \(\theta >0\) and \(0<\rho <1\), \(X_t\sim \textsf {Po}(\theta )\) with bivariate joint generating function
$$ {\textsf {E}}\left[ {s}^{X_1}~{z}^{X_2}\right] = \exp \left\{ \theta (1{-}\rho )({s}-1) + \theta (1{-}\rho )({z}-1) + \theta \rho ({s}{z}-1) \right\} $$and hence autocorrelation \({\textsf {Corr}}(X_s,X_t)= \rho ^{|s-t|}\) and recursive update
$$ X_t = \xi _t+\zeta _t,\text {\quad where } \xi _t \sim \textsf {Bi}(X_{t-1}, ~\rho )\text { and } \zeta _t\sim \textsf {Po}\big (\theta (1{-}\rho )); $$ -
4.
For some \(\alpha >0\), \(0<p<1\), and \(0<\rho <1\), \(X_t\sim \textsf {NB}(\alpha ,p)\), with bivariate joint generating function
$$ {\textsf {E}}\left[ {s}^{X_1}~{z}^{X_2}\right] = p^{2\alpha }[(1-q\rho )-q(1{-}\rho )(s+z)+q(q-\rho )sz]^{-\alpha } $$where \(q=1{-}p\), and hence autocorrelation \({\textsf {Corr}}(X_s,X_t)= \rho ^{|s-t|}\) and recursive update
$$ X_t = Y_t+\zeta _t,\text {\quad where } Y_t \sim \textsf {Bi}\big (X_{t-1}, ~\rho \,p/(1-\rho q)\big )\text { and } \zeta _t\sim \textsf {NB}\big (\alpha +Y_t,~p/(1-\rho q)\big ). $$
Note the limiting cases of autocorrelation \(\rho =1\) and \(\rho =0\) in cases 3., 4. are subsumed by the degenerate cases 1. and 2., respectively. The theorem follows from this.
From this theorem follows:
Theorem 2
Let \(\left\{ \mu ^\theta :~\theta \ge 0\right\} \) be an ID semigroup of probability distributions on the nonnegative integers \({\mathbb {N}}_0\) with \(\mu ^\theta (\{1\})>0\). Fix \(\theta >0\) and \(0<\rho <1\) and let \(\left\{ X_t\right\} \) be the “thinning process” of Eqn. (1) in Sec. () with the representation
for each \(t\in {\mathbb {Z}}\) with independent
Then \(X_t\) is Markov, stationary, time-reversible, and nonnegative integer valued, but it does not have infinitely-divisible marginal distributions of all orders unless \(\{\mu ^\theta \}\) is the Poisson family.
Proof
By construction \(X_t\) is obviously Markov and stationary. The joint distribution of the process at consecutive times is symmetric (see Eqn. ()) since the marginal and conditional pmfs
of \(X_t\) and \(X_t\mid X_{t-1}\) satisfy the symmetric relation
Applying this inductively, for any \(s<t\) and any \(\{x_s,\cdots ,x_t\}\subset {\mathbb {N}}_0\) we find
and so the distribution of \(X_t\) is time-reversible. Now suppose that it is also ID. Then by Theorem it must be one of the four specified processes: constant, iid, branching Poisson, or branching negative binomial.
Since \(\rho <1\) it cannot be the constant \(\left\{ X_t\equiv X_0\right\} \) process; since \(\rho >0\) it cannot be the independent \(\left\{ X_t\mathrel {\mathop {\sim }\limits ^{\textrm{iid}}}\mu ^\theta (dx)\right\} \) process. The joint generating function \(\phi (s,z)\) at two consecutive times for the negative binomial thinning process, given in Eqn. (), differs from that for the negative binomial branching process, given in Eqn. (). The only remaining option is the Poisson branching process of Sec. (). \(\square \)
Theorem 3
Let \(\left\{ \mu ^\theta :~\theta \ge 0\right\} \) be an ID semigroup of probability distributions on the nonnegative integers \({\mathbb {N}}_0\) with \(\mu ^\theta (\{1\})>0\). Fix \(\theta >0\) and \(0<\rho <1\) and let \(\left\{ X_t\right\} \) be the “random measure process” of Eqn. () in Sec. (). Then \(X_t\) is ID, stationary, time-reversible, and nonnegative integer valued, but it is not a Markov process unless \(\{\mu ^\theta \}\) is the Poisson family.
Proof
By construction \(X_t\) is ID, stationary, and time-reversible; suppose that it is also Markov. Then by Theorem it must be one of the four specified processes: constant, iid, branching Poisson, or branching negative binomial.
Since \(\rho <1\) it cannot be the constant \(\left\{ X_t\equiv X_0\right\} \) process; since \(\rho >0\) it cannot be the independent \(\left\{ X_t\mathrel {\mathop {\sim }\limits ^{\textrm{iid}}}\mu ^\theta (dx)\right\} \) process. The joint generating function \(\phi (s,z)\) at two consecutive times for the negative binomial random measure process coincides with that for the negative binomial thinning process, given in Eqn. (), and differs from that for the negative binomial branching process, given in Eqn. (). The only remaining option is the Poisson branching process of Sec. (). \(\square \)
4 Continuous Time
Now consider a nontrivial \({\mathbb {N}}_0\)-valued time-reversible stationary Markov processes \(X_t\) indexed by continuous time \(t\in {\mathbb {R}}\). The restriction of any such process to integer times \(t\in {\mathbb {Z}}\) will still be Markov, hence MISTI, so there are but two possibilities for the restriction of \(X_t\) to \(t\in {\mathbb {Z}}\): one with univariate Poisson marginal distributions, and one with univariate Negative Binomial distributions. Both do in fact exist, as shown below in Secs. and . For any positive integer \(q\in {\mathbb {N}}\) the restriction of such a process \(X_t\) to \(t\in {\mathbb {Z}}/q:=\{p/q:~p\in {\mathbb {Z}}\}\) will also still be Markov, hence MISTI, as one can see by considering the related MISTI process \(Y_p:=X_{p/q}\) indexed by \(p\in {\mathbb {Z}}\), so the restriction of \(X_t\) to \({\mathbb {Z}}/q\) must also be either the Poisson or the Negative Binomial MISTI process. Since the union of \(\{{\mathbb {Z}}/q_k\}\) for finitely-many \(\{q_k \in \mathbb {Q}\}\) is in \({\mathbb {Z}}/q_*\) for some \(q_*\in \mathbb {Q}\), it follows that the restriction of \(X_t\) to \(t\in \mathbb {Q}\) is also either the Poisson or the Negative Binomial MISTI process. Finally, one can show that \(X_t\) for each \(t\in {\mathbb {R}}\) is almost-surely equal to the limit from above of \(\{X_s:~s\in \mathbb {Q}, s>t\}\), and conclude that every \({\mathbb {N}}_0\)-valued time-reversible stationary Markov processes \(X_t\) indexed by continuous time \(t\in {\mathbb {R}}\) is also one of the four types described in Theorem . Now we construct the two nontrivial ones explicitly.
4.1 Continuous-Time Poisson Branching Process
Fix \(\theta >0\) and \(\lambda >0\) and construct a nonnegative integer-valued Markov process with generator
or, less precisely but more intuitively, for all \(i,j\in {\mathbb {N}}_0\) and \(\epsilon >0\),
\(X_t\) could be described as a linear death process with immigration. In Sec. () we verify that its univariate marginal distribution and autocorrelation are
and its restriction to integer times \(t\in {\mathbb {Z}}\) is precisely the process described in Sec. () item 3, with one-step autocorrelation \(\rho =e^{-\lambda }\).
4.2 Continuous-Time Negative Binomial Branching Process
Now fix \(\theta >0\), \(\lambda >0\), and \(0<p<1\) and construct a nonnegative integer-valued Markov process with generator
or, for all \(i,j\in {\mathbb {N}}_0\) and \(\epsilon >0\),
so \(X_t\) is a linear birth-death process with immigration. The univariate marginal distribution and autocorrelation (see Sec. ()) are now
and its restriction to integer times \(t\in {\mathbb {Z}}\) is precisely the process described in Sec. () item 4, with autocorrelation \(\rho =e^{-\lambda }\).
4.3 Markov Branching (Linear Birth/Death) Processes
The process \(X_t\) of Sec. () can also be described as the size of a population at time t if individuals arrive in a Poisson stream with rate \(\lambda \theta \) and die or depart independently after exponential holding times with rate \(\lambda \); as such, it is a continuous-time Markov branching process.
Similarly, that of Sec. () can be described as the size of a population at time t if individuals arrive in a Poisson stream with rate \(\lambda \alpha (1{-}p)/p\), give birth (introducing one new individual) independently at rate \(\lambda (1{-}p)/p\), and die or depart at rate \(\lambda /p\). In the limit as \(p\rightarrow 1\) and \(\alpha \rightarrow \infty \) with \(\alpha (1{-}p)\rightarrow \theta \) this will converge in distribution to the Poisson example of Sec. ().
4.4 Marginal Distributions
Here we verify that the Poisson and Negative Binomial distributions are the univariate marginal stationary distributions for the Markov chains with generators \({\mathfrak {A}}\) given in Eqn. (23) and Eqn. (24), respectively.
Denote by \(\pi ^0_i={\textsf {P}}[X_t=i]\) the pmf for \(X_t\) and by \(\pi ^\epsilon _i={\textsf {P}}[X_{t+\epsilon }=i]\) that for \(X_{t+\epsilon }\), and by \(\varphi _0(s)={\textsf {E}}[s^{X_t}]\) and \(\varphi _\epsilon (s)={\textsf {E}}[s^{X_{t+\epsilon }}]\) their generating functions. The stationarity requirement that \(\varphi _0(s)\equiv \varphi _\epsilon (s)\) will determine \(\varphi (s)\) and hence \(\{\pi _i\}\) uniquely.
4.4.1 Poisson
From Eqn. () for \(\epsilon >0\) we have
Multiplying by \(s^i\) and summing, we get:
so
and stationarity (\(\varphi _0(s)\equiv \varphi _\epsilon (s)\)) entails \(\lambda =0\) or \(\varphi _0'(s)/\varphi _0(s) \equiv \theta \), so \(\log \varphi _0(s) \equiv (s-1)\theta \) and:
so \(X_t\sim \textsf {Po}(\theta )\) is the unique stationary distribution.
4.4.2 Negative Binomial
From Eqn. () for \(\epsilon >0\) we have
so either \(\lambda =0\) (the trivial case where \(X_t\equiv X\)) or \(\lambda >0\) and:
and \(X_t\sim \textsf {NB}(\alpha ,p)\) is the unique stationary distribution.
4.4.3 Alternate Proof
A detailed-balance argument (Hoel et al., 1972, p. 105) shows that the stationary distribution \(\pi _i:={\textsf {P}}[X_t=i]\) for linear birth/death chains is proportional to
where \(\beta _j\) and \(\delta _j\) are the birth and death rates when \(X_t=j\), respectively. For the Poisson case, from Eqn. () this is
so \(X_t\sim \textsf {Po}(\theta )\), while for the Negative Binomial case from Eqn. () we have
so \(X_t\sim \textsf {NB}(\alpha ,p)\). In each case the proportionality constant is \(\pi _0=P[X_t=0]\): \(\pi _0=e^{-\theta }\) for the Poisson case, and \(\pi _0=p^\alpha \) for the negative binomial.
4.4.4 Autocorrelation
Aside from the two trivial (iid and constant) cases, MISTI processes have finite pth moments for all \(p<\infty \) and, in particular, have finite variance and well-defined autocorrelation. It follows by the Markov property and induction that the autocorrelation must be of the form
for some \(\rho \in [-1,1]\). In both the Poisson and negative binomial cases the one-step autocorrelation \(\rho \) is nonnegative; without loss of generality we may take \(0<\rho <1\).
5 Discussion
The condition \(\mu ^\theta (\{1\})>0\) introduced in Sec. () to avoid trivial technicalities is equivalent to a requirement that the support \(\mathop {\textrm{spt}}(\mu ^\theta ) ={\mathbb {N}}_0\) be all of the nonnegative integers. Without this condition, for any MISTI process \(X_t\) and any integer \(k\in {\mathbb {N}}\) the process \(Y_t=k\,X_t\) would also be MISTI, leading to a wide range of essentially equivalent processes.
The branching approach of Sec. () could be used to generate a wider class of continuous-time stationary Markov processes with ID marginal distributions (Vervaat, 1979; Steutel et al., 1983). If families of size \(k\ge 1\) immigrate independently in Poisson streams at rate \(\lambda _k\), with \(\sum _{k\ge 1} \lambda _k\log k<\infty \), and if individuals (after independent exponential waiting times) either die (at rate \(\delta >0\)) or give birth to some number \(j\ge 1\) of progeny (at rate \(\beta _j\ge 0\)), respectively, with \(\delta > \sum _{j\ge 1} j\,\beta _j\), then the population size \(X_t\) at time t will be a Markov, infinitely-divisible, stationary processes with nonnegative integer values. Unlike the MISTI processes, these may have infinite pth moments if \(\sum _{k\ge 1} \lambda _k k^p=\infty \) for some \(p>0\) and, in particular, may not have finite means, variances, or autocorrelations.
Unless \(\lambda _k =0\) and \(\beta _j=0\) for all \(k,j>1\), however, these will not be time-reversible, and hence not MISTI. Decreases in population size are always of unit size (necessary for the Markov property to hold), while increases might be of size \(k>1\) (if immigrating family sizes exceed one) or \(j>1\) (if multiple births occur).
References
Milton Abramowitz and Irene A. Stegun, editors. Handbook of Mathematical Functions With Formulas, Graphs, and Mathematical Tables, volume 55 of Applied Mathematics Series. National Bureau of Standards, Washington, D.C., 1964. ISBN 0-486-61272-4.
Mohammed A. Al-Osh and Abdulhamid A. Alzaid. First-order integer-valued autoregressive (INAR\((1)\)) process. J. Time Ser. Anal., 8:261–275, 1987. https://doi.org/10.1111/j.1467-9892.1987.tb00438.x.
Carol B. Edwards and John Gurland. A class of distributions applicable to accidents. J. Am. Stat. Assoc., 56(295):503–517, 1961. https://doi.org/10.2307/2282076.
Theodore E. Harris. The Theory of Branching Processes, volume 119 of Die Grundlehren der Mathematischen Wissenschaften. Springer-Verlag, Berlin, DE, 1963. ISBN 978-3-642-51868-3.
Paul G. Hoel, Sidney C. Port, and Charles J. Stone. Introduction to Stochastic Processes. Houghton Mifflin, Boston, MA, 1972. https://doi.org/10.1109/TSMC.1973.4309295.
Harry Joe. Time series models with univariate margins in the convolution-closed infinitely divisible class. J. Appl. Probab., 33(3):664–677, 1996. https://doi.org/10.2307/3215348.
Norman Lloyd Johnson, Samuel Kotz, and Narayanaswamy Balakrishnan. Discrete Multivariate Distributions. John Wiley & Sons, New York, NY, second edition, 1997. ISBN 0-471-12844-9.
Norman Lloyd Johnson, Adrienne W. Kemp, and Samuel Kotz. Univariate Discrete Distributions. John Wiley & Sons, New York, NY, third edition, 2005. ISBN 0-471-27246-9.
Peter A. W. Lewis. Generating negatively correlated gamma variates using the beta-gamma transformation. In Stephen D. Roberts, Jerry Banks, and Bruce W. Schmeiser, editors, Proceedings of the 1983 Winter Simulation Conference, pages 175–176, 1983.
Peter A. W. Lewis, Edward McKenzie, and D. K. Hugus. Gamma processes. Commun. Stat. Stoch. Models, 5:1–30, 1989. https://doi.org/10.1080/15326348908807096.
Edward McKenzie. Some simple models for discrete variate time series. Water Resources Bulletin, 21(4):645–650, 1985. https://doi.org/10.1111/j.1752-1688.1985.tb05379.x.
Edward McKenzie. Some ARMA models for dependent sequences of Poisson counts. Annals of Applied Probability, 20(4):822–835, 1988. https://doi.org/10.2307/1427362.
L. Chris G. Rogers and David Williams. Diffusions, Markov Processes, and Martingales, volume 1. Cambridge Univ. Press, Cambridge, UK, second edition, 2000. ISBN 0-521-77594-9. https://doi.org/10.1017/CBO9781107590120.
Frederik W. Steutel, Wim Vervaat, and Steven J. Wolfe. Integer-valued branching processes with immigration. Annals of Applied Probability, 15(4):713–725, 1983. https://doi.org/10.2307/1427320.
Wim Vervaat. On a stochastic difference equation and a representation of non-negative infinitely divisible random variables. Annals of Applied Probability, 11(4):750–783, 1979. https://doi.org/10.2307/1426858.
Robert L. Wolpert and Murad S. Taqqu. Fractional Ornstein-Uhlenbeck Lévy processes and the Telecom process: Upstairs and downstairs. Signal Processing, 85(8):1523–1545, 2005. https://doi.org/10.1016/j.sigpro.2004.09.016.
Acknowledgements
The author would like to thank Xuefeng Li, Avi Mandelbaum, Yosef Rinott, Larry Shepp, and Henry Wynn for helpful conversations, and particularly Larry Brown, who originated many of the ideas and conjectures that became theorems. Larry is sorely missed, both for his deep intellect and his delightful charm, generosity, and humanity. This work was supported in part by National Science Foundation grants DMS–1228317 and DMS-2015382 and National Air and Space Administration Applied Information Science Research Program grant NNX09AK60G.
Funding
No funding was received to assist with the preparation of this manuscript. The author has no relevent financial or non-financial interests to disclose.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wolpert, R.L. Markov Infinitely-Divisible Stationary Time-Reversible Integer-Valued Processes. Sankhya A (2024). https://doi.org/10.1007/s13171-024-00368-4
Received:
Published:
DOI: https://doi.org/10.1007/s13171-024-00368-4