Abstract
This paper concerns the estimation of stationary probability of ergodic semi-Markov chains based on an observation over a time interval. We derive asymptotic properties of the proposed estimator, when the time of observation goes to infinity, as consistency, asymptotic normality, law of iterated logarithm and rate of convergence in a functional setting. The proofs are based on asymptotic results on discrete-time semi-Markov random evolutions.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Semi-Markov processes have received extensive attention recently in that it is revealed as a very efficient tool for many applied science and engineering problems. As an extension of Markov processes and renewal processes, the sojourn time of the semi-Markov process can be any distribution, which provides great flexibility in modelling the real problems. Many applications can be found on queuing theory problems (Afanasyeva et al. 2012), reliability and survival analysis (Barbu and Limnios 2008), biological sequences (Barbu and Limnios 2008; Garcia-Maya and Limnios 2019), finance and insurance (Petersson 2016) and so on. The present paper concentrates on the discrete-time semi-Markov process, i.e., the semi-Markov chain.
Since Anselone (1960) first introduced the semi-Markov chain and Howard (1971) further refined its description, more and more researchers have been attracted and dedicated to this topic. Stenflo (1996) presented the periodicity concept of the semi-Markov chain. Mode and Pickens (1998) investigated the computational methods for the renewal theory of the semi-Markov chain and applied it to biology. Barbu et al. (2004) presented an efficient algorithm for solving the Markov renewal equation which is the basic tool in semi-Markov chain theory and applications. Girardin and Limnios (2006, 2018) presented entropy rate calculation of semi-Markov chains. Barbu and Limnios (2006), Barbu and Limnios (2008) presented estimation results with asymptotic properties of semi-Markov chains, and the theory and algorithms of hidden semi-Markov chains. Properties of the maximum likelihood estimator of hidden semi-Markov chains were analyzed by Trevezas and Limnios (1979), and Malefaki et al. (2010). Limnios (2011) and Limnios and Swishchuk (2013), Limnios and Swishchuk (2020) presented the asymptotic theory of random evolution in semi-Markov chain media.
One of the most important characteristics in the ergodic processes is their stationary probability measure. It is a significant problem to estimate the stationary distribution of a semi-Markov chain, because in applied sciences and engineering, we are usually interested in the equilibrium behavior of a system or device, which can be reached after a more or less long run, depending on some spectral properties of the transition operators.
Hence, to estimate the stationary distribution of a semi-Markov chain and find an estimator with asymptotic properties is meaningful and essential. This is the problem we face in this paper, i.e., how to estimate the stationary probability given an observation that consists of a trajectory of a semi-Markov chain in a time interval. We propose an empirical estimator and obtain its asymptotic properties in a functional setting.
Nevertheless, the stationary distribution of a semi-Markov process, in discrete or continuous-time, cannot be defined as in the Markov case, that is, \(\pi A=0\), where \(\pi \) is the stationary probability of the Markov process and A its generator, since we do not have a semigroup. The stationary probability of a semi-Markov process is defined either as a limit of the transition function of the process, or as a marginal distribution of a coupled process, e.g., the semi-Markov process together with the backward recurrence time process.
The stationary distribution for continuous-time semi-Markov processes is studied in the case of finite-state space case in Limnios et al. (2005) and general-state space in Limnios (2006) respectively. However, the present paper is the first to consider the estimation of the stationary distribution for semi-Markov chains in the case of general state space. The proof method used here, for consistency and asymptotic normality, is based on the random evolution theory. The law of iterated logarithm is based on results presented in Herkenrath et al. (2003).
Next Sect. 2, presents the semi-Markov chain setting. Section 3 defines the proposed estimator and the asymptotic results. Section 4 presents the proofs, and finally Sect. 5 presents some concluding remarks.
2 Stationary probability of semi-Markov chains
Semi-Markov chain. Let us denote by \(\mathbb {N}:= \{0,1,2,...\}\) the set of natural numbers, and a complete probability space with countably generated \(\sigma \)-algebra, say \((\Omega , \mathcal{F}, \mathbf {P})\), on which we define all stochastic processes and random variables consider here. It is worth noticing here that \(k\in \mathbb {N}\) denotes the calendar time, while \(n\in \mathbb {N}\) denotes the number of jumps. Let us consider a semi-Markov chain \((Z_k, k\in \mathbb {N})\), with state space a measurable space \((E,\mathcal{E})\). Let us also consider the embedded Markov renewal chain \((J_n, S_n, n\in \mathbb {N})\), where \(J_n\) is the embedded Markov chain, with state space \((E,\mathcal{E})\), and \(0=S_0<S_1<...<S_n<...\) the successive integer valued jump times, and the semi-Markov kernel q(x, B, k), \(x\in E\), \(B\in \mathcal{E}\), \(k\in \mathbb {N}\), that is,
where \(P(x,B):= \mathbf {P}(J_{n+1}\in B\mid J_n=x)\) is the transition kernel of the EMC J, and \(f(x,y,k):= \mathbf {P}(S_{n+1}-S_n=k\mid J_n=x,J_{n+1}=y)\) the conditional distribution of the sojourn time in state x when the next visited state is y.
The process \((N_k, k\in \mathbb {N})\), counts the jumps of the process up to time k, that is, \(N_k = \sup \{n : S_n\le k\}\). The processes \((Z_k, k\in \mathbb {N})\) and \((J_n, n\in \mathbb {N})\) are connected by the following relations:
Let us define also the survival function of the time the SMC Z spends in a state \(x\in E\), \(\bar{F}_x(k) := 1 - \sum _{\ell \le k} q(x,E,\ell )\), and the mean sojourn time in state \(x\in E\), that is, \(m(x):= \sum _{k\ge 0}\bar{F}_x(k)\). Denote also by \(f_x(\ell ):= q(x,E,\ell )\) the law of unconditional sojourn time in state \(x\in E\).
Backward recurrence time process. Consider now the backward recurrence time \(U_k := k - S_{N_k}\), \(k\ge 0\), and the coupled process \((Z_k,U_k)\), which is a Markov chain, with transition kernel operator \(P^\sharp \), see (Limnios and Swishchuk 2020), defined by acting on measurable and bounded test functions \(\varphi (x,k)\), can be written as follows,
where P is the transition kernel operator of the embedded Markov chain J, and \(\lambda (x,k):= \mathbf {P}_x(S_1 =k \mid S_1\ge k)\), the discrete time transition rate of the sojourn time in state \(x\in E\).
It is worth noticing that operator \(P^\sharp \), defined in (1), acting on test functions \(\varphi (x)\), can be written as
where I is the identity operator.
Markov Renewal Equation. The Markov renewal equation plays a crucial role, and it is defined as follows
where u is an unknown function, and v a given function, both defined on \(E\times \mathbb {N}\).
For a semi-Markov kernel q, we define the successive convolution powers, by induction, as follows
and
where \(\mathbf{1}_B\) is the indicator function of the set B.
Markov Renewal Theorem. First define the Markov renewal functions as follows
Let us give now the following important result, which is Shurenkov’s Markov renewal theorem (Shurenkov 1984) restated for discrete-time semi-Markov processes.
The following assumptions are needed in the sequel.
-
A1: The Markov chain \((J_n)\) is ergodic with stationary distribution \(\rho (B), B\in \mathcal{E}\).
-
A2: \(0< m < \infty \), with \(m := \int _E \rho (dx)m(x)\).
Theorem 1
(Markov Renewal Theorem). Under assumptions A1-A2, and the additional assumption that the function v(x, k) satisfies \(\int _E \rho (dx) \sum _{k\ge 0}\left| v(x,k) \right| < \infty \), the MRE (2) has a unique solution, given by \(u(x,k)=\psi * v(x,k)\), and
Limiting and stationary probability. Let us apply this theorem for the transition function of the semi-Markov chain calculation. The transition function of the semi-Markov process is defined as follows
and satisfies the following Markov renewal equation
Then, by the above Markov renewal theorem, we obtain the solution,
and the limit
The limit distribution \(\pi (B)\) is said to be the stationary distribution of the semi-Markov chain \(Z_k\).
Another way to define it, is to consider \(\pi ^\sharp \) the stationary distribution of the Markov process \((Z_k,U_k)\), where \(U_k= k-S_{N_k}\), the backward recurrence time of the SMC, that is
The stationary distribution \(\pi \) of Z, is now defined as the marginal distribution of \(\pi ^\sharp \), that is,
3 Estimation of stationary probability
Let us observe the semi-Markov chain Z on the time interval \([0,k]\subset \mathbb {N}\), that is the censured trajectory \((Z_\ell (\omega ), 0\le \ell \le k)\). Based on these data, we propose the following empirical estimator for the stationary probability of the SMC, \(\pi \), as follows
for any \(B\in \mathcal{E}\).
Convergence results considered here are in the weak functional sense in the space \(D_E[0,\infty )\) with Skorohod topology, denoted by \(\Rightarrow \). The following results are functional asymptotic properties of the estimator (3), namely, consistency, asymptotic normality, and convergence rate results.
In fact \(\hat{\pi }_{[t/\varepsilon ]}(B)\) and \(\hat{\pi }_{[t/\varepsilon ^2]}(B)\), \(t\ge 0\), \(\varepsilon >0\), where the [x] denotes the integer part of the real number x, are families of stochastic processes in continuous time \(t>0\).
Let us denote by \(L_0^2(\pi )\) the subspace of the centered second order elements of \(L^2(\pi )\), that is, \(v\in L_0^2(\pi )\) implies that \(\int _E v^2(x)\pi (dx) <\infty \) and \(\int _E v(x)\pi (dx) =0\).
Proposition 1
(Consistency). Under assumptions A1-A2, we have, for any fixed \(B\in \mathcal{E}\),
for \(t > 0\). And, we can also state this result in the following convergence in probability,
for any but fixed \(T\in \mathbb {R}_+\).
Let us now state the following assumption, (see, e.g., Limnios and Swishchuk 2013; Koroliuk and Limnios 2005).
-
C1: We assume moreover the following uniform integrability condition
$$\begin{aligned} \lim _{M\rightarrow \infty }\sup _{x\in E} \sum _{k\ge M}k^2 {f}_x(k) = 0 \end{aligned}$$which, of course, implies
$$\begin{aligned} \sup _{x\in E} \sum _{k\ge 1}k^2 {f}_x(k)< \infty . \end{aligned}$$
This assumption is needed for tightness of the semi-Markov chains.
For notational simplicity, let us define the process
for any but fixed \(B\in \mathcal{E}\) and \(\xi _0^\varepsilon =0\) for any \(\varepsilon >0\).
Proposition 2
(Asymptotic normality). Under assumptions A1-A2 and C1, we have
where \(B_t\), \(t\ge 0\), is the standard Brownian motion, and
and \(0< b< \infty \), where \(R_0\) is the potential operator of the discrete generator \(P^\sharp -I\), and I is the identity operator.
Remark 3.1
In the weak convergence scheme, the initial value, \(\xi ^\varepsilon _0\), must satisfy the following two conditions, (see, e.g., Koroliuk and Limnios 2005, p. 93),
In our case, both conditions are trivially satisfied. We could also consider \(\xi _0^\varepsilon = \varepsilon r\), for a constant r.
It is worth noticing that from the above Proposition 2, for an observation time of order \(\varepsilon ^{-2}\), we have approximately
Let us now define K, a subset of C[0, 1], the space of continuous functions on [0, 1], with sup-norm, \(\left\| \cdot \right\| \), as follows
where \(\dot{x}(t)\) denotes the derivative of the function x with respect to t.
Let us suppose that the time \(k=0\) is not necessary a jump time, that is, we define \(S_0 \le 0< S_1<S_2< \cdots<S_n < \cdots \), with \(U_0= - S_0 \ge 0\). In that case, we have a slight modification of the definition of \(U_k\), for \(k< S_1\), as follows
Moreover we suppose that the process \((Z_k,U_k)\) is a stationary Markov chain, that is, its initial distribution is equal to its stationary distribution \(\pi ^\sharp (dx\times \{\ell \})= \rho (dx)\overline{F}_x(\ell )/m\), defined previously.
The following assumption is needed in the sequel.
-
C2: Denote \(\left\| \cdot \right\| _2\) the usual norm of the space \(L^2(\pi ^\sharp )\). We assume that
$$\begin{aligned} \sum _{n\ge 1}\left\| (P^\sharp )^{n}v \right\| _2 < \infty , \quad v\in L^2_0(\pi ^\sharp ) \end{aligned}$$where the operator \(P^\sharp \) is defined in (1).
For example, for an one state semi-Markov chain (i.e., a renewal process in discrete-time, Barbu and Limnios 2008), the above condition is trivially satisfied since the first member of the relation is bounded by mv where m is here the mean value of the distribution f.
Proposition 3
(Law of the Iterated Logarithms). Under assumptions A1-A2 and C1-C2, and the additional condition that the process (Z, U) is stationary, the limit points of the following family of processes, as \(\varepsilon \downarrow 0\), \((0<\varepsilon <1)\),
are exactly the set K, \(\mathbf {P}\)-a.s.
As a divergence result, it means that the convergence of the estimator \(\hat{\pi }_{[t/\varepsilon ]}(B)\), in the worst case, is of order
Proposition 4
(Rates of convergence). The rates of convergence, for consistency and asymptotic normality, are given by
for \(0\le t \le T\), where \(C_r:= C_r(T, \left\| R_0 \right\| )\), with \(r=1,2\), are two positive constants.
The above result can be also stated, for both cases, as
And, from the above, we can easily obtain that the bias of estimator (3), goes to zero with a rate O(1/k), that is, as \(k\rightarrow \infty \),
Remark 3.2
It is worth noticing here that all the above results can be also obtained by considering the stochastic functional \(\beta _\varepsilon (t) := \alpha _\varepsilon (t) + \varepsilon b^{-1}(t/\varepsilon ^2 - [t/\varepsilon ^2])a(Z_{[t/\varepsilon ^2]+1})\) in the space \(C[0,\infty )\) of the continuous functions equipped with the uniform topology, instead of \(\alpha _\varepsilon (t)\). In fact, since \(a(\cdot )\) is a bounded function, we have, as \(\varepsilon \rightarrow 0\), \(\varepsilon (t/\varepsilon ^2 - [t/\varepsilon ^2])a(Z_{[t/\varepsilon ^2]+1}) \rightarrow 0\), a.s.
4 Proofs
In this section we will prove the results presented in the previous section 3. In order to prove the results for Proposition 1, 2 and 4, we are using results in paper Limnios and Swishchuk (2013), in particular Result 4.2 and Corollary 5.1. For the asymptotic normality in the stationary case, see also (Herkenrath et al. 2003), Theorem 1. Here in fact we don’t need the process (Z, U) to be stationary, while in Proposition 3, we consider that this process is stationary and we use results from (Herkenrath et al. 2003). For Theorem 1, in Herkenrath et al. (2003), see also (Heyde and Scott 1973; Scott 1973). See also Anisimov (2008), Gikhman and Skorohod (1974), Girardin and Limnios (2018), Shurenkov (1984), Stroock and Varadhan (1979).
Let us now define the projection operator \(\Pi \) of transition operator \(P^\sharp \) by
Define also the potential operator \(R_0\) of \(P^\sharp -I\) by
where \(P^\sharp \) is the transition kernel operator of the MC \((Z_k, U_k)\), as defined by (1).
Proof of Proposition 1
Consider an additive functional of the semi-Markov chain, defined as follows,
where the function \(a: E \longrightarrow \mathbb {R}\), is measurable and bounded.
Consider the additive functional in the series scheme, \(Y_t^\varepsilon \),
and \(Y_0^\varepsilon \) satisfying conditions (4), with limit say u, then we have the following averaging result, see (Limnios and Swishchuk 2013),
where \(\hat{a}= \int _E \pi (dx)a(x)\).
Now from the above result, for \(a:= \mathbf{1}_B\), we get directly the proof of the first part of Proposition 1.
The second part of this proposition is obtained directly since the above limit is a constant (see, e.g., Billingsley 1968). \(\square \)
Proof of Proposition 2
Consider now the continuous time processes \(\xi _t^\varepsilon \), defined as follows
and \(\xi _0^\varepsilon \) satisfy conditions (4), with limit initial point u.
Then for \(a\in L_0(\pi )\), which is equivalent to the balance condition, \(\Pi a=0\), in Limnios and Swishchuk (2013), Result 4.2, we have,
where \(b^2 = 2\overline{a}_0 -\hat{a}_2 \), with
and \(B_t, t\ge 0\), is a standard Brownian motion.
Now putting \(a(x):= \mathbf{1}_B(x) - \hat{a}\), \(x\in E\), which obviously satisfy the balance condition \(\Pi a =0\), we get the desired result. \(\square \)
Proof of Proposition 3
Law of the Iterated Logarithms.
Define
with \(a_k:= a(Z_k)\) and \(S_k:= a_1+...+a_k\).
Let us now consider the processes
By comparison of Theorem 1, in Herkenrath et al. (2003), and the above Proposition 2, we have that \(\sigma _a=b\). Under assumptions A1-A2, C1-C2 and \(\sigma ^2_a>0\), and from Herkenrath et al. (2003), Theorem 1, (see also Heyde and Scott 1973; Scott 1973), we have that the sequence
as a subset of C[0, 1] is \(\mathbf {P}_{\pi ^\sharp }-\)a.s. relative compact (in the uniform topology) and the set of its limit points is exactly K.
Now by taking \(a= \mathbf{1}_B - \hat{a}\), we conclude the proof of Proposition 3. \(\square \)
Proof of Proposition 4
This is an easy consequence of Corollary 5.1 in Limnios and Swishchuk (2013). \(\square \)
5 Concluding remarks
The finite or countable state spaces can be deduced directly as particular cases of the present one. For the stationary case, the estimator (3) can be considered in any interval, that is, for any fixed integer \(c\in \mathbb {N}\),
its asymptotic properties remain still true. Here we have as initial distribution for the process Z at time c its limit distribution which satisfy conditions (4).
The semi-Markov chains are important for applications, and also for numerical calculus schemes of continuous-time processes. We can also use the estimator and its properties, proposed above, to estimate the stationary probability of a continuous-time semi-Markov process. If Q(t) is the semi-Markov kernel of a continuous-time semi-Markov process (see, e.g., Limnios and Oprişan 2001), we can consider it in the discrete-time scheme, with time step \(h>0\), as a SMC with semi-Markov kernel, q(k), given by the relations \(q(k)= Q(kh)-Q((k-1)h)\), \(k\ge 1\), and \(q(0)=0\).
Finally, it is worth noticing that all results presented in this paper are equally valuable when Z is an ergodic Markov chain with transition probability kernel P, since this process is a particular case of the semi-Markov chain with semi-Markov kernel \(Q(x,B,k):= P(x,B)[P(x,\{x\})]^{k-1}\), \(k\ge 1\).
References
Afanasyeva L, Bashtova E, Bulinskaya E (2012) Limit theorems for semi-Markov queues and their applications. Commun Stat Simul Comput 41(6):688–709
Anisimov VV (2008) Switching processes in queueing models, iSTE. Wiley, London
Anselone PM (1960) Ergodic theory for discrete semi-Markov chains. Duke Math J 27(1):33–40
Barbu V, Boussemart M, Limnios N (2004) Discrete time semi-Markov model for reliability and survival analysis. Commun Stat-Theory Methods 33(11):2833–2868
Barbu V, Limnios N (2006) Empirical estimation for discrete time semi-Markov processes with applications in reliability. J Nonparametric Stat 18(7–8):483–498
Barbu V, Limnios N (2008) Semi-Markov Chains and Hidden Semi-Markov Models. Toward Applications. Their use in Reliability and DNA Analysis, Lecture Notes in Statistics, vol. 191, Springer, New York
Billingsley P (1968) Convergence of probability measures. Wiley, New York
Garcia-Maya BI, Limnios N (2019) Identification of words in biological sequences under the semi-Markov hypothesis. J Comput Biol 27(5):683–697
Gikhman II, Skorohod AV (1974) Theory of stochastic processes, vol 2. Springer, Berlin
Girardin V, Limnios N (2006) Entropy for semi-Markov processes with Borel state spaces: asymptotic equirepartition properties and invariance principles. Bernoulli 12(2):1–19
Girardin V, Limnios N (2018) Applied probability—from random sequences to stochastic processes. Springer, Berlin
Heyde CC, Scott DJ (1973) Invariance principles for the law of iterated algorithm for martingales and processes with stationary increments. Ann Probab 1:428–436
Herkenrath U, Iosifescu M, Rudolph A (2003) A note on invariance principles for iterated random functions. J Appl Prob 40:834–837
Howard R (1971) Dynamic probabilistic systems, volume II: semi-Markov and decision processes. Wiley, New York
Koroliuk VS, Limnios N (2005) Stochastic systems in merging phase space. World Scientific, Singapore
Limnios N (2006) Estimation of the stationary distribution of semi-Markov processes with Borel state space. Stat Probab Lett 76(14):1536–1542
Limnios N (2011) Discrete-time semi-Markov random evolutions - average and diffusion approximation of difference equations and additive functionals. Commun Stat Theory Methods 40(19–20):3396–3406
Limnios N, Ouhbi B, Sadek A (2005) Empirical estimator of stationary distribution for semi-Markov processes. Commun Stat Theory Methods 34(3):987–995
Limnios N, Oprişan G (2001) Semi-Markov processes and reliability. Birkhäuser, Boston
Limnios N, Swishchuk A (2013) Discrete-time semi-Markov random evolutions and their applications. Adv Appl Probab 45(1):214–240
Limnios N, Swishchuk A (2020) Discrete-time semi-Markov random evolutions in asymptotic reduced random media with applications. Mathematics 8(6):963
Malefaki S, Trevezas S, Limnios N (2010) An EM and a stochastic version of the EM algorithm for nonparametric Hidden semi-Markov models. Commun Stat Simul Comput 39(2):240–261
Mode CJ, Pickens GT (1998) Computational methods for renewal theory and semi-Markov processes with illustrative examples. Am Stat 42(2):143–152
Petersson M (2016) Perturbed discrete time stochastic models, Doctoral Thesis, Mathematical Statistics, Stockholm University
Scott DJ (1973) Central limit theorems for martingales and processes with stationary increments using a Skorokhod representation approach. Adv Appl Prob 5:119–137
Shurenkov VM (1984) On the theory of Markov renewal. Theory Prob Appl 19(2):247–265
Skorohod AV (1989) Asymptotic Methods in the Theory of Stochastic Differential Equations, AMS, vol. 78, Providence
Stenflo O (1996) Iterated function systems controlled by a semi-Markov Chain. Theory Stoch Process 18:305–313
Stroock DW, Varadhan SRS (1979) Multidimensional diffusion processes. Springer, Berlin
Trevezas S, Limnios N (1979) Maximum likelihood estimation for general hidden semi-Markov processes with backward recurrence time dependence. J Math Sci 163(3):262
Acknowledgements
The authors are grateful to the anonymous referees for their detailed remarks and suggestions which led to significant improvements in the presentation of this paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Limnios, N., Wu, B. Estimation of stationary probability of semi-Markov Chains. Stat Inference Stoch Process 25, 355–364 (2022). https://doi.org/10.1007/s11203-021-09255-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11203-021-09255-3
Keywords
- Semi-Markov chain
- Stationary probability
- Estimation
- Consistency
- Asymptotic normality
- Law of iterated logarithm
- Rate of convergence