Abstract
Chapter 4 deals with the probability distribution of αε( ⋅) through the corresponding forward equation and is mainly an analytical approach, whereas the current chapter is largely probabilistic in nature. The central theme of this chapter is limit results of unscaled as well as scaled sequences of occupation measures, which include the law of large numbers for an unscaled sequence, exponential upper bounds, and asymptotic distribution of a suitably scaled sequence of occupation times.
Access provided by Autonomous University of Puebla. Download chapter PDF
Keywords
- Occupation Measure
- Principal Solution Matrix
- Martingale Problem
- Perturbed Liapunov Function
- Quasi-stationary Distribution
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
1 Introduction
Chapter 4 deals with the probability distribution of αε( ⋅) through the corresponding forward equation and is mainly an analytical approach, whereas the current chapter is largely probabilistic in nature. The central theme of this chapter is limit results of unscaled as well as scaled sequences of occupation measures, which include the law of large numbers for an unscaled sequence, exponential upper bounds, and asymptotic distribution of a suitably scaled sequence of occupation times. It further exploits the deviation of the functional occupation times from its quasi-stationary distribution. We obtain estimates of centered deviations, prove the convergence of a properly scaled and centered sequence of occupation times, characterize the limit process by deriving the limit distribution and providing explicit formulas for the mean and covariance functions, and provide exponential bounds for the normalized process. It is worthwhile to note that the limit covariance function depends on the initial-layer terms in contrast with most of the existing results of central limit type.
The rest of the chapter is arranged as follows. We first study the asymptotic properties of irreducible Markov chains in Section 5.2. In view of the developments in Remarks 4.34 and 4.39, the Markov chain with recurrent states is the most illustrative and representative one. As a result, in the remaining chapters, we mainly treat problems associated with this model. Starting in Section 5.3.1, we consider Markov chains with weak and strong interactions with generators consisting of multiple irreducible blocks. After treating aggregation of the Markov states, we study the corresponding exponential bounds. We deal with asymptotic distributions. Then in Section 5.4, we treat Markov chains with generators that are merely measurable. Next, remarks on inclusion of transient and absorbing states are provided in Section 5.5. Applications of the weak convergence results and a related stability problem are provided in Section 5.6. Finally, Section 5.7 concludes the chapter with notes and further remarks.
2 The Irreducible Case
The notion of occupation measure is set forth first. We consider a sequence of unscaled occupation measures and establish its convergence in probability to that of the accumulative quasi-stationary distribution. This is followed by exponential bounds of the function occupation time and moment estimates. In addition, asymptotic normality is derived. Although the prelimit process has nonzero mean and is nonstationary, using the results of Section 4.2, the quasi-stationary regime is established after a short period (of order O(ε)). We also calculate explicitly the covariance representation of the limit process, and prove that the process αε( ⋅) satisfies a mixing condition. The tightness of the sequence and the w.p.1 continuity of the sample paths of the limit process are proved by estimating the fourth moment. The limit of the finite-dimensional distributions is then calculated and shown to be Gaussian. By proving a series of lemmas, we derive the desired asymptotic normality.
As was mentioned in previous chapters, the process αε( ⋅) arises from pervasive practical use that involves a rapidly fluctuating finite-state Markov chain. In these applications, the asymptotic behavior of the Markov chain αε( ⋅) has a major influence. Further investigation and understanding of the asymptotic properties of αε( ⋅), in particular, the probabilistic structures, play an important role in the in-depth study.
In Section 4.2, using singular perturbation techniques, we examined the asymptotic properties of \({p}_{i}^\varepsilon (t) = P({\alpha }^\varepsilon (t) = i)\). It has been proved that p ε(t) = (p 1 ε(t), …, p m ε(t)) converges to the quasi-stationary distribution ν(t) as ε → 0 for each t > 0 and p ε(t) admits an asymptotic expansion in terms of ε. To gain further insight, we ask whether there is a limit result for the occupation measure \(\int \nolimits_{0}^{t}{I}_{\{{\alpha }^\varepsilon (s)=i\}}ds\). If a convergence is expected to take place, then what is the rate of convergence? Does one have a central limit theorem associated with the αε( ⋅)-process? The answers to these questions are affirmative. We will prove a number of limit results related to an unscaled sequence, and a suitably scaled and normalized sequence. Owing to the asymptotic expansions, the scaling factor is \(\sqrt \varepsilon \). The limit process is Gaussian with zero mean, and the covariance of the limit process depends on the asymptotic expansion in an essential way, which reflects one of the distinct features of the central limit theorem. It appears that it is virtually impossible to calculate the asymptotic covariance of the Gaussian process without the help of the asymptotic expansion, which reveals a salient characteristic of the two-time-scale Markov chain.
A related problem is to examine the exponential bounds on the scaled occupation measure process. This is similar to the estimation of the moment generating function. Such estimates have been found useful in studying hierarchical controls of manufacturing systems. Using the asymptotic expansion and the martingale representation of finite-state Markov chains, we are able to establish such exponential bounds for the scaled occupation measures.
2.1 Occupation Measure
Let \((\Omega,\mathcal{F},P)\) denote the underlying probability space. As in Section 4.2, αε( ⋅) is a nonstationary Markov chain on \((\Omega,\mathcal{F},P)\) with finite-state space \(\mathcal{M} =\{ 1,\ldots,m\}\) and generator \({Q}^\varepsilon (t) = Q(t)/\varepsilon \).
For each \(i \in \mathcal{M}\), let β i ( ⋅) denote a bounded Borel measurable deterministic function and define a sequence of centered (around the quasi-stationary distribution) occupation measures Z i ε(t) as
Set Z ε(t) = (Z 1 ε(t), …, Z m ε(t)). It is a measure of the functional occupancy for the process αε( ⋅). Our interest lies in the asymptotic properties of the sequence defined in (5.1). To proceed, we first present some conditions and preliminary results needed in the sequel.
Note that a special choice of β i ( ⋅) is β i (t) = 1, for t ∈ [0, T]. To insert β i ( ⋅) in sequence allows one to treat various situations in some applications. For example, in the manufacturing problem, β i (t) is often given by a function of a control process; see Chapter 8 for further details.
2.2 Conditions and Preliminary Results
To proceed, we make the following assumptions.
-
(A5.1)
For each t ∈ [0, T], Q(t) is weakly irreducible.
-
(A5.2)
( ⋅) is continuously differentiable on [0, T], and its derivative is Lipschitz.
Recall that \({p}^\varepsilon (t) = (P({\alpha }^\varepsilon (t) = 1),\ldots,P({\alpha }^\varepsilon (t) = m))\) and let
Use P ε(t, t 0) to denote the transition matrix (p ij ε(t, t 0)). The following lemma is on the asymptotic expansion of P ε(t, t 0).
Lemma 5.1
Assume (A5.1) and (A5.2) . Then there exists a positive constant κ 0 such that for each fixed 0 ≤ T < ∞,
uniformly in (t 0 ,t) where 0 ≤ t 0 ≤ t ≤ T and
In addition, assume Q(⋅) to be twice continuously differentiable on [0,T] with the second derivative being Lipschitz. Then
uniformly in (t 0 ,t), where 0 ≤ t 0 ≤ t ≤ T,
and
where φ 1 (t) is given in (4.13) ( with \(\tau := (t - {t}_{0})/\varepsilon \) ) . Furthermore, for i = 0,1, the P i (⋅) are (2 − i) times continuously differentiable on [0,T] and there exist constants K > 0 and κ 0 > 0 such that
uniformly for t 0 ∈ [0,T].
Remark 5.2
Recall that ν(t) and φ1(t) are row vectors. As a result, P0(⋅) and P1(⋅) have identical rows. This is a consequence of the convergence of pε(t) to the quasi-stationary distribution and the asymptotic expansions.
Proof of Lemma 5.1: It suffices to verify (5.3) because (5.2) can be derived similarly. The asymptotic expansion of P ε(t, t 0) can be obtained as in Section 4.2. Thus only the exponential bound (5.4) needs to be proved. The main task is to verify the uniformity in t 0. To this end, it suffices to treat each row of Q i (τ, t 0) separately. For a fixed i = 0, 1, let
denote any row of Q i (τ, t 0) and η0(t 0) the corresponding row in Q i (0, t 0) with
Then η(τ, t 0) satisfies the differential equation
By virtue of the assumptions of Lemma 5.1 and the asymptotic expansion, it follows that η0(t 0) is uniformly bounded and η0(t 0)1 = 0.
Define
Then Lemma A.6 implies that \(\widehat{k} > 0\). In view of Theorem 4.5, it suffices to show that for all \(\tau \geq 0\) and for some constant K > 0 independent of t 0,
To verify (5.5), note that for any ς0 ∈ [0, T],
Solving this differential equation by treating η(τ, t 0)[Q(t 0) − Q(ς0)] as the driving term, we have
In view of (A5.2), for some K 0 > 0,
Noting that η0(t 0)1 = 0 and that P 0(t) has identical rows, we have
Thus the equation in (5.6) is equivalent to
From Lemma A.2, we have
for some constants K 1 and K 2 which may depend on ς0 but are independent of t 0. By Gronwall’s inequality (see Hale [79, p. 36]),
for all t 0 ∈ [0, T] and τ > 0.
If (5.5) does not hold uniformly, then there exist τ n > 0 and t n ∈ [0, T] such that
Since T is finite, we may assume t n → ς0, as n → ∞. This contradicts (5.7) for n large enough satisfying \(\vert {t}_{n} - {\varsigma }_{0}\vert <\widehat{ \kappa }/(2{K}_{2})\) and K 1 < n. Thus the proof is complete. □
Unscaled Occupation Measure
To study the unscaled occupation measure Z i ε(t) in (5.1), we define a related sequence \(\{\widehat{{Z}}^\varepsilon (t)\}\) of \({\mathbb{R}}^{m}\)-valued processes with its ith component \(\widehat{{Z}}_{i}^\varepsilon (t)\) given by
Assume the conditions (A5.1) and (A5.2). We claim that for any δ > 0,
Note that (5.8) follows from (5.9) using Tchebyshev’s inequality. The verification of (5.9), which mainly depends on a mixing property of the underlying sequence, is almost the same as the moment estimates in the proof of asymptotic normality in Lemma 5.13. The details of the verifications of (5.8) and (5.9) are omitted here.
With (5.9) in hand for any δ > 0, to study the asymptotic properties of Z ε( ⋅), it remains to show that
In fact, it is enough to work with each component of Z ε(t). Note that both Z ε(t) and \(\widehat{Z}^\varepsilon(t)\) are bounded. This together with the boundedness of β(t) and Lemma 5.1 implies that for each \(i \in \mathcal{M}\),
as ε → 0, which yields the desired results.
The limit result above is of the law-of-large-numbers type. What has been obtained is that as ε → 0,
for 0 < t ≤ T. In fact, a somewhat stronger result on uniform convergence in terms of the second moment is established. To illustrate, suppose that \({\alpha }^\varepsilon (t) = \alpha (t/\varepsilon )\) such that α( ⋅) is a stationary process with stationary distribution \(\overline{\nu } = ({\overline{\nu }}_{1},\ldots,{\overline{\nu }}_{m})\). Then via a change of variable \(\varsigma = s/\varepsilon \), we have
for 0 < t ≤ T. This is exactly the continuous-time version of the law of large numbers.
Example 5.3
Let us return to the singularly perturbed Cox process of Section 3.3. Recall that the compensator of the singularly perturbed Cox process is given by
where ai > 0 for i = 1,…,m. Assume that all the conditions in Lemma 5.1 hold. Then Theorem 4.5 implies that P(α ε (t) = i) → ν i (t) as ε → 0. What we have discussed thus far implies that for each \(i \in \mathcal{M}\),
Moreover,
In the rest of this chapter, we treat suitably scaled occupation measures; the corresponding results for the Cox process can also be derived.
With the limit results in hand, the next question is this: How fast does the convergence take place? The rate of convergence issue together with more detailed asymptotic properties is examined fully in the following sections.
2.3 Exponential Bounds
This section is devoted to the derivation of exponential bounds for the normalized occupation measure (or occupation time) n ε( ⋅). Given a deterministic process β( ⋅), we consider the “centered” and “scaled” functional occupation-time process n ε(t, i) defined by
In view of Lemma 5.1, we have, for 0 ≤ s ≤ t ≤ T,
for some κ0 > 0. Note that the big O( ⋅) usually depends on T. Let K T denote an upper bound of
for 0 ≤ s ≤ t ≤ T. For convenience, we use the notation O 1(y) to denote a function of y such that | O 1(y) | ∕ | y | ≤ 1. The rationale is that K T represents the magnitude of the bounding constant and the rest of the bound is in terms of a function with norm bounded by 1. Using this notation and K T , we write
Let y(t) = ( y ij (t)) and z( t) = (z i (t)) denote a matrix-valued function and a vector-valued function defined on [0, T], respectively. Their norms are defined by
For future use, define β(t) = diag(β 1(t), …, β m (t)). The following theorem is concerned with the exponential bound of n ε(t) for ε sufficiently small.
Theorem 5.4
Assume that (A5.1) and (A5.2) are satisfied. Then there exist ε 0 > 0 and K > 0 such that for all 0 < ε ≤ ε 0 , T ≥ 0, and any bounded and measurable deterministic function β(⋅) = diag (β 1 (⋅),…,β m (⋅)), the following exponential bound holds:
where θ T is a constant satisfying
with κ 0 being the exponential constant in Lemma 5.1.
Remark 5.5
Note that the constants ε 0 and K are independent of T. This is a convenient feature of the estimate in certain applications. The result is in terms of a fixed but otherwise arbitrary T, which is particularly useful for systems in an infinite horizon.
Proof: The proof is divided into several steps.
Step 1. In the first step, we show that (5.13) holds when the “sup” is absent. Let χε( ⋅) denote the indicator vector of αε( ⋅), that is,
It is well known (see Elliott [56]) that w ε(t) = (w 1 ε(t), …, w m ε(t)), for t ≥ 0, is a σ{α ε(s) : s ≤ t}-martingale. In view of a result of Kunita and Watanabe [134] (see also Ikeda and Watanabe [91, p. 55]), one can define a stochastic integral with respect to w ε(t). Moreover, the solution of the linear stochastic differential equation
is given by
where P ε(t, s) is the principal matrix solution to the equation
representing the transition probability matrix.
Note that for t ≥ s ≥ 0,
Using this and (5.11), we have
The last equality above follows from the observation that
and
Recall that β( t) = diag(β1(t), …, β m (t)). Then it follows that
where b( s, t) is a measurable function and | b(s, t) | ≤ 1 for all s and t. Dividing both sides by (T + 1), we obtain
Therefore, we have
In view of the choice of θ T , it follows that
Recall that
It suffices to work out the estimate for each component w i ε (t). That is, it is enough to show that for each i = 1, …, m,
for all measurable functions b( ⋅, ⋅) with | b(s, t) | ≤ 1 and 0 ≤ t ≤ T. For each t 0 ≥ 0, let b 0(s) = b(s, t 0).
For any nonnegative random variable ξ,
By virtue of the inequality above, we have
To proceed, let us concentrate on the estimate of
For each i = 1, …, m, let
and let \(\widetilde{{q}}_{i}(\cdot )\) denote the only solution to the following equation (see Elliott [55, Chapter 13])
where \(\widetilde{{q}}_{i}({s}^{-})\) is the left-hand limit of \(\widetilde{{q}}_{i}\) at s and ζ is a positive constant to be determined later. In what follows, we suppress the i-dependence and write \(\widetilde{{p}}_{i}(\cdot )\) and \(\widetilde{{q}}_{i}(\cdot )\) as \(\widetilde{p}(\cdot )\) and \(\widetilde{q}(\cdot )\) whenever there is no confusion.
Note that \(\widetilde{p}(t)\) , for t ≥ 0, is a local martingale. Since
is a local martingale, we have \(E\widetilde{q}(t) \leq 1\mbox{ for all }t \geq 0\). Moreover, \(\widetilde{q}(t)\) can be written as follows (see Elliott [55, Chapter 13]):
where \(\Delta \widetilde{p}(s) :=\widetilde{ p}(s) -\widetilde{ p}({s}^{-})\), with \(\vert \Delta \widetilde{p}(s)\vert \leq 1\).
Now observe that there exist positive constants ζ0 and κ1 such that for 0 < ζ ≤ ζ0 and for all s > 0,
Combining (5.19) and (5.20), we obtain
where N(t) is the number of jumps of \(\widetilde{p}(s)\) in s ∈ [0, t]. Since N(t) is a monotonically increasing process, we have
Note also that for each i = 1, …, m,
Consider the first term on the right-hand side of the inequality above. Let \({a}_{j} = j(T + 1)/(8{\kappa }_{1}\varepsilon )\). Then
The last inequality follows from the local martingale property (see Elliott [55, Theorem 4.2]).
Now if we choose \(\zeta = 4\sqrt\varepsilon /\sqrt{T + 1}\), then
In view of the construction of Markov chains in Section 2.4, there exists a Poisson process N 0 ( ⋅) with parameter (i.e., mean) a ∕ ε for some a > 0, such that N(t) ≤ N 0(t). Assume a = 1 without loss of generality (otherwise one may replace ε by εa − 1 ). Using the Poisson distribution of N 0(t), we have
In view of Stirling’s formula (see Chow and Teicher [30] or Feller [60]), for ε small enough,
where ⌊ a j ⌋ is the integer part of a j and
Thus, for j ≥ j 0,
Repeating the same argument for the martingale \((-\widetilde{p}(\cdot ))\), we get for j ≥ j 0,
Combining the above two inequalities yields that for j ≥ j 0,
Then by (5.18),
where K 0 is the sum corresponding to those terms with j ≤ j 0. Now choose ε small enough that \(e{\gamma }_{0}^{1/(8{\kappa }_{1}\varepsilon )} \leq 1/2\). Then
Since t 0 is arbitrary, we may take t 0 = t in the above inequality. Then
Combining this inequality with (5.16) leads to
Step 2. Recall that
Then, for each \(i \in \mathcal{M}\), n ε(t, i) is nearly a martingale, i.e., for ε small enough,
Here \(O \sqrt{\varepsilon}\) is deterministic, i.e., it does not depend on the sample point ω. The reason is that it is obtained from the asymptotic expansions. In fact, for all \({i}_{0} \in \mathcal{M}\),
So, (5.21) follows.
Step 3. We show that for each a > 0,
First of all, note that ϕ(x) = | x | is a convex function. There exists a vector function ϕ0(x) bounded by 1 such that
for all x and a. Noting that \(O(\sqrt\varepsilon ) = -O(\sqrt\varepsilon )\) , we have
Moreover, note that e ax is also convex. It follows that
Step 4. Let x ε(t) = exp(a | n ε(t, i) | ) for a > 0. Then, for any \({\mathcal{F}}_{t}\) stopping time τ ≤ T,
Note that x ε(t) is continuous. Therefore, it suffices to show the above inequality when τ takes values in a countable set {t 1 , t 2, …}. To this end, note that, for each t i ,
For all \(A \in {\mathcal{F}}_{\tau }\) , we have \(A \cap \{ \tau = {t}_{i}\} \in {\mathcal{F}}_{{t}_{i}}\) . Therefore,
Thus
and (5.22) follows.
Step 5. Let \(a = \theta /\sqrt{{(T + 1)}^{3}}\) in Step 3. Then, for ε small enough, there exists K such that
for all x > 0.
In fact, let τ = inf{t > 0 : x ε(t) ≥ x}. We adopt the convention that τ = ∞ if { t > 0 : x ε(t) ≥ x} = ∅. Then we have
and write
Moreover, in view of the definition of τ, we have
It follows that
Thus, (5.23) follows.
Finally, to complete the proof of (5.13), note that, for 0 < κ < 1,
It follows that
This completes the proof. □
Next we give several corollaries to the theorem. Such estimates are useful for establishing exponential bounds of asymptotic optimal hierarchical controls in manufacturing models (see Sethi and Zhang [192]).
Corollary 5.6
In Theorem 5.4 , if Q(t) = Q, a constant matrix, then the following stronger estimate holds:
Moreover, the constant θ T = θ is independent of T for T > 0.
Proof: If Q(t) = Q, then φ1(t) in Lemma 5.1 is identically 0. Therefore, the estimate (5.11) can be replaced by
As a result, the estimate in (5.15) can be replaced by
The proof of (5.24) follows in essentially the same way as that of Theorem 5.4 (from equation (5.15) on).
To see that θ T in (5.24) is independent of T, it suffices to note that in (5.11) the constant K T is independent of T, which can be seen by examining closely Example 4.16. □
Corollary 5.7
Under the conditions of Theorem 5.4 , there exist constants K j , such that for j = 1,2,…,
Moreover, if Q(t) = Q, then for some K j independent of T and
Proof: Since (5.26) follows from a similar argument to that of Corollary 5.6, it suffices to verify (5.25) using Theorem 5.4. Note that for each j = 1, 2, …, there exists K j 0 such that for all x, we have x 2j ≤ K j 0 e x. Thus,
Taking expectations on both sides of the above inequality yields the desired estimate. □
Corollary 5.8
Under the conditions of Theorem 5.4 , for each 0 < δ < 1∕2, we have
Moreover, if Q(t) = Q, then θ T = θ is independent of T and
Proof: Using Theorem 5.4, we obtain
This proves (5.27). Similarly, (5.28) follows from Corollary 5.6. □
2.4 Asymptotic Normality
Recall that the ith component of n ε( ⋅) is given by
It is expected that the sequence of centered and scaled occupation measures will display certain “central limit type” phenomena. The goal here is to study the asymptotic properties of n ε( ⋅) as ε → 0. To be more specific, we show that n ε( ⋅) converges to a Gaussian process as ε goes to 0. The following theorem is the main result of this section.
Theorem 5.9
Suppose that (A5.1) is satisfied and Q(⋅) is twice continuously differentiable in [0,T] with the second derivative being Lipschitz. Then for t ∈ [0,T], the process n ε (⋅) converges weakly to a Gaussian process n(⋅) with independent increments such that
where A(t) = (A ij (t)) with
and Q 0 (r,t) = (q 0,ij (r,t)).
Remark 5.10
In view of (5.29) and the independent increment property of n(t), it follows that
The form of the covariance matrix (between t1 and t2) reveals the nonstationarity of the limit process n(⋅). Note that the limit covariance of n(t) given in (5.31) is an integral of the function A(s) defined in (5.30). For simplicity, with a slight abuse of notation, we shall also call A(t) as the covariance. This convention will be used throughout the chapter.
Remark 5.11
The additional assumptions on the second derivative of Q(⋅) in Theorem 5.9 are required for computing or characterizing the function A(⋅). It is not crucial for the convergence of n ε (⋅); see Remark 5.44 in Section 5.3.3 for details.
Proof of Theorem 5.9: We divide the proof into several steps, which are presented by a number of lemmas.
Step 1. Show that the limit of the mean of n ε( ⋅) is 0.
Lemma 5.12
For each t ∈ [0,T],
Proof: Using Theorem 4.5 and the boundedness of β i ( ⋅), for t ∈ [0, T],
for each \(i \in \mathcal{M}\). □
Step 2. Calculate the limit covariance function.
Lemma 5.13
For each t ∈ [0,T],
where A(t) is given by (5.30).
Proof: For each \(i,j \in \mathcal{M}\),
Let
and let
Then it follows that
Note that if (ς, r) ∈ D 1, then ς ≥ r and
Hence, for (ς, r) ∈ D 1 we have
Using Theorem 4.5 and Lemma 5.1, for (ς, r) ∈ D 1,
In the above, φ ℓ i and ψ ℓ i denote the ith components of the vectors φℓ and ψ ℓ , respectively. By elementary integration, we have
Thus, it follows that
Exchanging the order of integration leads to
Making a change of variables (via \(\varsigma - r = \varepsilon s\)) yields
We note that β i ( ⋅) is bounded and β i (r + ε s) → β i (r) in L 1 for each r ∈ [0, T], as ε → 0. Since q 0, ji ( ⋅) decays exponentially fast, as in Lemma 5.1, we have
Therefore, we obtain
Similarly, we can show that
Combining (5.33) and (5.34), we obtain
with A(t) = ( A ij (t)) given by (5.30). □
Step 3. Establish a mixing condition for the sequence { n ε( ⋅)}.
Lemma 5.14
For any ς ≥ 0 and σ{α ε (s) : s ≥ t + ς}-measurable η with |η|≤ 1,
Remark 5.15
It follows from (5.35) that for any σ{αε(s) : 0 ≤ s ≤ t}-measurable ξ with |ξ|≤ 1 and η given in Lemma 5.14,
We will make crucial use of (5.35) and (5.36) in what follows.
Proof of Lemma 5.14: For any
let
Then in view of the Markovian property of α ε ( ⋅),
Similarly, we have
We first show that
for some positive constants K and κ that are independent of \(i,j \in \mathcal{M}\) and t ∈ [0, T].
To verify (5.37), it suffices to show that for any \(k \in \mathcal{M}\),
Since P 0 ( ⋅) and P 1( ⋅) have identical rows, the asymptotic expansion in (5.3) implies that \({p}_{ij}^\varepsilon (t + \zeta,t) - {p}_{kj}^\varepsilon (t + \zeta,t)\) is determined by Q 0 (ζ ∕ ε, t). By virtue of the asymptotic expansion (see Theorem 4.5 and Lemma 5.1), there exist a K 1 > 0 and a κ 0 > 0 such that
Choose N > 0 sufficiently large that K 1 exp( − κ 0 N) < 1. Then for ε > 0 sufficiently small, there is a 0 < ρ < 1 such that
To proceed, subdivide \([t + N\varepsilon,t + \varsigma ]\) into intervals of length Nε.
In view of the Chapman–Kolmogorov equation,
for any \({l}_{1} \in \mathcal{M}\) . Iterating on the inequality above, we arrive at
Choose \(\kappa = -1/(2N)\log \rho \) , and note that κ > 0. Then for any ς satisfying k 0 Nε ≤ ς < ( k 0 + 1)Nε,
Thus (5.37) holds. This implies that αε( ⋅) is a mixing process with exponential mixing rate. By virtue of Lemma A.16, (5.35) holds. □
Step 4. Prove that the sequence n ε( ⋅) is tight, and any weakly convergent subsequence of {n ε( ⋅)} has continuous paths with probability 1.
Lemma 5.16
The following assertions hold:
-
(a)
{n ε(t); t ∈ [0, T]} is tight in \(D([0,T]; {\mathbb{R}}^{m})\), where \(D([0,T]; {\mathbb{R}}^{m})\) denotes the space of functions that are defined on [0, T] and that are right continuous with left limits.
-
(b)
The limitn( ⋅) of any weakly convergent subsequence ofn ε( ⋅) has continuous sample paths with probability 1.
Proof: For \(i \in \mathcal{M}\) , define
By virtue of Theorem 4.5,
Thus \({n}^\varepsilon (t,i) =\widetilde{ {n}}^\varepsilon (t,i) + O(\sqrt\varepsilon )\) , and as a result the tightness of { n ε( ⋅)} will follow from the tightness of \(\{\widetilde{{n}}^\varepsilon (\cdot )\}\) (see Kushner [139, Lemma 5, p. 50]).
For the tightness of \(\{\widetilde{{n}}^\varepsilon (\cdot )\}\), in view of Kushner [139, Theorem 5, p. 32], it suffices to show that
To verify this assertion, it is enough to prove that for each \(i \in \mathcal{M}\), \(\widetilde{{n}}^\varepsilon (\cdot,i)\) satisfies the condition.
Fix \(i \in \mathcal{M}\) and for any 0 ≤ t ≤ T, let
We have suppressed the i and ε dependence in θ( t) for ease of presentation. Let \(D =\{ ({s}_{1},{s}_{2},{s}_{3},{s}_{4}) : t \leq {s}_{i} \leq t + \varsigma,i = 1,2,3,4\}\). It follows that
Let (i 1 , i 2, i 3 , i 4) denote a permutation of (1, 2, 3, 4) and
Then it is easy to see that \(D = \cup {D}_{{i}_{1}{i}_{2}{i}_{3}{i}_{4}}\). This and (5.40) leads to
where D 0 = D 1234.
Note that
By virtue of (5.36) and Eθ( t) = 0, t ≥ 0,
Similarly, we have
Therefore, it follows that
The elementary inequality \({(a + b)}^{1/2} \leq {a}^{1/2} + {b}^{1/2}\) for nonnegative numbers a and b yields that
In view of (5.36), we obtain
Similarly, by virtue of (5.35) and the boundedness of θ(s),
and
By virtue of the estimates above, we arrive at
The estimate (5.39) then follows from (5.42) and (5.43), and so does the desired tightness of {n ε ( ⋅)}.
Since {n ε( ⋅)} is tight, by Prohorov’s theorem, we extract a convergent subsequence, and for notational simplicity, we still denote the sequence by {n ε ( ⋅)} whose limit is n( ⋅). By virtue of Kushner [139, Theorem 5, p. 32] or Ethier and Kurtz [59, Proposition 10.3, p. 149], n( ⋅) has continuous paths with probability 1. □
Remark 5.17
Step 4 implies that both nε(⋅) and n(⋅) have continuous sample paths with probability 1. It follows, in view of Prohorov’s theorem (see Billingsley [13]), that nε(⋅) is tight in \(C([0,T]; {\mathbb{R}}^{m})\).
Step 5. Show that the finite-dimensional distributions of n ε( ⋅) converge to that of a Gaussian process with independent increments.
This part of the proof is similar to Khasminskii [112] (see also Friedlin and Wentzel [67, pp. 224]). Use ι to denote the imaginary number \({\iota }^{2} = -1\). To prove the convergence of the finite-dimensional distributions, we use the characteristic function Eexp(ι⟨ z, n ε(t)⟩), where \(z \in {\mathbb{R}}^{m}\) and ⟨ ⋅. ⋅⟩ denotes the usual inner product in \({\mathbb{R}}^{m}\). Owing to the mixing property and repeated applications of Remark 5.15, for arbitrary positive real numbers s l and t l satisfying
we have
as ε → 0, for \({z}_{l} \in {\mathbb{R}}^{m}\). This, in turn, implies that the limit process n( ⋅) has independent increments. Moreover, in view of Lemma 5.16, the limit process has continuous path with probability 1. In accordance with a result in Skorohod [197, p. 7], if a process with independent increments has continuous paths w.p.1, then it must necessarily be a Gaussian process. This implies that the limits of the finite-dimensional distribution of n( ⋅) are Gaussian.
Consequently, n( ⋅) is a process having Gaussian finite-dimensional distributions, with mean zero and covariance ∫0 t A( s)ds given by Lemma 5.13. Moreover, the limit does not depend on the chosen subsequence. Thus n ε( ⋅) converges weakly to the Gaussian process n( ⋅). This completes the proof of the theorem. □
To illustrate, we give an example in which the covariance function of the limit process can be calculated explicitly.
Example 5.18
Let \({\alpha }^\varepsilon (t) \in \mathcal{M} =\{ 1,2\}\) be a two-state Markov chain with a generator
where μ1(t) ≥ 0, μ2(t) ≥ 0, and μ1(t) + μ2(t) > 0 for each t ∈ [0,T]. Moreover, μ1(⋅) and μ2(⋅) are twice continuously differentiable with Lipschitz continuous second derivatives. It is easy to see that assumptions (A5.1) and (A5.2) are satisfied. Therefore the desired asymptotic normality follows.
In this example,
Moreover,
Thus,
2.5 Extensions
In this section, we generalize our results in the previous sections including asymptotic expansions, asymptotic normality, and exponential bounds, to the Markov chain αε ( ⋅) with generator given by \({Q}^\varepsilon (t) = Q(t)/\varepsilon +\widehat{ Q}(t)\) with weakly irreducible generator Q(t). Recall that the vector of probabilities \({p}^\varepsilon (t) = (P({\alpha }^\varepsilon (t) = 1),\ldots,P({\alpha }^\varepsilon (t) = m))\) satisfies the differential equation
To proceed, the following conditions are needed.
-
(A5.3)
Both Q(t) and \(\widehat{Q}(t)\) are generators. For each t ∈ [0, T], Q(t) is weakly irreducible.
-
(A5.4)
For some positive integer n 0, Q( ⋅) is ( n 0 + 1)-times continuously differentiable on [0, T] and \(({d}^{{n}_{0}+1}/d{t}^{{n}_{0}+1})Q(\cdot )\) is Lipschitz. Moreover, \(\widehat{Q}(\cdot )\) is n 0-times continuously differentiable on [0, T] and \(({d}^{{n}_{0}}/d{t}^{{n}_{0}})\widehat{Q}(\cdot )\) is Lipschitz.
Similarly to Section 4.2 for \(k = 1,\ldots,{n}_{0} + 1\), the outer expansions lead to equations
with constraints
and
The initial-layer correction terms are
where
with initial conditions
Theorem 5.19
Suppose that (A5.3) and (A5.4) are satisfied. Then
-
(a)
φ i ( ⋅) is \(({n}_{0} + 1 - i)\) -times continuously differentiable on [0, T],
-
(b)
for each i, there is a \(\widehat{\kappa } > 0\) such that
$${\left |{\psi }_{i}\left ({ t \over \varepsilon } \right )\right |} \leq K\exp \left (-\frac{\widehat{\kappa }t} \varepsilon \right ),\,\mbox{ and}$$ -
(c)
the approximation error satisfies
$${ \sup }_{t\in [0,T]}{\left |{p}^\varepsilon (t) -{\sum }_{i=0}^{{n}_{0} }\varepsilon ^{i}{\varphi }_{ i}(t) -{\sum }_{i=0}^{{n}_{0} }\varepsilon ^{i}{\psi }_{ i}\left ({ t \over \varepsilon } \right )\right |} \leq K\varepsilon ^{{n}_{0}+1}.$$(5.46)
The proof of this theorem is similar to those of Theorem 4.5, and is thus omitted. We also omit the proofs of the following two theorems because they are similar to that of Theorem 5.4 and Theorem 5.9, respectively.
Theorem 5.20
Suppose (A5.3) and (A5.4) are satisfied with n 0 = 0. Then there exist positive constants ε 0 and K such that for 0 < ε ≤ ε 0,\(i \in \mathcal{M}\) , and for any deterministic process β i (⋅) satisfying |β i (t)|≤ 1 for all t ≥ 0, we have
where θ T and n ε (⋅) are as defined previously.
Corollary 5.21
Consider \({Q}^\varepsilon = Q/\varepsilon +\widehat{ Q}\) with constant generators Q and \(\widehat{Q}\) such that Q is weakly irreducible. Then (5.25) and (5.27) hold with constants K and K j independent of T.
Theorem 5.22
Suppose (A5.3) and (A5.4) are satisfied with n 0 = 1. Then for t ∈ [0,T], the process n ε (⋅) converges weakly to a Gaussian process n(⋅) such that
where A(t) = (A ij (t)) with
and Q 0 (r,t) = (q 0,ij (r,t)) satisfying
with P 0 (t) = (ν′(t),…,ν′(t))′.
Remark 5.23
In view of Theorem 5.22 , the asymptotic covariance is determined by the quasi-stationary distribution ν(t) and Q 0 (r,t). Both ν(t) and Q 0 (r,t) are determined by Q(t), the dominating term in Q ε (t). In the asymptotic normality analysis, it is essential to have the irreducibility condition of Q(t), whereas the role of \(\widehat{Q}(t)\) is not as important. If Q(t) is weakly irreducible, then there exists an ε 0 > 0 such that \({Q}^\varepsilon (t) = Q(t)/\varepsilon +\widehat{ Q}(t)\) is weakly irreducible for 0 < ε ≤ ε 0, as shown in Sethi and Zhang [192, Lemma J.10].
By introducing another generator \(\widehat{Q}(t)\) , we are dealing with a singularly perturbed Markovian system with fast and slow motions. Nevertheless, the entire system under consideration is still weakly irreducible. This irreducibility allows us to extend our previous results with minor modifications.
Although most of the results in this section can be extended to the case with \({Q}^\varepsilon (t) = Q(t)/\varepsilon +\widehat{ Q}(t)\) , there are some exceptions. For example, Corollary 5.6 would not go through because even with constant matrix \(\widehat{Q}(t) =\widehat{ Q}\) , φ 1 (t) in Lemma 5.1 does not equal 0 when \(\widehat{Q}\not =0\).
One may wonder what happens if Q(t) in Q ε (t) is not weakly irreducible. In particular, one can consider the case in which Q(t) consists of several blocks of irreducible submatrices. Related results of asymptotic normality and the exponential bounds are treated in subsequent sections.
3 Markov Chains with Weak and Strong Interactions
For brevity, unless otherwise noted, in the rest of the book, whenever the phrase “weak and strong interaction” is used, it refers to the case of two-time-scale Markov chains with all states being recurrent. Similar approaches can be used for the other cases as well. The remainder of the chapter concentrates on exploiting detailed structures of the weak and strong interactions. In addition, it deals with convergence of the probability distribution with merely measurable generators.
We continue our investigation of asymptotic properties of the Markov chain αε ( ⋅) generated by Q ε( ⋅), with
where \(\widetilde{Q}(t) = \mathrm{diag}(\widetilde{{Q}}^{1}(t),\ldots,\widetilde{{Q}}^{l}(t))\) is a block-diagonal matrix such that \(\widehat{Q}(t)\) and \(\widetilde{{Q}}^{k}(t)\), for k = 1, …, l, are themselves generators. The state space of α ε ( ⋅) is given by
For each k = 1, …, l, let \({\mathcal{M}}_{k} =\{ {s}_{k1},\ldots,{s}_{k{m}_{k}}\}\), representing the group of states corresponding to \(\widetilde{{Q}}^{k}(t)\).
The results in Section 5.3.1 reveal the structures of the Markov chains with weak and strong interactions based on the following observations. Intuitively, for small ε, the Markov chain αε( ⋅) jumps more frequently within the states in \({\mathcal{M}}_{k}\) and less frequently from \({\mathcal{M}}_{k}\) to \({\mathcal{M}}_{j}\) for j ≠ k. Therefore, the states in \({\mathcal{M}}_{k}\) can be aggregated and represented by a single state k (one may view the state k as a super state). That is, one can approximate αε( ⋅) by an aggregated process, say, \({\overline{\alpha }}^\varepsilon (\cdot )\). Furthermore, by examining the tightness and finite-dimensional distribution of \({\overline{\alpha }}^\varepsilon (\cdot )\), it will be shown that \({\overline{\alpha }}^\varepsilon (\cdot )\) converges weakly to a Markov chain \(\overline{\alpha }(\cdot )\) generated by
Section 5.3.2 continues the investigation along the line of estimating the error bounds of the approximation. Our interest lies in finding how closely one can approximate an unscaled sequence of occupation measures. The study is through the examination of appropriate exponential-type bounds. To take a suitable scaled sequence, one first centers the sequence around the “mean,” and then compares the actual sequence of occupation measures with this “mean.” In contrast to the results of Section 5.2, in lieu of taking the difference of the occupation measure with that of a deterministic function, it is compared with a random process. One of the key points here is the utilization of solutions of linear time-varying stochastic differential equations, in which the stochastic integration is with respect to a square-integrable martingale.
In comparison with the central limit theorem obtained in Section 5.2, it is interesting to know whether these results still hold under the structure of weak and strong interactions. The answer to this question is in Section 5.3.3, which also contains further study on related scaled sequences of occupation measures. The approach is quite different from that of Section 5.2. We use the martingale formulation and apply the techniques of perturbed test functions. It is interesting to note that the limit process is a switching diffusion process, which does not have independent increments. When the generator is weakly irreducible as in Section 5.2, the motion of jumping around the grouped states disappears and the diffusion becomes the dominant force.
We have considered only Markov chains with smooth generators up to now. However, there are cases in certain applications in which the generators may be merely measurable. Section 5.4 takes care of the scenario in which the Markov chains are governed by generators that are only measurable. Formulation via weak derivatives is also discussed briefly. Finally the chapter is concluded with a few more remarks. Among other things, additional references are given.
3.1 Aggregation of Markov Chains
This section deals with an aggregation of αε( ⋅). The following assumptions will be needed:
-
(A5.5)
For each k = 1, …, l and t ∈ [0, T], \(\widetilde{{Q}}^{k}(t)\) is weakly irreducible.
-
(A5.6)
\(\widetilde{Q}(\cdot )\) is differentiable on [0, T] and its derivative is Lipschitz. Moreover, \(\widehat{Q}(\cdot )\) is also Lipschitz.
The assumptions above guarantee the existence of an asymptotic expansion up to zeroth order. To prepare for the subsequent study, we first provide the following error estimate. Since only the zeroth-order expansion is needed here, the estimate is confined to such an approximation. Higher-order terms can be obtained in a similar way.
Lemma 5.24
Assume (A5.5) and (A5.6) . Let P ε (t,t 0 ) denote the transition probability of α ε (⋅). Then for some κ 0 > 0,
where
where ν k (t) is the quasi-stationary distribution of \(\widetilde{{Q}}^{k}(t)\) , and Θ(t,t 0) \(= ({{\vartheta}}_{ij}(t,{t}_{0})) \in {\mathbb{R}}^{l\times l}\) is the solution to the following initial value problem:
Proof: The proof is similar to those of Lemma 5.1 and Theorem 4.29, except that the notation is more involved. □
Define an aggregated process of αε ( ⋅) on [0, T] by
The idea to follow is to treat a related Markov chain having only l states. The transitions among its states correspond to the jumps from one group \({\mathcal{M}}_{k}\) to another \({\mathcal{M}}_{j}\),j≠k, in the original Markov chain.
Theorem 5.25
Assume (A5.5) and (A5.6) . Then, for any i = 1,…,l, j = 1,…,m i , and bounded and measurable deterministic function β ij (⋅),
Proof: For any i, j and 0 ≤ t ≤ T, let
We have suppressed the i, j dependence of ηε( ⋅) for notational simplicity. Loosely speaking, the argument used below is a Liapunov stability one, and ηε( ⋅) can be viewed as a Liapunov function. By differentiating ηε( ⋅), we have
The definition of \({\overline{\alpha }}^\varepsilon (\cdot )\) yields that \(\{{\overline{\alpha }}^\varepsilon (t) = i\} =\{ {\alpha }^\varepsilon (t) \in {\mathcal{M}}_{i}\}\). Thus,
where \({\Phi }^\varepsilon (t,r) = {\Phi }_{1}^\varepsilon (t,r) + {\Phi }_{2}^\varepsilon (t,r)\) with
and
Note that the Markov property of αε( ⋅) implies that for 0 ≤ r ≤ t,
In view of the asymptotic expansion, we have
It follows that
Combining (5.55) and (5.56) leads to
Similarly, we can show that
by noting that
and
for any k = 1, …, m i . Therefore,
This together with ηε (0) = 0 implies that η ε(t) = O(ε). □
Theorem 5.25 indicates that νk(t) together with \({\overline{\alpha }}^\varepsilon (\cdot )\) approximates well the Markov chain αε( ⋅) in an appropriate sense. Nevertheless, in general, {αε( ⋅)} is not tight. The following example provides a simple illustration.
Example 5.26
Let α ε (⋅) ∈{ 1,2} denote a Markov chain generated by
for some λ,μ > 0. Then α ε (⋅) is not tight.
Proof: If α ε (⋅) is tight, then there exists a sequence ε k → 0 such that \({\alpha }^{\varepsilon _{k}}(\cdot )\) converges weakly to a stochastic process \(\alpha (\cdot ) \in D([0,T];\mathcal{M})\) . In view of the Skorohod representation (without changing notation for simplicity), Theorem A.11 , we may assume \({\alpha }^{\varepsilon _{k}}(\cdot ) \rightarrow \alpha (\cdot )\) w.p.1. It follows from Lemma A.41 that
for all t ∈ [0,T]. Moreover, similarly as in Theorem 5.25 , we obtain
where (ν 1 ,ν 2 ) is the stationary distribution of α ε (⋅) and ν 1 + 2ν 2 is the mean with respect to the stationary distribution. As a consequence, it follows that \(\alpha (t) = {\nu }_{1} + 2{\nu }_{2}\) for all t ∈ [0,T] w.p.1. Let
Then for t ∈ [0,T],
Hence, under the Skorohod topology
This contradicts the fact that \({\alpha }^{\varepsilon _{k}}(\cdot ) \rightarrow \alpha (\cdot ) = {\nu }_{ 1} + 2{\nu }_{2}\) w.p.1. Therefore, α ε(⋅) cannot be tight. □
Although αε ( ⋅) is not tight because it fluctuates in \({\mathcal{M}}_{k}\) very rapidly for small ε, its aggregation \({\overline{\alpha }}^\varepsilon (\cdot )\) is tight, and converges weakly to \(\overline{\alpha }(t)\),t ≥ 0, a Markov chain generated by \(\overline{Q}(t)\), t ≥ 0, where \(\overline{Q}(t)\) is defined in (5.48). The next theorem shows that \({\overline{\alpha }}^\varepsilon (\cdot )\) can be further approximated by \(\overline{\alpha }(\cdot )\).
Theorem 5.27
Assume (A5.5) and (A5.6) . Then \({\overline{\alpha }}^\varepsilon (\cdot )\) converges weakly to \(\overline{\alpha }(\cdot )\) in \(D([0,T];\overline{\mathcal{M}})\) , as ε → 0.
Proof: The proof is divided into two steps. First, we show that \({\overline{\alpha }}^\varepsilon (\cdot )\) defined in (5.51) is tight in \(D([0,T];\overline{\mathcal{M}})\). The definition of \({\overline{\alpha }}^\varepsilon (\cdot )\) implies that
Consider the conditional expectation
Since \(\{{\overline{\alpha }}^\varepsilon (t + s) = k\} =\{ {\alpha }^\varepsilon (t + s) \in {\mathcal{M}}_{k}\}\), it follows that
Therefore, we obtain
Note that \(\lim \limits_{t\rightarrow 0}{{\vartheta}}_{ik}(t + s,s) = 0\) for i≠ k.
Thus, the Markov property of αε( ⋅) implies
Recall that \({\overline{\alpha }}^\varepsilon (\cdot )\) is bounded. The tightness of \({\overline{\alpha }}^\varepsilon (\cdot )\) follows from Kurtz’ tightness criterion (see Lemma A.17).
To complete the proof, it remains to show that the finite-dimensional distributions of \({\overline{\alpha }}^\varepsilon (\cdot )\) converge to that of \(\overline{\alpha }(\cdot )\). In fact, for any
we have
In view of Lemma 5.24, for each k, we have
Moreover, note that
It follows that
where \({\sum }_{{j}_{1},\ldots,{j}_{n}} = \sum \limits_{{j}_{1}=1}^{{m}_{{i}_{1}}}\cdots {\sum }_{{j}_{ n}=1}^{{m}_{{i}_{n}}}\) and \(\widetilde{{{\vartheta}}}_{{i}_{1}}({t}_{1})\) denotes the initial distribution (also known as absolute probability in the literature of Markov chains). Thus, \({\overline{\alpha }}^\varepsilon (\cdot ) \rightarrow \overline{\alpha }(\cdot )\) in distribution. □
This theorem implies that \({\overline{\alpha }}^\varepsilon (\cdot )\) converges to a Markov chain, although \({\overline{\alpha }}^\varepsilon (\cdot )\) itself is not a Markov chain in general. If, however, the generator Q ε(t) has some specific structure, then \({\overline{\alpha }}^\varepsilon (\cdot )\) is a Markov chain. The following example demonstrates this point.
Example 5.28
Let \(\widetilde{Q}(t) = (\widetilde{{q}}_{ij}(t))\) and \(\overline{Q}(t) = ({\overline{q}}_{ij}(t))\) denote generators with the corresponding state spaces \(\{{a}_{1},\ldots,{a}_{{m}_{0}}\}\) and {1,…,l}, respectively. Consider
where \({I}_{{m}_{0}}\) is the m 0 × m 0 identity matrix. In this case
Then \({\overline{\alpha }}^\varepsilon (\cdot )\) is a Markov chain generated by \(\overline{Q}(t)\) . In fact, let
Note that s ij = (i,a j ) for j = 1,…,m 0 and i = 1,…,l. In view of Lemma 2.4 , we obtain that
is a martingale. Postmultiplying (multiplying from the right) (5.60) by
and noting that \(\{{\overline{\alpha }}^\varepsilon (t) = i\} =\{ {\alpha }^\varepsilon (t) \in {\mathcal{M}}_{i}\}\) and
we obtain that
is still a martingale. In view of the special structure of Q ε (t) in (5.59),
and
Therefore, (5.60) implies that
is a martingale. This implies, in view of Lemma 2.4, that \({\overline{\alpha }}^\varepsilon (\cdot )\) is a Markov chain generated by \(\overline{Q}(t)\) , t ≥ 0.
3.2 Exponential Bounds
For each i = 1, …, l, j = 1, …, m i , \(\alpha \in \mathcal{M}\), and t ≥ 0, let β ij (t) be a bounded, Borel measurable, deterministic function and let
Consider normalized occupation measures
where
In this section, we establish the exponential error bound for n ε( ⋅), a sequence of suitably scaled occupation measures for the singularly perturbed Markov chains with weak and strong interactions.
In view of Theorem 4.29, there exists κ0 > 0 such that
Similar to Section 5.3.2, for fixed but otherwise arbitrary T > 0, let
We may write (5.62) in terms of K T and O 1( ⋅) as follows:
where | O 1(y) | ∕ | y | ≤ 1. The notation of K T and O 1( ⋅) above emphasizes the separation of the dependence of the constant and a “norm 1” function. Essentially, K T serves as a magnitude of the bound indicating the size of the bounding region, and the rest is absorbed into the function O 1( ⋅).
Theorem 5.29
Assume (A5.5) and (A5.6) . Then there exist ε 0 > 0 and K > 0 such that for 0 < ε ≤ ε 0 , T ≥ 0, and for any bounded, Borel measurable, and deterministic process β ij (⋅),
where θ T is any constant satisfying
and where |⋅| T denotes the matrix norm as defined in (5.12) , that is,
similarly for \(\vert \widehat{Q}{\vert }_{T}\).
Remark 5.30
This theorem is a natural extension to Theorem 5.4 . Owing to the existence of the weak and strong interactions, slightly stronger conditions on K T and θ T are made in (5.63) and (5.66). Also the exponential constant in (5.65) is changed to (T + 1)3.
Proof of Theorem 5.29: Here the proof is again along the lines of Theorem 5.4. Since Steps 2-5 in the proof are similar to those of Theorem 5.4, we will only give the proof for Step 1.
Let χε ( ⋅) denote the vector of indicators corresponding to α ε ( ⋅), that is,
Then w ε( ⋅) defined by
is an \({\mathbb{R}}^{m}\)-valued martingale. In fact, w ε ( ⋅) is square integrable on [0, T]. It then follows from a well-known result (see Elliott [55] or Kunita and Watanabe [134]) that a stochastic integral with respect to w ε(t) can be defined. In view of the defining equation (5.67), the linear stochastic differential equation
makes sense. Recall that P ε(t, s) is the principal matrix solution of the matrix differential equation
The solution of this stochastic differential equation is
Use \({{\vartheta}}_{ij}(t,s)\) defined in Lemma 5.24 and write \(\Theta (t,s) = ({{\vartheta}}_{ij}(t,s))\) . Then it is easy to check that
Set
and
Then it follows that
Moreover, postmultiplying both sides of (5.67) by \(\widetilde{\mathrm{1}\mathrm{l}}\) yields that
Here \({w}^\varepsilon (\cdot )\widetilde{\mathrm{1}\mathrm{l}}\) is also a square-integrable martingale. Note that \(\widetilde{Q}(s)\widetilde{\mathrm{1}\mathrm{l}} = 0\) and hence
We obtain from (5.73) that
Since Θ(t, s) is the principal matrix solution to
similar to (5.68), solving the stochastic differential equation for \(\widetilde{{\chi }}^\varepsilon (\cdot )\) leads to the equation:
Let us now return to the last two terms in (5.70) and use (5.71), (5.72), and (5.74) to obtain
Combining this with (5.70), we have
where
Note that the matrix P ε(t, s) is invertible but P 0(t, s) is not. The idea is to approximate the noninvertible matrix P 0(t, s) by the invertible P ε(t, s). Let
and
Then ϕε(0) = 0 and ϕε(t) satisfies the following equation:
The properties of the principal matrix solution imply that
Set
Owing to the properties of the principal matrix solution, for any t ∈ [0, T], we have
ψ ε (0) = 0 and ψ ε(t) satisfies the equation
The solution to this equation is given by
where \(\check{{\Phi }}^\varepsilon (t,s)\) is the principal matrix solution to
Postmultiplying both sides of (5.78) by P ε(t, 0) yields
where
Thus it follows that
Again using (5.77), we have
This implies that \(\check{{\Psi }}^\varepsilon (t,s)\) is the principal matrix solution to the differential equation
Therefore, all entries of \(\check{{\Psi }}^\varepsilon (t,s)\) are bounded below from 0 and bounded above by 1, and these bounds are uniform in 0 ≤ s ≤ t ≤ T. Thus, \(\vert \check{{\Psi }}^\varepsilon (t,s){\vert }_{T} \leq 1\).
Multiplying both sides of (5.79) by the m ×m matrix
from the right and integrating over the interval [0, ς], for each ς ∈ [0, T], we have
By changing the order of integration, we write the last term in the above expression as
Therefore, it follows that
where
Moreover, in view of the fact that \(\vert \check{{\Psi }}^\varepsilon (t,s){\vert }_{T} \leq 1\), it is easy to see that
Note that n ε ( ⋅) can be written in terms of χ ε ( ⋅) and \({\overline{\chi }}^\varepsilon (\cdot )\) as
By virtue of (5.81), it follows that
Note that in view of the definition of η1 ε( ⋅) in (5.76),
Thus, in view of (5.82),
Thus, in view of (5.63) and (5.66), for some ε0 > 0, and all 0 < ε ≤ ε 0,
Moreover, using (5.64), as in the proof of Theorem 5.4, we obtain that
for
Finally, combine (5.81), (5.83), and (5.85) to obtain
This completes the proof. □
Remark 5.31
It is easily seen that the error bound so obtained has a form similar to that of the martingale inequality. If nε(⋅) were a martingale, the inequality would be obtained much more easily since exp (⋅) is a convex function. As in Section 5.2, the error bound is still a measure of “goodness” of approximation. However, one cannot compare the unscaled occupation measures with a deterministic function. A sensible alternative is to use an approximation by the aggregated process that is no longer deterministic. The exponential bounds obtained tell us exactly how closely one can carry out the approximation. It should be particularly useful for many applications in stochastic control problems with Markovian jump disturbances under discounted cost criteria.
The next two corollaries show that the error bound can be improved under additional conditions by having smaller exponential constants, e.g., \({(T + 1)}^{3/2}\) or \({(T + 1)}^{5/2}\) instead of ( T + 1)3.
Corollary 5.32
Assume that the conditions of Theorem 5.29 hold. Let \(\widetilde{Q}(t) = (\widetilde{{q}}_{ij}(t))\) and \(\overline{Q}(t) = ({\overline{q}}_{ij}(t))\) denote generators with the corresponding state spaces \(\{{a}_{1},\ldots,{a}_{{m}_{0}}\}\) and {1,…,l}, respectively. Consider
where \({I}_{{m}_{0}}\) is the m 0 × m 0 identity matrix. Then there exist positive constants ε 0 and K such that for 0 < ε ≤ ε 0 , and T ≥ 0,
Proof: Under the special structure of the generator Q ε, it is easy to see that
where \(\widetilde{\mathrm{1}\mathrm{l}}\) now takes the form
Note that under current conditions on the fast-changing part of the generator \(\widetilde{Q}(t)\),
where I l denotes the l-dimensional identity matrix. This together with (5.72) implies that
It follows from (5.71) that
Then (5.75) becomes
The rest of the proof follows exactly that of Theorem 5.29. □
Corollary 5.33
Assume the conditions of Theorem 5.29 . Suppose \(\widetilde{Q}(t) =\widetilde{ Q}\) and \(\widehat{Q}(t) =\widehat{ Q}\) for some constant matrices \(\widetilde{Q}\) and \(\widehat{Q}\) . Then there exist positive constants ε 0 and K such that for 0 < ε ≤ ε 0 , and T ≥ 0,
Remark 5.34
Note that in view of Corollary 4.31 , one can show under the condition \(\widetilde{Q}(t) =\widetilde{ Q}\) and \(\widehat{Q}(t) =\widehat{ Q}\) that there exists a constant K such that
In this case, θ T can be taken as
That is, compared with the general result, the constant K T can be further specified as \({K}_{T} = K(T + 1)\).
Proof of Corollary 5.33: Note that when the generators are time independent, the quasi-stationary distribution νi(t) is also independent of time and is denoted by νi. In this case, the argument from (5.75) to (5.80) can be replaced by the following. Let
Then it can be shown that
This implies that
Let \({\phi }^\varepsilon (t) = ({\chi }^\varepsilon (t) -{\overline{\chi }}^\varepsilon (t)) - {\eta }^\varepsilon (t)\). Then ϕε( ⋅) satisfies the equation
Solving for ϕε( ⋅), we obtain
Writing \({\chi }^\varepsilon (t) -{\overline{\chi }}^\varepsilon (t)\) in terms of ϕε (t) and η ε(t) yields,
The rest of the proof follows that of Theorem 5.29. □
Similar to Section 5.2, we derive estimates that are analogous to Corollary 5.7 and Corollary 5.8. The details are omitted, however.
3.3 Asymptotic Distributions
In Section 5.2, we obtained a central limit theorem for a class of Markov chains generated by \({Q}^\varepsilon (t) = Q(t)/\varepsilon +\widehat{ Q}(t)\) with a weakly irreducible Q( t). In this case for sufficiently small ε > 0, Q ε(t) is weakly irreducible. What, if anything, can be said about the weak and strong interaction models, when \(\widetilde{Q}(t)\) is not weakly irreducible? Is there a central limit theorem for the corresponding occupation measure when one has a singularly perturbed Markov chain with weak and strong interactions? This section deals with such an issue; our interest lies in the asymptotic distribution as ε → 0. It is shown that the asymptotic distribution of the corresponding occupation measure can be obtained. However, the limit distribution is no longer Gaussian, but a Gaussian mixture, and the proof is quite different from that of the irreducible case in Section 5.2.
For each i = 1, …, l, j = 1, …, m i , \(\alpha \in \mathcal{M}\) , and t ≥ 0, let β ij (t) be a bounded Borel measurable deterministic function. Use W ij (t, α) defined in (5.61) and the normalized occupation measure
with
We will show in this section that n ε( ⋅) converges weakly to a switching diffusion modulated by \(\overline{\alpha }(\cdot )\). The procedure is as follows:
-
(a)
Show that \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\) is tight;
-
(b)
verify that the limit of a subsequence of \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\) is a solution to a martingale problem that has a unique solution;
-
(c)
characterize the solution of the associated martingale problem;
-
(d)
construct a switching diffusion that is also a solution to the martingale problem and therefore the limit of \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\).
To accomplish our goal, these steps are realized by proving a series of lemmas. Recall that \({\mathcal{F}}_{t}^\varepsilon = \sigma \{{\alpha }^\varepsilon (s) :\; 0 \leq s \leq t\}\) denotes the filtration generated by αε( ⋅). The lemma below is on the order estimates of the conditional moments, and is useful for getting the tightness result in what follows.
Lemma 5.35
Assume (A5.5) and (A5.6) . Then for all 0 ≤ s ≤ t ≤ T and ε small enough, the following hold:
Proof: First, note that for any fixed i, j,
Moreover, in view of the definition of W ij (t, α) and the Markov property, we have, for 0 ≤ s ≤ r,
In view of Lemma 5.24, in particular, similar to (5.55) and (5.56), for all i 0 = 1, …, l and \({j}_{0} = 1,\ldots,{m}_{{i}_{0}}\),
Thus owing to Lemma A.42, we have
Note also that
This implies (a).
To verify (b), fix and suppress i, j and define
Then by the definition of n ij ( ⋅),
In accordance with the definition of \({\overline{\alpha }}^\varepsilon (\cdot )\),\({\overline{\alpha }}^\varepsilon (t) = i\) iff \({\alpha }^\varepsilon (t) \in {\mathcal{M}}_{i}\) . In what follows, we use \({\alpha }^\varepsilon (t) \in {\mathcal{M}}_{i}\) and \({\overline{\alpha }}^\varepsilon (t) = i\) interchangeably. Set
Then as in the proof of Theorem 5.25,
Using Lemma 5.24, we obtain
for all i 0 = 1, …, l and \({j}_{0} = 1,\ldots,{m}_{{i}_{0}}\). Then from Lemma A.42, we obtain
As a consequence, we have
Integrating both sides over [s, t] and recalling ηε(s) = 0 yields
This completes the proof of the lemma. □
The next lemma is concerned with the tightness of \(\{({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\}\).
Lemma 5.36
Assume (A5.5) and (A5.6) . Then \(\{({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\}\) is tight in \(D([0,T]; {\mathbb{R}}^{m} \times \overline{\mathcal{M}})\).
Proof: The proof uses Lemma A.17. We first verify that the condition given in Remark A.18 holds. To this end, note that \(0 \leq {\overline{\alpha }}^\varepsilon (t) \leq l\) for all t ∈ [0, T]. Moreover, by virtue of Theorem 5.25, for each δ > 0 and each rational t ≥ 0,
where the last inequality is due to Theorem 5.25. Thus if we choose \({K}_{t,\delta } > \sqrt{KT/\delta }\), (A.6) will follow.
It follows from Lemma 5.35 and (5.58) that for all t ∈ [0, T],
Using (5.86) and (5.87), Theorem A.17 yields the desired result. □
The tightness of \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\) and Prohorov’s theorem allow one to extract convergent subsequences. We next show that the limit of such a subsequence is uniquely determined in distribution. An equivalent statement is that the associated martingale problem has a unique solution. The following lemma is a generalization of Theorem 5.25 and is needed for proving such a uniqueness property.
Lemma 5.37
Let ξ(t,x) be a real-valued function that is Lipschitz in (t,x) \(\in {\mathbb{R}}^{m+1}\) . Then
where \({W}_{ij}(t,\alpha ) = ({I}_{\{\alpha ={s}_{ij}\}} - {\nu }_{j}^{i}(t){I}_{\{\alpha \in {\mathcal{M}}_{i}\}}){\beta }_{ij}(t)\) as defined in (5.61).
Remark 5.38
This lemma indicates that the weighted occupation measure (with weighting function ξ(t,αε(t))) defined above goes to zero in mean square uniformly in t ∈ [0,ς]. If ξ(⋅) were a bounded and measurable deterministic function not depending on αε(⋅) or nε(⋅), this assertion would follow from Theorem 5.25 easily. In the current situation, it is a function of n ε (⋅) and therefore a function of α ε (⋅), which results in much of the difficulty. Intuitively, if we can “separate” the functions W ij (⋅) and ξ(⋅) in the sense treating ξ(⋅) as deterministic, then Theorem 5.25 can be applied to obtain the desired limit. To do so, subdivide the interval [0,ς] into small intervals so that on each of the small intervals, the two functions can be separated. To be more specific, on each partitioned interval, use a piecewise-constant function to approximate ξ(⋅), and show that the error goes to zero. In this process, the Lipschitz condition of ξ(t,x) plays a crucial role.
Proof of Lemma 5.37: For 0 < δ < 1 and 0 < ς ≤ T, let \(N = [\varsigma /\varepsilon ^{1-\delta }]\). Use a partition of [0, ς] given by
of [0, ς], where \({t}_{k} = \varepsilon ^{1-\delta }k\) for k = 0, 1, …, N and \({t}_{N+1} = \varsigma \) . Consider a piecewise-constant function
Let W ij ε(t) = W ij (t, α ε(t)). Then
We now estimate the first term on the second line above. In view of the Cauchy inequality and the boundedness of W ij ε (t), it follows, for 0 ≤ ς ≤ T, that
Note that Theorem 5.25 implies
for a positive constant K and for all t ∈ [0, T]. Therefore, in view of the Lipschitz condition of ξ( ⋅), we have
Noting that \({t}_{2} = 2\varepsilon ^{1-\delta } = O(\varepsilon ^{1-\delta })\) , it follows that
Using the definition of \(\widetilde{\xi }(t)\) , the Lipschitz property of ξ( t, x) in ( t, x), the choice of the partition of [0, ς], and Lemma 5.35, we have
Let us estimate the second term on the second line in (5.88). Set
Then the derivative of \(\widetilde{{\eta }}^\varepsilon (t)\) is given by
For 0 ≤ t ≤ t 2, in view of the Lipschitz property and Theorem 5.25, we obtain
If t k ≤ t < t k + 1, for k = 2, …, N, then using the same argument gives us
and
Recall that \({\mathcal{F}}_{t}^\varepsilon = \sigma \{{\alpha }^\varepsilon (s) :\; 0 \leq s \leq t\}\) . For \(s \leq {t}_{k-1} < {t}_{k} \leq t < {t}_{k+1}\),
Moreover, in view of the definition of \(\widetilde{\xi }(\cdot )\) and the proof of Lemma 5.35, we have for some κ0 > 0,
Combine this with (5.89) to obtain
Therefore,
uniformly on [0, T], which implies, together with \(\widetilde{{\eta }}^\varepsilon (0) = 0\), that
This completes the proof. □
To characterize the limit of \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\), consider the martingale problem associated with \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\). Note that
where
Let \({\mathcal{G}}^\varepsilon (t)\) be the operator
for all f( ⋅, ⋅, α) ∈ C 1, 1, where ∇ x denotes the gradient with respect to x and ⟨ ⋅, ⋅⟩ denotes the usual inner product in Euclidean space. It is well known that (see Davis [41, Chapter 2])
is a martingale.
We use the perturbed test function method (see Ethier and Kurtz [59] and Kushner [139]) to study the limit as ε → 0. To begin with, we define a functional space on \({\mathbb{R}}^{m} \times \overline{\mathcal{M}}\)
For any real-valued function \({f}^{0}(\cdot,i) \in {C}_{L}^{2}\), define
and consider the function
where h( t, x, α) is to be specified later. The main idea is that by appropriate choice of h( ⋅), the perturbation is small and results in the desired cancelation in the calculation.
In view of the block-diagonal structure of \(\widetilde{Q}(t)\) and the definition of \(\overline{f}(x,\alpha )\), it is easy to see that
Applying the operator \({\mathcal{G}}^\varepsilon (t)\) to the function f( ⋅) defined in (5.92) yields that
defines a martingale.
The basic premise of the perturbed test function method is to choose the function h( ⋅) that cancels the “bad” terms of order 1 ∕ \sqrt{ε}:
Note that as mentioned previously, \(\widetilde{Q}(t)\) has rank m − l. Thus the dimension of the null space is l; that is, \(N(\widetilde{Q}(t)) = l\). A crucial observation is that in view of the Fredholm alternative (see Lemma A.37 and Corollary A.38), a solution of (5.93) exists iff the matrix \((\langle W(s,{s}_{ij}),{\nabla }_{x}\overline{f}(x,{s}_{ij})\rangle )\) is orthogonal to \(\widetilde{\mathrm{1}{\mathrm{l}}}_{{m}_{1}},\ldots,\widetilde{\mathrm{1}{\mathrm{l}}}_{{m}_{l}}\) , the span of \(N(\widetilde{Q}(t))\) (see Remark 4.23 for the notation). Moreover, since f 0( ⋅, i) is C L 2,h( ⋅) can be chosen to satisfy the following properties assuming β ij ( ⋅) to be Lipschitz on [0, T]:
Such an \(h(\cdot )\) leads to
being a martingale. For each s, x, α, define
With f 0 ∈ C L 2 , it is easy to see that g(s, x, α) is Lipschitz in (s, x). This function will be used in defining the operator for the limit problem later.
Remark 5.39
Note that the choice of h(⋅) in (5.93) is not unique. If h 1 (⋅) and h 2(⋅) are both solutions to (5.93), then the irreducibility of \(\widetilde{{Q}}^{i}(s)\) implies that, for each i = 1,…,l,
for some scalar functions h 0 (s,x,i). Although the choice of h is not unique, the resulting function g(s,x,α) is well defined. As in Remark 4.23, the consistency condition or solvability condition due to Fredholm alternative is in force. Therefore, if h 1 and h 2 are both solutions to (5.93), then
for \(\alpha \in {\mathcal{M}}_{i}\) and i = 1,…,l.
Using g(s, x, α) defined above, we obtain
In view of Lemma 5.37, the term in the fourth line above goes to zero in mean square uniformly in t ∈ [0, T]. Let
Then it follows that
Therefore, as ε → 0, we have
uniformly in t ∈ [0, T].
Furthermore, we have
Again, Lemma 5.37 implies that the third line above goes to 0 in mean square uniformly in t ∈ [0, T]. The last term above equals
where \(\overline{Q}(s) = \mathrm{diag}({\nu }^{1}(t),\ldots,{\nu }^{l}(t))\widehat{Q}(s)\widetilde{\mathrm{1}\mathrm{l}}\). It follows that as ε → 0,
uniformly in t ∈ [0, T].
We next examine the function \(\overline{g}(s,x,i)\) closely. Using the block-diagonal structure of \(\widetilde{Q}(s)\), we can write (5.93) in terms of each block \(\widetilde{{Q}}^{j}(s)\). For j = 1, …, l,
Note that \(\widetilde{{Q}}^{j}(s)\) is weakly irreducible so \(\mathrm{rank}(\widetilde{{Q}}^{j}(s)) = {m}_{j} - 1\). As in Remark 4.9, equation (5.98) has a solution since it is consistent and the solvability condition in the sense of Fredholm alternative is satisfied. We can solve (5.98) using exactly the same technique as in Section 4.2 for obtaining the φ i (t), that is, replacing one of the rows of the augmented matrix in (5.98) by (1, 1, …, 1, 0), which represents the equation \({\sum }_{k=1}^{{m}_{j}}h(s,x,{s}_{jk}) = 0\). The coefficient matrix of the resulting equation then has full rank; one readily obtains a solution. Equivalently, the solution may be written as
Note that
Recall the notation for the partitioned vector x = ( x 1, …, x l) where x j is an m j -dimensional vector and \({x}^{j} = ({x}_{1}^{j},\ldots,{x}_{{m}_{j}}^{j})\). For the partial derivatives, use the notation
Then h(s, x, s jk ) is a functional of ∂ j, j1 f 0(x, j),...\({\partial }_{j,{m}_{j}}{f}^{0}(x,j)\). It follows that g( s, x, s jk ) is a functional of \({\partial }_{j,{j}_{1}{j}_{2}}^{2}{f}^{0}(x,j)\) , for j 1, j 2 = 1, …, m j , and so is \(\overline{g}(s,x,j)\). Write
for some continuous functions \({a}_{{j}_{1}{j}_{2}}(s,j)\).
Lemma 5.40
Assume (A5.5) and (A5.6) . Suppose \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\) converges weakly to \((n(\cdot ),\overline{\alpha }(\cdot ))\) . Then for f 0 (⋅,i) ∈ C L 2,
is a martingale.
Proof: Define
The martingale property implies that
for any 0 ≤ t 1 ≤ ⋯ ≤ t k ≤ s ≤ t and any bounded and continuous functions z 1 ( ⋅), …, z k ( ⋅).
In view of the choice of h( ⋅), it follows that all the three terms
converge to 0 in mean square. Recall (5.96), (5.97), and
Denote the weak limit of H ε( ⋅) by \(\overline{H}(\cdot )\). We have
where \(\overline{H}(\cdot )\) is given by
Thus \((n(\cdot ),\overline{\alpha }(\cdot ))\) is a solution to the martingale problem. □
Lemma 5.41
Let \(\mathcal{L}\) denote the operator given by
Then the martingale problem with operator \(\mathcal{L}\) has a unique solution.
Proof: In view of Lemma A.14, we need only verify the uniqueness in distribution of \((n(t),\overline{\alpha }(t))\) for each t ∈ [0, T]. Let
where \(\theta \in {\mathbb{R}}^{m}\),\({\theta }_{0} \in \mathbb{R}\),\(j \in \mathcal{M}\) , and ι is the pure imaginary number with \({\iota }^{2} = -1\).
For fixed j 0, k 0 , let \({F}_{{j}_{0}{k}_{0}}(x,j) = {I}_{\{j={j}_{0}\}}f(x,{k}_{0})\) . Then
Moreover, note that
Furthermore, we have
Let
Then in view of (5.100) and (5.101),
Let
Rewrite (5.102) in terms of ϕ( ⋅) as
where ϕ(0) = (ϕ jk (0)) with \({\phi }_{jk}(0) = E{I}_{\{\overline{\alpha }(0)=j\}}f(0,k)\), and B( t) is a matrix-valued function whose entries are defined by the integrand of (5.102). The equation for ϕ(t) is a linear ordinary differential equation. It is well known that such a differential equation has a unique solution. Hence, ϕ(t) is uniquely determined. In particular,
is uniquely determined for all θ, θ0, so is the distribution of \((n(t),\overline{\alpha }(t))\) by virtue of the uniqueness theorem and the inversion formula of the characteristic function (see Chow and Teicher [30]). □
The tightness of \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\) together with Lemma 5.40 and Lemma 5.41 implies that \(({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\) converges weakly to \((n(\cdot ),\overline{\alpha }(\cdot ))\). We will show that n( ⋅) is a switching diffusion, i.e., a diffusion process modulated by a Markov process such that the covariance of the diffusion depends on the Markov jump process. Precisely, owing to the presence of the jump Markov chains, the limit process does not possess the independent increment property shared by many processes. A moment of reflection reveals that, necessarily, the coefficients in \(\overline{g}(s,x,i)\) must consist of a symmetric nonnegative definite matrix serving as a covariance matrix. The following lemma verifies this assertion.
Lemma 5.42
For s ∈ [0,T] and j = 1,…,l, the matrix
is symmetric and nonnegative definite.
Proof: Let \({\eta }^{j} = ({\eta }_{j1},\ldots,{\eta }_{j{m}_{j}})^{\prime}\) and \({x}^{j} = ({x}_{j1},\ldots,{x}_{j{m}_{j}})^{\prime}\). Define
Then the corresponding \(\overline{g}(\cdot )\) defined in (5.99) has the following form:
Moreover, let f j (x, k) = f j (x), independent of k. Then for all k = 1, …, l,
To verify the nonnegativity of A(s, j), it suffices to show that
for all 0 ≤ s ≤ t ≤ T. Recall that f j (x) is a quadratic function. In view of (5.94) and the proof of Lemma 5.40, it then follows that
We are in a position to show that the limit is nonnegative. Let
Then
For t ≥ s ≥ 0, using
we have
We next show that the last term goes to 0 as ε → 0. In fact, in view of (a) in Lemma 5.35, it follows that
and hence
Using (b) in Lemma 5.35, we derive the following inequalities
The Cauchy–Schwarz inequality then leads to
As a result for some K > 0, we have
as ε → 0. The nonnegativity of A(s, j) follows.
To show that A( s, j) is symmetric, consider
Then, we have
for all t ∈ [0, T]. Thus, A(s, j) is symmetric. □
Next, we derive an explicit representation of the nonnegative definite matrix A( s, j) similar to that of Theorem 5.9. Recall that given a function f 0( ⋅), one can find h( ⋅) as in (5.93). Using this h( ⋅), one defines f( ⋅) as in (5.95) which leads to \(\overline{g}(\cdot )\) given in (5.99). In view of the result in Theorem 5.9 for a single block of the irreducible matrix \(\widetilde{{Q}}^{j}(t)\) together with the computations of \(\overline{g}(s,x,j)\) , it follows that A(s, j) = 2A 0(s, j), where
with
and
Applying Lemma 5.42 to the case of \(\widetilde{Q}(s)\) a single block irreducible matrix \(\widetilde{{Q}}^{j}(s)\) , it follows that A 0(s, j) is symmetric and nonnegative definite. Hence, standard results in linear algebra yield that there exists an m j ×m j matrix σ 0 (s, j) such that
Note that the definition of \(\overline{g}(s,x,j)\) is independent of \(\widehat{Q}(t)\), so for determining A 0(s, j), we may consider \(\widehat{Q}(t) = 0\) . Note also that
The foregoing statements suggest that in view of (5.104), the desired covariance matrix is given by
where \({0}_{{m}_{k}\times {m}_{k}}\) is the m k ×m k zero matrix. That is, it is a matrix with the jth block-diagonal submatrix equal to σ0(s, j) and the rest of its elements equal to zero.
Theorem 5.43
Assume that (A5.5) holds. Suppose \(\widetilde{Q}(\cdot )\) is twice differentiable with Lipschitz continuous second derivative and \(\widehat{Q}(\cdot )\) is differentiable with Lipschitz continuous derivative. Let β ij (⋅) be bounded and Lipschitz continuous deterministic functions. Then n ε (⋅) converges weakly to a switching diffusion n(⋅), where
and w(⋅) is a standard m-dimensional Brownian motion.
Proof: Let
and \(\overline{\alpha }(\cdot )\) be a Markov chain generated by \(\overline{Q}(t)\). Then for all f 0 ( ⋅, i) ∈ C L 2,
is a martingale. This and the uniqueness of the martingale problem in Lemma 5.41 yields that \((\widetilde{n}(\cdot ),\overline{\alpha }(\cdot ))\) has the same probability distribution as \((n(\cdot ),\overline{\alpha }(\cdot ))\). This proves the theorem. □
Remark 5.44
Note that the Lipschitz condition on β ij(⋅) is not required in analyzing the asymptotic normality in Section 5.3.3. It is needed in this section because the perturbed test function method typically requires smoothness conditions of the associated processes.
It appears that the conditions in (A5.5) and (A5.6) together with the Lipschitz property of β ij (⋅) are sufficient for the convergence of n ε (⋅) to a switching diffusion n(⋅). The additional assumptions on further derivatives of \(\widetilde{Q}(\cdot )\) and \(\widehat{Q}(\cdot )\) are needed for computing the covariance of the limit process n(⋅).
Remark 5.45
If \(\overline{\alpha }(\cdot )\) were a deterministic function, n(⋅) above would be a diffusion process in the usual sense. However since the limit \(\overline{\alpha }(\cdot )\) is a Markov chain, the diffusion process is modulated by this jump process; the resulting distribution has the features of the “continuous” diffusion process and the “discrete” Markov chain limit.
In this section, we use the perturbed test function method, which is quite different from the approach of Section 5.2. The method used in that section, which might be called a direct approach, is interesting in its own right and makes a close connection between asymptotic expansion and asymptotic normality. It is effective whenever it can be applied. One of the main ingredients is that the direct approach makes use of the mixing properties of the scaled occupation measures heavily. In fact, using asymptotic expansion, it was shown that the scaled sequence of occupation measures is a mixing process with exponential mixing rate. For the weak and strong interaction cases presented, the mixing condition, and even approximate mixing conditions, no longer hold. To illustrate, consider Example 4.20 with constant jump rates and calculate
By virtue of the proof of Theorem 5.25 , a straightforward but tedious calculation shows that
for the weak and strong interaction models because E[n ε,′ (s)(n ε (t) − n ε (s))] depends on P 1 (t,s), generally a nonzero function. A direct consequence is that the limit process does not have independent increments in general. It is thus difficult to characterize the limit process via the direct approach. The perturbed test function method, on the other hand, can be considered as a combined approach. It uses enlarged or augmented states by treating the scaled occupation measure n ε (⋅) and the Markov chain α ε (⋅) together. That is, one considers a new state variable with two components (x,α). This allows us to bypass the verification of mixing-like properties such that the limit process is characterized by means of solutions of appropriate martingale problems via perturbed test functions, which underlies the rationale and essence of the approach. As a consequence, the limit process is characterized via the limit of the underlying sequence of operators.
Note that if \(\widetilde{Q}(t)\) itself is weakly irreducible (i.e., \(\widetilde{Q}(t)\) consists of only one block), then the covariance matrix is given by (5.30). In this case, since there is only one group of recurrent states, the jump behavior due to the limit process \(\overline{\alpha }(\cdot )\) will disappear. Moreover, owing to the fast transition rate \(\widetilde{Q}(t)/\varepsilon \), the singularly perturbed Markov chain rapidly reaches its quasi-stationary regime. As a result, the jump behavior does not appear in the asymptotic distribution, and the diffusion becomes the dominant factor. Although the method employed in this chapter is different from that of Section 5.2, the result coincides with that of Section 5.2 under irreducibility. We state this in the following corollary.
Corollary 5.46
Assume that the conditions of Theorem 5.43 are fulfilled with l = 1 (i.e., \(\widetilde{Q}(t)\) has only one block). Then n ε (⋅) converges weakly to the diffusion process
where w(⋅) is an m-dimensional standard Brownian motion with covariance
given by (5.30).
To further illustrate, consider the following example. This problem is concerned with a singularly perturbed Markov chain with four states divided into two groups. It has been used in modeling production planning problems with failure-prone machines. As was mentioned, from a modeling point of view, it may be used to depict the situation that two machines operate in tandem, in which the operating conditions (the machine capacity) of one of the machines change much faster than the other; see also related discussions in Chapters 7 and 8.
Example 5.47
Let α ε (⋅) be a Markov chain generated by
Then
Let \(\overline{\alpha }(\cdot )\) be a Markov chain generated by \(\overline{Q}(t)\) , t ≥ 0. In this example,
and
The limit of n ε (⋅) is given by
where w(⋅) is a standard Brownian motion taking values in \({\mathbb{R}}^{4}\).
4 Measurable Generators
In Section 4.2, we considered the asymptotic expansions of probability distributions. A natural requirement of such expansions is that the generator Q ε(t) be smooth enough to establish the desired error bounds. It would be interesting to consider the case in which the generator Q ε(t), t ≥ 0, is merely measurable. The method used in this section is very useful in some manufacturing problems; see Sethi and Zhang [192]. Moreover, the results are used in Section 8.6 to deal with a control problem under relaxed control formulation. Given only the measurability of Q ε(t), there seems to be little hope to obtain an asymptotic expansion. Instead of constructing an asymptotic series of the corresponding probability distribution, we consider the convergence of P(α ε(t) = s ij ) under the framework of convergence of
Since the phrase “weak convergence” is reserved throughout the book for the convergence of probability measures, to avoid confusion, we refer to the convergence above as convergence in the weak sense on \({L}^{2}([0,T]; \mathbb{R})\) or convergence under the weak topology of \({L}^{2}([0,T]; \mathbb{R})\).
4.1 Case I: Weakly Irreducible \(\widetilde{Q}(t)\)
Let \({\alpha }^\varepsilon (\cdot ) \in \mathcal{M} =\{ 1,\ldots,m\}\) denote the Markov chain generated by
where both \(\widetilde{Q}(t)\) and \(\widehat{Q}(t)\) are generators.
We assume the following conditions in this subsection.
-
(A5.7)
\(\widetilde{Q}(t)\) and \(\widehat{Q}(t)\) are bounded and Borel measurable. Moreover, \(\widetilde{Q}(t)\) is weakly irreducible.
Remark 5.48
In fact, both the boundedness and the Borel measurability in (A5.7) are redundant. Recall that our definition of generators (see Definition 2.2) uses the q-Property, which includes both the Borel measurability and the boundedness. Thus, (A5.7) requires only weak irreducibility. Nevertheless, we retain both boundedness and measurability for those who read only this section. Similar comments apply to assumption (A5.8) in what follows.
Define the probability distribution vector
and the transition matrix
Then using the martingale property in Lemma 2.4, we have
and
The next two lemmas are concerned with the asymptotic properties of p ε(t) and P ε(t, s).
Lemma 5.49
Assume (A5.7) . Then for each i, j, and T > 0, P(α ε (t) = i) and \(P({\alpha }^\varepsilon (t) = i\vert {\alpha }^\varepsilon (s) = j)\) both converge weakly to ν i (t) on \({L}^{2}([0,T]; \mathbb{R})\) and \({L}^{2}([s,T]; \mathbb{R})\) , respectively, that is, as ε → 0,
and
for all \(f(\cdot ) \in {L}^{2}([0,T]; \mathbb{R})\) and \({L}^{2}([s,T]; \mathbb{R})\) , respectively.
Proof: We only verify (5.108); the proof of (5.109) is similar. Recall that
Since \({p}^\varepsilon (\cdot ) \in {L}^{2}([0,T]; {\mathbb{R}}^{m})\) (space of square-integrable functions on [0, T] taking values in \({\mathbb{R}}^{m}\) ), for each subsequence of ε → 0 there exists (see Lemma A.36) a further subsequence of ε → 0 (still denoted by ε for simplicity), and for such ε, the corresponding { p ε( ⋅)} converges (in the weak sense on \({L}^{2}([0,T]; {\mathbb{R}}^{m})\)) to some \(p(\cdot ) = ({p}_{1}(\cdot ),\ldots,{p}_{m}(\cdot )) \in {L}^{2}([0,T]; {\mathbb{R}}^{m})\), that is,
for any \(({f}_{1}(\cdot ),\ldots,{f}_{m}(\cdot ))^{\prime} \in {L}^{2}([0,T]; {\mathbb{R}}^{m})\). Moreover,
almost everywhere. Since \(\widetilde{Q}(\cdot ) \in {L}^{2}([0,T]; {\mathbb{R}}^{m\times m})\), we have for 0 ≤ s ≤ t ≤ T,
Thus, using (5.106) we obtain
Since s and t are arbitrary, it follows immediately that
By virtue of (5.110), the irreducibility of \(\widetilde{Q}(t)\) implies p(t) = ν( t) almost everywhere. Thus the limit is independent of the chosen subsequence. Therefore, p ε ( ⋅) → ν( ⋅) in the weak sense on \({L}^{2}([0,T]; {\mathbb{R}}^{m})\). □
Theorem 5.50
Assume (A5.7) . Then for any bounded deterministic function β i (⋅) and for each \(i \in \mathcal{M}\) and t ≥ 0,
Proof: Let
Then as in the proof of Theorem 5.25, we can show that
where
By virtue of Lemma 5.49, P(αε(s) = i) → ν i (s) in the weak sense on \({L}^{2}([0,T]; \mathbb{R})\) and therefore as ε → 0,
Similarly, in view of the convergence of
under the weak topology of \({L}^{2}([r,t]; \mathbb{R})\), we have
This concludes the proof of the theorem. □
4.2 Case II: \(\widetilde{Q}(t) = \mathrm{diag}(\widetilde{{Q}}^{1}(t),\ldots,\widetilde{{Q}}^{l}(t))\)
This subsection extends the preceding result to the cases in which \(\widetilde{Q}(t)\) is a block-diagonal matrix with irreducible blocks. We make the following assumptions:
-
(A5.8)
\(\widehat{Q}(t)\) and \(\widetilde{{Q}}^{i}(t)\), for i = 1, …, l, are bounded and Borel measurable. Moreover, \(\widetilde{{Q}}^{i}(t)\),i = 1, …, l, are weakly irreducible.
Lemma 5.51
Assume (A5.8) . Then the following assertions hold:
-
(a)
For each i = 1, …, l and j = 1, …, m i , P(αε(t) = s ij ) converges in the weak sense to \({\nu }_{j}^{i}(t){{\vartheta}}^{i}(t)\) on \({L}^{2}([0,T]; \mathbb{R})\) , that is,
$${\int }_{0}^{T}[P({\alpha }^\varepsilon (t) = {s}_{ ij}) - {\nu }_{j}^{i}(t){{\vartheta}}^{i}(t)]f(t)dt \rightarrow 0,$$(5.112)for all \(f(\cdot ) \in {L}^{2}([0,T]; \mathbb{R})\) , where
$$({{\vartheta}}^{1}(t),\ldots,{{\vartheta}}^{l}(t)) = {p}_{ 0}\widetilde{\mathrm{1}\mathrm{l}} +{ \int }_{0}^{t}({{\vartheta}}^{1}(s),\ldots,{{\vartheta}}^{l}(s))\overline{Q}(s)ds.$$ -
(b)
For each i, j, i 1, j 1,\(P({\alpha }^\varepsilon (t) = {s}_{ij}\vert {\alpha }^\varepsilon (s) = {s}_{{i}_{1}{j}_{1}})\) converges in the weak sense to \({\nu }_{j}^{i}(t){{\vartheta}}_{ii}(t,s)\) on \({L}^{2}([s,T]; \mathbb{R})\) , that is,
$${\int }_{s}^{T}[P({\alpha }^\varepsilon (t) = {s}_{ ij}\vert {\alpha }^\varepsilon (s) = {s}_{{ i}_{1}{j}_{1}}) - {\nu }_{j}^{i}(t){{\vartheta}}_{ ii}(t,s)]f(t)dt \rightarrow 0,$$(5.113)for all \(f(\cdot ) \in {L}^{2}([s,T]; \mathbb{R})\) , where \({{\vartheta}}_{ij}(t,s)\) is defined in Lemma 5.24 (see (5.50)).
Proof: We only derive (5.112); the proof of (5.113) is similar. Let
where \({p}_{ij}^\varepsilon (t) = P({\alpha }^\varepsilon (t) = {s}_{ij})\) . Since \({p}^\varepsilon (\cdot ) \in {L}^{2}([0,T]; {\mathbb{R}}^{m})\) , there exists (see Lemma A.36) a subsequence of ε → 0 (still denoted by ε for simplicity), such that corresponding to this ε, p ε(t) converges to some \(p(\cdot ) \in {L}^{2}([0,T]; {\mathbb{R}}^{m})\) under the weak topology. Let
Then 0 ≤ p ij (t) ≤ 1 and ∑ i, j p ij (t) = 1 almost everywhere. Similarly as in the proof of Lemma 5.49, for 0 ≤ t ≤ T,
The irreducibility of \(\widetilde{{Q}}^{k}(t)\),k = 1, …, l, implies that
for some functions \({{\vartheta}}^{1}(t),\ldots,{{\vartheta}}^{l}(t)\).
In view of (5.106), we have
Since \(\widetilde{Q}(s)\widetilde{\mathrm{1}\mathrm{l}} = 0\), it follows that
Owing to the convergence of p ε(t) → p( t) under the weak topology of \({L}^{2}([0,T]; {\mathbb{R}}^{m})\), we have
Using (5.114) and noting that
we have
The uniqueness of the solution then yields the lemma. □
Theorem 5.52
Assume (A5.8) . Then for any i = 1,…,l, j = 1,…,m i , and bounded deterministic function β ij (t), t ≥ 0,
Proof: Let η(t) be defined as in (5.52). Then we can show similarly as in the proof of Theorem 5.25 that
where \({\Phi }^\varepsilon (t,r) = {\Phi }_{1}^\varepsilon (t,r) + {\Phi }_{2}^\varepsilon (t,r)\) with Φ 1 ε(t, r) and Φ 2 ε(t, r) defined by (5.53) and (5.54), respectively.
Note that by changing the order of integration,
Since the β ij ( ⋅) are bounded uniformly on [0, T], \({\beta }_{ij}(\cdot ) \in {L}^{2}([0,T]; \mathbb{R})\). As a result, Lemma 5.51 implies that
Hence as ε → 0,
Similarly,
The proof is complete. □
Theorem 5.53
Assume (A5.8) . Then \({\overline{\alpha }}^\varepsilon (\cdot )\) converges weakly to \(\overline{\alpha }(\cdot )\) on \(D([0,T];\overline{\mathcal{M}})\), as ε → 0.
Proof: Recall that χε(t) denotes the vector of indicator functions
and let
Then \({\overline{\chi }}_{i}^\varepsilon (t) = {I}_{\{{\overline{\alpha }}^\varepsilon (t)=i\}}\) for i = 1, …, l.
We show that \({\overline{\chi }}^\varepsilon (\cdot )\) is tight in D l[0, T] first. Let \({\mathcal{F}}_{t}^\varepsilon = \sigma \{{\alpha }^\varepsilon (r) :\; r \leq t\}\). Then in view of the martingale property associated with αε( ⋅), we have, for 0 ≤ s ≤ t,
Right multiplying both sides of the equation by \(\widetilde{\mathrm{1}\mathrm{l}}\) and noting that \(\widetilde{Q}(r)\widetilde{\mathrm{1}\mathrm{l}} = 0\), we obtain
Note that
It follows from (5.115) that
Note also that (I A )2 = I A for any set A. We have, in view of (5.116),
for each i = 1, …, l. Hence,
Therefore, by Lemma A.17, \({\overline{\chi }}^\varepsilon (\cdot )\) is tight.
The tightness of \({\overline{\chi }}^\varepsilon (\cdot )\) implies that for any sequence ε k → 0, there exists a subsequence of {ε k } (still denoted by {ε k }) such that \({\overline{\chi }}^{\varepsilon _{k}}(\cdot )\) converges weakly. We next show that the limit of such a subsequence is uniquely determined by \(\overline{Q}(\cdot ) := \mathrm{diag}({\nu }^{1}(\cdot ),\ldots,{\nu }^{l}(\cdot ))\widehat{Q}(\cdot )\widetilde{\mathrm{1}\mathrm{l}}\).
Note that
In view of Theorem 5.52, we have, as ε → 0,
Now by virtue of (5.115),
for 0 ≤ t 1 ≤ ⋯ ≤ t j ≤ s ≤ t and bounded and continuous functions z 1 ( ⋅), …, z j ( ⋅).
Let \(\overline{\chi }(\cdot )\) denote the limit in distribution of \({\overline{\chi }}^{\varepsilon _{k}}(\cdot )\). Then in view of (5.117) and the continuity of \({\int }_{s}^{t}\eta (r)\overline{Q}(r)dr\) with respect to η( ⋅) (see Lemma A.40), we have \({\overline{\chi }}^\varepsilon (\cdot ) \rightarrow \overline{\chi }(\cdot )\) as ε k → 0, and \(\overline{\chi }(\cdot )\) satisfies
It is easy to see that \(\overline{\chi }(\cdot ) = ({\overline{\chi }}_{1}(\cdot ),\ldots,{\overline{\chi }}_{l}(\cdot ))\) is an l-valued measurable process having sample paths in D l[0, T] and satisfying \({\overline{\chi }}_{i}(t) = 0\) or 1 and \({\overline{\chi }}_{1}(\cdot ) + \cdots +{ \overline{\chi }}_{l}(\cdot ) = 1\) w.p.1. Let
or in an expanded form,
Then \(\overline{\alpha }(\cdot )\) is a process with sample paths in \(D([0,T];\overline{\mathcal{M}})\) and
Therefore, \(\overline{\alpha }(\cdot )\) is a Markov chain generated by \(\overline{Q}(\cdot )\) . As a result, its distribution is uniquely determined by \(\overline{Q}(\cdot )\) . It follows that \({\overline{\alpha }}^\varepsilon (\cdot )\) converges weakly to \(\overline{\alpha }(\cdot )\). □
Remark 5.54
Note that Theorem 5.53 gives the same result as Theorem 5.27 under weaker conditions. The proofs are quite different. The proof of Theorem 5.53 is based on martingale properties associated with the Markov chain, whereas the proof of Theorem 5.27 follows the traditional approach, i.e., after the tightness is verified, the convergence of finite-dimensional distributions is proved.
Remark 5.55
In view of the development in Chapter 4, apart from the smoothness conditions, one of the main ingredients is the use of the Fredholm alternative. One hopes that this will carry over (under suitable conditions) to the measurable generators. A possible approach is the utilization of the formulation of weak derivatives initiated in the study of partial differential equations (see Hutson and Pym [90]).
Following the tactics of the weak sense formulation, for some T < ∞ and for given \(g(\cdot ) \in {L}^{2}([0,T]; \mathbb{R})\) , a function \(f(\cdot ) \in {L}^{2}([0,T]; \mathbb{R})\) is a weak solution of \((d/dt)f = g\) if
for any C ∞ -functions on [0,T] vanishing on the boundary together with their derivatives (denoted by \(\phi \in {C}_{0}^{\infty }([0,T]; \mathbb{R})\) ). Write the weak solution as \((d/dt)f\mbox{ w} =g\).
Recall that L loc 2 is the set of functions that lie in \({L}^{2}(S; \mathbb{R})\) for every closed and bounded set S ⊂ (0,T). A function f(⋅) ∈ L loc 2 has a jth-order weak derivative if there is a function g(⋅) ∈ L loc 2 such that
for all \(\phi \in {C}_{0}^{\infty }([0,T]; \mathbb{R})\) . The function g(⋅) above is called the jth-order weak derivative of f(⋅), and is denoted by D j f = g.
To proceed, define the space of functions H n as
Equip H n with an inner product and a norm as
One can then work under such a framework and proceed to obtain the asymptotic expansion of the probability distribution. It seems that the conditions required are not much different from those in the case of smooth generators; we will not pursue this issue further.
5 Remarks on Inclusion of Transient and Absorbing States
So far, the development in this chapter has focused on Markov chains with only recurrent states (either a single weakly irreducible class or a number of weakly irreducible classes). This section extends the results obtained to the case that a transient class or a group of absorbing states is included.
5.1 Inclusion of Transient States
Consider the Markov chain \({\alpha }^\varepsilon (\cdot ) \in \mathcal{M}\), where its generator is still given by (5.47) and the state space of αε(t) is given by
with \({\mathcal{M}}_{i} =\{ {s}_{i1},\ldots,{s}_{i{m}_{i}}\}\) and \({\mathcal{M}}_{{_\ast}} =\{ {s}_{{_\ast}1},\ldots,{s}_{{_\ast}{m}_{{_\ast}}}\}\). In what follows, we present results concerning the asymptotic distributions of scaled occupation measures and properties of measurable generators. While main assumptions and results are provided, the full proofs are omitted. The interested reader can derive the results using the ideas presented in the previous sections.
To proceed, assume that \(\widetilde{Q}(t)\) is a generator of a Markov chain satisfying
such that for each t ∈ [0, T] and each i = 1, …, l, \(\widetilde{{Q}}^{i}(t)\) is a generator with dimension m i ×m i , \(\widetilde{{Q}}_{{_\ast}}(t)\) is an m ∗ ×m ∗ matrix, \(\widetilde{{Q}}_{{_\ast}}^{i}(t) \in {\mathbb{R}}^{{m}_{{_\ast}}\times {m}_{i}}\), and \({m}_{1} + {m}_{2} + \cdots + {m}_{l} + {m}_{{_\ast}} = m\). We impose the following conditions.
-
(A5.9)
For all t ∈ [0, T], and i = 1, …, l, \(\widetilde{{Q}}^{i}(t)\) are weakly irreducible, and \(\widetilde{{Q}}_{{_\ast}}(t)\) is Hurwitz (i.e., all of its eigenvalues have negative real parts). Moreover, \(\widetilde{Q}(\cdot )\) is differentiable on [0, T] and its derivative is Lipschitz; \(\widehat{Q}(\cdot )\) is Lipschitz continuous on [0, T].
Use the partition
where
and write
where
and
In what follows, if \({a}_{{m}_{i}}(t)\) is time independent, we will simply write it as \({a}_{{m}_{i}}\). The requirement on \(\widetilde{{Q}}_{{_\ast}}(t)\) in (A5.9) implies that the corresponding states are transient. The Hurwitzian property also has the following interesting implication: For each t ∈ [0, T], and each i = 1, …, l, \({a}_{{m}_{i}}(t) = ({a}_{{m}_{i},1}(t),\ldots,{a}_{{m}_{i},{m}_{{_\ast}}}(t))^{\prime} \in {\mathbb{R}}^{{m}_{{_\ast}}\times 1}\). Then
for each j = 1, …, m ∗ . That is, for each t ∈ [0, T] and each j = 1, …, l, \(({a}_{{m}_{1},j}(t),\ldots,{a}_{{m}_{l},j}(t))\) can be considered a probability row vector. To see this, note that
which has nonnegative components. It follows from the definition that \({a}_{{m}_{i}}(t) \geq 0\). Furthermore,
Thus (5.122) follows. Similar to the development in the section for the case of weak and strong interactions, we can derive the following results.
Theorem 5.56
Define
and assume (A5.9) . Then for each j = 1,…,m i,
Next, for each fixed t ∈ [0, T], let ξ be a random variable uniformly distributed on [0, 1] that is independent of αε( ⋅). For each j = 1, …, m ∗ , define an integer-valued random variable ξ j (t) by
Now redefine the aggregated process \({\overline{\alpha }}^\varepsilon (\cdot )\) by
Note that the state space of \({\overline{\alpha }}^\varepsilon (t)\) is \(\overline{\mathcal{M}} =\{ 1,\ldots,l\}\), and that \({\overline{\alpha }}^\varepsilon (\cdot ) \in D([0,T];\overline{\mathcal{M}})\). Similar to the weak and strong interaction case, but with more effort, we can obtain the following result.
Theorem 5.57
Under conditions (A5.9), \({\overline{\alpha }}^\varepsilon (\cdot )\) converges weakly to \(\overline{\alpha }(\cdot )\) , a Markov chain generated by \({\overline{Q}}_{{_\ast}}(\cdot )\) given by (5.120).
Next, for t ≥ 0, and \(\alpha \in \mathcal{M}\), let β ij (t) be bounded Borel measurable deterministic functions, and let
Consider the normalized occupation measure
where
We can then proceed to obtain the asymptotic distribution.
Theorem 5.58
Assume (A5.9) , and suppose \(\widetilde{Q}(\cdot )\) is twice differentiable with Lipschitz continuous second derivative. Moreover, \(\widehat{Q}(\cdot )\) is differentiable with Lipschitz continuous derivative. Let β ij (⋅) (for i = 1,…,l, j = 1,…,m i ) be bounded and Lipschitz continuous deterministic functions. Then n ε (⋅) converges weakly to a switching diffusion n(⋅), where
where σ(s,i) is similar to (5.105) with the following modifications:
and w(⋅) is a standard m-dimensional Brownian motion.
Finally, we confirm that the case of the generators being merely measurable can be treated as well. We state this as the following theorem.
Theorem 5.59
Assume the generator is given by (5.47) with \(\widetilde{Q}(\cdot )\) given by (5.120) such that \(\widetilde{Q}\) and \(\widehat{Q}\) are measurable and bounded and that \(\widetilde{{Q}}^{i}(t)\) is weakly irreducible for each \(i = 1,\ldots,l\) . Then the following assertions hold:
-
For any i = 1, …, l, j = 1, …, m i , and bounded deterministic function β ij (t), t ≥ 0,
$$E\left ({\int }_{0}^{T}\left ({I}_{\{{ \alpha }^\varepsilon (t)={s}_{ij}\}} - {\nu }_{j}^{i}(t){I}_{\{{\overline{\alpha }}^{ \varepsilon }(t)=i\}}\right ){\beta }_{ij}(t)dt\right )^{2} \rightarrow 0,\mbox{ as }\varepsilon \rightarrow 0.$$ -
\({\overline{\alpha }}^\varepsilon (\cdot )\) converges weakly to \(\overline{\alpha }(\cdot )\), a Markov chain generated by \({\overline{Q}}_{{_\ast}}(\cdot )\).
5.2 Inclusion of Absorbing States
Consider the Markov chain \({\alpha }^\varepsilon (\cdot ) \in \mathcal{M}\), where the generator of αε( ⋅) is still given by (5.47) with
where \({0}_{{m}_{a}\times {m}_{a}}\) is the m a ×m a zero matrix, the state space of αε(t) is given by
with \({\mathcal{M}}_{i} =\{ {s}_{i1},\ldots,{s}_{i{m}_{i}}\}\) and \({\mathcal{M}}_{a} =\{ {s}_{a1},\ldots,{s}_{a{m}_{a}}\}\), and \({m}_{1} + {m}_{2} + \cdots + {m}_{l} + {m}_{a} = m\). Assume the following conditions.
-
(A5.10)
For all t ∈ [0, T] and i = 1, …, l, \(\widetilde{{Q}}^{i}(t)\) is weakly irreducible. Furthermore, \(\widetilde{Q}(\cdot )\) is differentiable on [0, T] and its derivative is Lipschitz. Moreover, \(\widehat{Q}(\cdot )\) is Lipschitz continuous on [0, T].
Define
Assume that the conditions in (A5.10) are satisfied. Then we can prove the following:
-
(a)
As ε → 0,
$${p}^\varepsilon (t) = ({\vartheta}(t),{{\vartheta}}^{a}(t))\mathrm{diag}({\nu }^{1}(t),\ldots,{\nu }^{l}(t),{I}_{{ m}_{a}}) + O\left (\varepsilon +\exp (-{\kappa }_{0}t/\varepsilon )\right ),$$where
$$\begin{array}{rl} &{\vartheta}(t) = ({{\vartheta}}^{1}(t),\ldots,{{\vartheta}}^{l}(t))) \in {\mathbb{R}}^{1\times l}\ \mbox{ and } \\ &{{\vartheta}}^{a}(t) = ({{\vartheta}}_{ 1}^{a}(t),\ldots,{{\vartheta}}_{{ m}_{a}}^{a}(t)) \in {\mathbb{R}}^{1\times {m}_{a} },\end{array}$$satisfying
$$\begin{array}{rl} { d({\vartheta}(t),{{\vartheta}}^{a}(t)) \over dt} = ({\vartheta}(t),{{\vartheta}}^{a}(t))\overline{Q}(t),\ \ ({\vartheta}(0),{{\vartheta}}^{a}(0)) = {p}^\varepsilon (0)\widetilde{\mathrm{1}{\mathrm{l}}}_{ a} \end{array}$$where \(\overline{Q}(t)\) is given in (5.131) and p ε(0) = (p ε, 1(0), …, p ε, l(0), p ε, a(0)) with \({p}^{\varepsilon,i}(0) \in {\mathbb{R}}^{1\times {m}_{i}}\) and \({p}^{\varepsilon,a}(0) \in {\mathbb{R}}^{1\times {m}_{a}}\).
-
(b)
For the transition probability P ε(t, t 0), we have
$${P}^\varepsilon (t,{t}_{ 0}) = {P}^{0}(t,{t}_{ 0}) + O\left (\varepsilon +\exp (-{\kappa }_{0}(t - {t}_{0})/\varepsilon ))\right ),$$(5.132)for some κ0 > 0, where
$${P}^{0}(t,{t}_{ 0}) =\widetilde{ \mathrm{1}{\mathrm{l}}}_{a}\Theta (t,{t}_{0})\mathrm{diag}({\nu }^{1}(t),\ldots,{\nu }^{l}(t),{I}_{{ m}_{a}}),$$and
$${ d\Theta (t,{t}_{0}) \over dt} = \Theta (t,{t}_{0})\overline{Q}(t),\ \ \Theta ({t}_{0},{t}_{0}) = I.$$
To proceed, we aggregate the states in \({\mathcal{M}}_{i}\) for i = 1, …, l as one state leading to the definition of the following process:
For each j = 1, …, m i , we also define a sequence of centered occupation measures by
For t ≥ 0 and \(\alpha \in \mathcal{M}\), let
Consider the normalized occupation measure
where
Note that
where
We can obtain the following results.
Theorem 5.60
Assume (A5.10) . Then the following assertions hold.
-
(a)
For all i = 1, …, l and j = 1, …, m i , corresponding to the recurrent states, sup t ∈ [0, T] E | O ij ε(t) | 2 = O(ε).
-
(b)
\({\overline{\alpha }}^\varepsilon (\cdot )\) converges weakly to \(\overline{\alpha }(\cdot )\), a Markov chain generated by \(\overline{Q}(\cdot )\).
-
(c)
Define the generator \(\mathcal{L}\) by
$$\begin{array}{rl} \mathcal{L}{f}^{0}(x,\alpha )& ={ 1 \over 2} {\sum }_{{j}_{1},{j}_{2}=1}^{{m}_{\alpha } }{a}_{{j}_{1}{j}_{2}}(s,\alpha ){\partial }_{\alpha,{j}_{1}{j}_{2}}^{2}{f}^{0}(x,\alpha ) \\ &\ + \sum \limits_{j=1}^{{m}_{a} }{b}_{j}(s,\alpha ){\partial }_{a,j}{f}^{0}(x,\alpha ) + \overline{Q}(s){f}^{0}(x,\cdot )(\alpha )\end{array}$$Then the sequence \({Y }^\varepsilon (\cdot ) = ({n}^\varepsilon (\cdot ),{\overline{\alpha }}^\varepsilon (\cdot ))\) converges weakly to \(\overline{Y }(\cdot ) = (n(\cdot ),\overline{\alpha }(\cdot ))\) that is a solution of the martingale problem with operator \(\mathcal{L}\).
Next, assume that \(\widetilde{Q}(\cdot )\) and \(\widehat{Q}(\cdot )\) are bounded and measurable and \(\widetilde{{Q}}^{i}(t)\) for each \(i = 1,\ldots,l\) is weakly irreducible. Then
converges in the weak topology of \({L}^{2}([0,T]; {\mathbb{R}}^{m})\) (with \(m = \sum \limits_{i=1}^{l}{m}_{i} + {m}_{a})\) to
where p 0, a is the subvector in the initial data p 0 corresponding the absorbing state.
Note that in deriving the asymptotic distribution of the scaled occupation measures, we need to compute the asymptotic covariance of the limit process. That is, we need to evaluate the limit of
where
It can be shown that
as ε → 0, where for i = 1, …, l, \({\overline{W}}^{r}(t) ={ \overline{W}}^{r}(t,i) ={ \int }_{0}^{t}\widehat{{W}}^{r}(s,i)ds\) with
with σ(s, i) the m i ×m i matrix such that \(\sigma (s,i)\sigma ^{\prime}(s,i) = A(s,i)\mbox{ for }i = 1,\ldots,l,\) and
where δ jk = 1 if j = k, δ jk = 0 if j ≠ k. The detailed proof of Theorem 5.60 can be found in Yin, Zhang, and Badowski [241].
6 Remarks on a Stability Problem
So far, our study has been devoted to systems with two time scales in a finite interval. In many problems arising in networked control systems, stability is often a main concern. A related problem along this line is in Badowski and Yin [5].
It is interesting to note that intuitive ideas sometimes are not necessarily true for systems with switching, for example, if one put together two stable systems by using, for instance, Markovian switching. Our intuition may lead to the conclusion that the combined systems should also be stable. Nevertheless, this is, in fact, not true. Such an idea was illustrated in Wang, Khargonekar, and Beydoun [212] for deterministically switched systems; see also Chapter 1 of this book concerning this matter.
As a variation of the system in [212], we consider the following example. Suppose that αε( ⋅) is a continuous-time Markov chain with state space \(\mathcal{M} =\{ 1,2\}\) and generator \({Q}^\varepsilon = Q/\varepsilon \), where \(Q = \left (\begin{array}{cc} - 1& 1\\ \\ \\ 1 & -1 \end{array} \right )\). Consider a controlled system
with state feedback u(t) = K(αε(t))x(t). Then we obtain the equivalent representation
Suppose that
Note that both matrices are Hurwitz (i.e., their eigenvalues have negative real parts). A question of interest is this: Is system (5.139) stable? The key to understanding the system is to examine
where both G(1) and G(2) are stable matrices.
Since Q is irreducible, the stationary distribution associated with Q is given by \((1/2,1/2)\). As a result, as ε → 0, using our weak convergence result, x ε( ⋅) converges weakly to x( ⋅), which is a solution of the system
In addition, for any T < ∞, using the large deviations result obtained in He, Yin, and Zhang [84], we can show that for any δ > 0, there is a c 1 > 0 such that
where \({\rho }_{0,T}(x,y) {=\sup }_{0\leq t\leq T}\vert x(t) - y(t)\vert \).
Note that \(\overline{G}\) is an unstable matrix with eigenvalues − 210 and 10. Thus for (5.141), the critical point (0, 0)′ is a saddle point. But why should the stability of the averaged system dominate that of the original system? To see this, from a result of differential equations, there is a nonsingular matrix H such that \(H\overline{G}{H}^{-1} = \Lambda = \mathrm{diag}(-210,10)\). Clearly, the stability of (5.141) is equivalent to that of
where \(y = Hx = ({y}_{1},{y}_{2})^{\prime}\). The stability of (5.141) is equivalent to that of (5.143), which is completely decoupled and \({y}_{1}(t) =\exp (-210t){y}_{1}(0) \rightarrow 0\) and y 2(t) = exp(10t)y 2(0) → ∞. To see how the original system (5.140) behaves, we apply the same transformation to get
For the transformed system (5.143), by choosing \(V (y) = {y}_{2}^{2}/2\), we obtain \(\dot{V }(y(t)) = 10{y}_{2}^{2} > 0\) for all y 2 ≠ 0. Define \({L}^\varepsilon z(t) {=\lim }_{\delta \downarrow 0}{E}_{t}^\varepsilon [z(t + \delta ) - z(t)]/\delta \) for a real-valued function z(t) that is continuously differentiable, where E t ε denotes the conditioning on the \({\mathcal{F}}_{t}^\varepsilon = \sigma \{{\alpha }^\varepsilon (s) : s \leq t\}\). With \(V (y) = {y}_{2}^{2}/2\), we have
where \({V ^{\prime}}_{y}(y) = (0,{y}_{2}) \in {\mathbb{R}}^{1\times 2}\). Using perturbed Liapunov function techniques as done in Badowski and Yin [5], define a perturbation
It can be shown that V 2 ε(y, t) = O(ε)V (y). In addition,
Define
Evaluate L ε V ε(y ε(t), t). Upon cancelation, for sufficiently small ε, we can make
It then follows that
Taking expectation of the left- and right-hand sides above leads to
which in turn yields that
Similar to the previous development, choose \(V (y) = {y}_{1}^{2}/2\), define
and redefine
Using the upper bound O(ε)V (y ε(t)) ≤ (y 1 ε(t))2 this time and calculating L ε V ε(y ε(t), t), we obtain
which in turn yields that
This yields that (5.144) and hence (5.140) are unstable in probability (see Yin and Zhu [244, p. 220] for a definition). In fact, it can be seen that the trivial solution of the original system is also a saddle.
In the same spirit of the last example, consider a system given by
where Q is as in the last example. Then it can be shown that x ε( ⋅) converges weakly to x( ⋅) that is a solution of the following system
Neither G(1) nor G(2) is a stable matrix, but the system (5.146) is a stable one. The stability analysis is again carried out using perturbed Liapunov function methods. Here exactly the same kind of argument as in [5] can be applied. Using the techniques of perturbed Liapunov functions, we can show that the stability of the averaged system “implies” that of the original system.
These two examples illustrate that one can combine two stable systems using Markovian switching to produce an unstable limit system. Likewise, one can combine two unstable systems to produce a limit stable systems. More importantly, using our weak convergence result of this chapter and the large deviations results in He, Yin, and Zhang [84], combined with the perturbed Liapunov function argument, we can give the reason why such a thing can happen.
7 Notes
This chapter concerns sequences of functional occupation measures. It includes convergence of an unscaled sequence (in probability) and central-limit-type results for suitably scaled sequences. For a general introduction to central limit theorems, we refer to the book by Chow and Teicher [30] and the references therein. In the stationary case, that is, Q(t) = Q, a constant matrix, the central limit theorem may be obtained as in Friedlin and Wentzell [67]. Some results of central limit type for discrete Markov chains are in Dobrushin [50] (see also the work of Linnik on time-inhomogeneous Markov chains [147]). Work in the context of random evolution, with primary concern the central limit theorem involving a singularly perturbed Markov chain, is in Pinsky [176]; see also Kurtz [135, 137] for related discussions and the martingale problem formulation. Exponential bounds for Markov processes are quite useful in analyzing the behavior of the underlying stochastic processes. Some results in connection with diffusions can be found in Kallianpur [102]. Corollary 5.8 can be viewed as a large deviations result. For extensive treatment of large deviations, see Varadhan [207].
The central theme here is limit results of unscaled as well as scaled sequences of occupation measures, which include the law of large numbers for an unscaled sequence, exponential upper bounds, and asymptotic distribution of a suitably scaled sequence of occupation times. Results in Section 5.2 are based on the paper of Zhang and Yin [252]; however, a somewhat different approach to the central limit theorem was used in [252]. Some of the results in Section 5.3 are based on Zhang and Yin [253]. The result on exponential error bound in Section 5.3 is a natural extension for the irreducible generators. Such result holds uniformly in t ∈ [0, T] for fixed but otherwise arbitrary T > 0. The main motivation for treating T as a parameter stems from various control and optimization problems with discounted cost over the infinite horizon. In such a situation, the magnitude of the bound counts. Thus detailed information on the bounding constant is helpful for dealing with the near optimality of the underlying problem. Section 5.3 also presents a characterization of the limit process using martingale problem formulations. Much of the foundation of this useful approach is in the work of Stroock and Varadhan [203]. Using perturbed operators to study limit behavior may be traced back to Kurtz [135]. The general idea of perturbed test functions was used in Blankenship and Papanicolaou [16], and Papanicolaou, Stroock, and Varadhan [168]. It was further developed and extended by Kushner [139] for various stochastic systems, and singularly perturbed systems in Kushner [140]; see also Kushner and Yin [145] for related stochastic approximation problems, and Ethier and Kurtz [59] and Kurtz [137] for related work in stochastic processes. The results of this section have benefited from the discussion with Thomas Kurtz, who suggested treating the pair of processes (n ε( ⋅), αε( ⋅)) together, which led to the current version. Earlier treatment of a pair of processes may be found in the work of Kesten and Papanicolaou [110] for stochastic acceleration.
The results on asymptotic properties for the inclusion of transient states can be found in Yin, Zhang, and Badowski [239]; the results for the case of generators being measurable can be found in the work of Yin, Zhang, and Badowski [240]; the results on asymptotic properties of occupation measures with absorbing states can be found in Yin, Zhang, and Badowski [241].
References
M. Abbad, J.A. Filar, and T.R. Bielecki, Algorithms for singularly perturbed limiting average Markov control problems, IEEE Trans. Automat. Control AC-37 (1992), 1421–1425.
R. Akella and P.R. Kumar, Optimal control of production rate in a failure-prone manufacturing system, IEEE Trans. Automat. Control AC-31 (1986), 116–126.
W.J. Anderson, Continuous-Time Markov Chains: An Application-Oriented Approach, Springer-Verlag, New York, 1991.
E. Altman, K.E. Avrachenkov, and R. Nunez-Queija, Pertrubation analysis for denumerable Markov chains with applications to queueing models, Adv. in Appl. Probab., 36 (2004), 839–853.
G. Badowski and G. Yin, Stability of hybrid dynamic systems containing singularly perturbed random processes, IEEE Trans. Automat. Control, 47 (2002), 2021–2031.
G. Barone-Adesi and R. Whaley, Efficient analytic approximation of American option values, J. Finance, 42 (1987), 301–320.
A. Bensoussan, J.L. Lion, and G.C. Papanicolaou, Asymptotic Analysis of Periodic Structures, North-Holland, Amsterdam, 1978.
A. Bensoussan, Perturbation Methods in Optimal Control, J. Wiley, Chichester, 1988.
L.D. Berkovitz, Optimal Control Theory, Springer-Verlag, New York, 1974.
A.T. Bharucha-Reid, Elements of the Theory of Markov Processes and Their Applications, McGraw-Hill, New York, 1960.
T.R. Bielecki and J.A. Filar, Singularly perturbed Markov control problem: Limiting average cost, Ann. Oper. Res. 28 (1991), 153–168.
T.R. Bielecki and P.R. Kumar, Optimality of zero-inventory policies for unreliable manufacturing systems, Oper. Res. 36 (1988), 532–541.
P. Billingsley, Convergence of Probability Measures, J. Wiley, New York, 1968.
T. Björk, Finite dimensional optimal filters for a class of Ito processes with jumping parameters, Stochastics, 4 (1980), 167–183.
W.P. Blair and D.D. Sworder, Feedback control of a class of linear discrete systems with jump parameters and quadratic cost criteria, Int. J. Control, 21 (1986), 833–841.
G.B. Blankenship and G.C. Papanicolaou, Stability and control of stochastic systems with wide band noise, SIAM J. Appl. Math. 34 (1978), 437–476.
H.A.P. Blom and Y. Bar-Shalom, The interacting multiple model algorithm for systems with Markovian switching coefficients, IEEE Trans. Automat. Control, AC-33 (1988), 780–783.
N.N. Bogoliubov and Y.A. Mitropolskii, Asymptotic Methods in the Theory of Nonlinear Oscillator, Gordon and Breach, New York, 1961.
E.K. Boukas and A. Haurie, Manufacturing flow control and preventive maintenance: A stochastic control approach, IEEE Trans. Automat. Control AC-35 (1990), 1024–1031.
P. Brémaud, Point Processes and Queues, Springer-Verlag, New York, 1981.
P.E. Caines and H.-F. Chen, Optimal adaptive LQG control for systems with finite state process parameters, IEEE Trans. Automat. Control, AC-30 (1985), 185–189.
S.L. Campbell, Singular perturbation of autonomous linear systems, II, J. Differential Equations 29 (1978), 362–373.
S.L. Campbell and N.J. Rose, Singular perturbation of autonomous linear systems, SIAM J. Math. Anal. 10 (1979), 542–551.
M. Caramanis and G. Liberopoulos, Perturbation analysis for the design of flexible manufacturing system flow controllers, Oper. Res. 40 (1992), 1107–1125.
M.-F. Chen, From Markov Chains to Non-equilibrium Particle Systems, 2nd ed., World Scientific, Singapore, 2004.
S. Chen, X. Li, and X.Y. Zhou, Stochastic linear quadratic regulators with indefinite control weight costs, SIAM J. Control Optim. 36 (1998), 1685–1702.
C.L. Chiang, An Introduction to Stochastic Processes and Their Applications, Kreiger, Hungtington, 1980.
T.-S. Chiang and Y. Chow, A limit theorem for a class of inhomogeneous Markov processes, Ann. Probab. 17 (1989), 1483–1502.
P.L. Chow, J.L. Menaldi, and M. Robin, Additive control of stochastic linear systems with finite horizon, SIAM J. Control Optim. 23 (1985), 859–899.
Y.S. Chow and H. Teicher, Probability Theory, Springer-Verlag, New York, 1978.
K.L. Chung, Markov Chains with Stationary Transition Probabilities, 2nd Ed., Springer-Verlag, New York, 1967.
F. Clarke, Optimization and Non-smooth Analysis, Wiley Interscience, New York, 1983.
O.L.V. Costa and F. Dufour, Singular perturbation for the discounted continuous control of piecewise deterministic Markov processes, Appl. Math. Optim., 63 (2011), 357–384.
O.L.V. Costa and F. Dufour, Singularly perturbed discounted Markov control processes in a general state space, SIAM J. Control Optim., 50 (2012), 720–747.
P.J. Courtois, Decomposability: Queueing and Computer System Applications, Academic Press, New York, NY, 1977.
D.R. Cox and H.D. Miller, The Theory of Stochastic Processes, J. Wiley, New York, 1965.
M.G. Crandall, C. Evans, and P.L. Lions, Some properties of viscosity solutions of Hamilton-Jacobi equations, Trans. Amer. Math. Soc. 282 (1984), 487–501.
M.G. Crandall, H. Ishii, and P.L. Lions, User’s guide to viscosity solutions of second order partial differential equations, Bull. Amer. Math. Soc. 27 (1992), 1–67.
M.G. Crandall and P.L. Lions, Viscosity solutions of Hamilton-Jacobi equations, Trans. Amer. Math. Soc. 277 (1983), 1–42.
I. Daubechies, Ten Lectures on Wavelets, CBMS-NSF Regional Conf. Ser. Appl. Math., SIAM, Philadelphia, PA, 1992.
M.H.A. Davis, Markov Models and Optimization, Chapman & Hall, London, 1993.
M.V. Day, Boundary local time and small parameter exit problems with characteristic boundaries, SIAM J. Math. Anal. 20 (1989), 222–248.
F. Delebecque, A reduction process for perturbed Markov chains, SIAM J. Appl. Math., 48 (1983), 325–350.
F. Delebecque and J. Quadrat, Optimal control for Markov chains admitting strong and weak interactions, Automatica 17 (1981), 281–296.
F. Delebecque, J. Quadrat, and P. Kokotovic, A unified view of aggregation and coherency in networks and Markov chains, Internat. J. Control 40 (1984), 939–952.
C. Derman, Finite State Markovian Decision Processes, Academic Press, New York, 1970.
G.B. Di Masi and Yu.M. Kabanov, The strong convergence of two-scale stochastic systems and singular perturbations of filtering equations, J. Math. Systems, Estimation Control 3 (1993), 207–224.
G.B. Di Masi and Yu.M. Kabanov, A first order approximation for the convergence of distributions of the Cox processes with fast Markov switchings, Stochastics Stochastics Rep. 54 (1995), 211–219.
J.L. Doob, Stochastic Processes, Wiley Classic Library Edition, Wiley, New York, 1990.
R.L. Dobrushin, Central limit theorem for nonstationary Markov chains, Theory Probab. Appl. 1 (1956), 65–80, 329–383.
E.B. Dynkin, Markov Processes, Springer-Verlag, Berlin, 1965.
N. Dunford and J.T. Schwartz, Linear Operators, Interscience, New York, 1958.
E.B. Dynkin and A.A. Yushkevich, Controlled Markov Processes, Springer-Verlag, New York, 1979.
W. Eckhaus, Asymptotic Analysis of Singular Perturbations, North-Holland, Amsterdam, 1979.
R.J. Elliott, Stochastic Calculus and Applications, Springer-Verlag, New York, 1982.
R.J. Elliott, Smoothing for a finite state Markov process, in Lecture Notes in Control and Inform. Sci., 69, 199–206, Springer-Verlag, New York, 1985.
R.J. Elliott, L. Aggoun, and J. Moore, Hidden Markov Models: Estimation and Control, Springer-Verlag, New York, 1995.
A. Erdélyi, Asymptotic Expansions, Dover, New York, 1956.
S.N. Ethier and T.G. Kurtz, Markov Processes: Characterization and Convergence, J. Wiley, New York, 1986.
W. Feller, An Introduction to Probability Theory and Its Applications, J. Wiley, New York, Vol. I, 1957; Vol. II, 1966.
W.H. Fleming, Functions of Several Variables, Addison-Wesley, Reading, 1965.
W.H. Fleming, Generalized solution in optimal stochastic control, in Proc. URI Conf. on Control, 147–165, Kingston, RI, 1982.
W.H. Fleming and R.W. Rishel, Deterministic and Stochastic Optimal Control, Springer-Verlag, New York, 1975.
W.H. Fleming and H.M. Soner, Controlled Markov Processes and Viscosity Solutions, Springer-Verlag, New York, 1992.
W.H. Fleming, S.P. Sethi, and H.M. Soner, An optimal stochastic production planning problem with randomly fluctuating demand, SIAM J. Control Optim. 25 (1987), 1494–1502.
W.H. Fleming and Q. Zhang, Risk-sensitive production planning of a stochastic manufacturing system, SIAM J. Control Optim., 36 (1998), 1147–1170.
M.I. Friedlin and A.D. Wentzel, Random Perturbations of Dynamical Systems, Springer-Verlag, New York, 1984.
C.W. Gardiner, Handbook of Stochastic Methods for Physics, Chemistry, and the Natural Sciences, 2nd Ed., Springer-Verlag, Berlin, 1985.
V.G. Gaitsgori and A.A. Pervozvanskii, Aggregation of states in a Markov chain with weak interactions, Kybernetika 11 (1975), 91–98.
D. Geman and S. Geman, Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images, IEEE Trans. Pattern Anal. Machine Intelligence 6 (1984), 721–741.
S.B. Gershwin, Manufacturing Systems Engineering, Prentice-Hall, Englewood Cliffs, 1994.
M.K. Ghosh, A. Arapostathis, and S.I. Marcus, Ergodic control of switching diffusions, SIAM J. Control Optim., 35 (1997), 1952–1988.
I.I. Gihman and A.V. Skorohod, Introduction to the Theory of Random Processes, W.B. Saunders, Philadelphia, 1969.
P. Glasserman, Gradient Estimation via Perturbation Analysis, Kluwer, Boston, MA, 1991.
R. Goodman, Introduction to Stochastic Models, Benjamin/Cummings, Menlo Park, CA, 1988.
R.J. Griego and R. Hersh, Random evolutions, Markov chains, and systems of partial differential equations, Proc. Nat. Acad. Sci. U.S.A. 62 (1969), 305–308.
R.J. Griego and R. Hersh, Theory of random evolutions with applications to partial differential equations, Trans. Amer. Math. Soc. 156 (1971), 405–418.
X. Guo and O. Hernàndez-Lerma, Continuous-time Markov Decision Processes: Theory and Applications, Springer, Heidelberg, 2001.
J.K. Hale, Ordinary Differential Equations, R.E. Krieger Publishing Co., 2nd Ed., Malabar, 1980.
P. Hänggi, P. Talkner, and M. Borkovec, Reaction-rate theory: Fifty years after Kramers, Rev. Modern Phys. 62 (1990), 251–341.
J.M. Harrison and M.I. Reiman, Reflected Brownian motion on an orthant, Ann. Probab. 9 (1981), 302–308.
U.G. Haussmann and Q. Zhang, Stochastic adaptive control with small observation noise, Stochastics Stochastics Rep. 32 (1990), 109–144.
U.G. Haussmann and Q. Zhang, Discrete time stochastic adaptive control with small observation noise, Appl. Math. Optim. 25 (1992), 303–330.
Q. He, G. Yin, and Q. Zhang, Large Deviations for Two-Time-Scale Systems Driven by Nonhomogeneous Markov Chains and LQ Control Problems, SIAM J. Control Optim., 49, (2011), 1737–1765.
R. Hersh, Random evolutions: A survey of results and problems, Rocky Mountain J. Math. 4 (1974), 443–477.
F.S. Hillier and G.J. Lieberman, Introduction to Operations Research, McGraw-Hill, New York, 1989.
Y.C. Ho and X.R. Cao, Perturbation Analysis of Discrete Event Dynamic Systems, Kluwer, Boston, MA, 1991.
A. Hoyland and M. Rausand, System Reliability Theory: Models and Statistical Methods, J. Wiley, New York, 1994.
Z. Hou and Q. Guo, Homogeneous Denumerable Markov Processes, Springer-Verlag, Berlin, 1980.
V. Hutson and J.S. Pym, Applications of Functional Analysis and Operator Theory, Academic Press, London, 1980.
N. Ikeda and S. Watanabe, Stochastic Differential Equations and Diffusion Processes, North-Holland, Amsterdam, 1981.
A.M. Il’in, Matching of Asymptotic Expansions of Solutions of Boundary Value Problems, Trans. Math. Monographs, Vol. 102, Amer. Math. Soc., Providence, 1992.
A.M. Il’in and R.Z. Khasminskii, Asymptotic behavior of solutions of parabolic equations and ergodic properties of nonhomogeneous diffusion processes, Math. Sbornik. 60 (1963), 366–392.
A.M. Il’in, R.Z. Khasminskii, and G. Yin, Singularly perturbed switching diffusions: Rapid switchings and fast diffusions, J. Optim. Theory Appl. 102 (1999), 555–591.
M. Iosifescu, Finite Markov Processes and Their Applications, Wiley, Chichester, 1980.
H. Ishii, Uniqueness of unbounded viscosity solutions of Hamilton-Jacobi equations, Indiana Univ. Math. J. 33 (1984), 721–748.
Y. Ji and H.J. Chizeck, Controllability, stabilizability, and continuous-time Markovian jump linear quadratic control, IEEE Trans. Automatic Control, 35 (1990), 777–788.
Y. Ji and H.J. Chizeck, Jump linear quadratic Gaussian control in continuous time, IEEE Trans. Automat. Control AC-37 (1992), 1884–1892.
J. Jiang and S.P. Sethi, A state aggregation approach to manufacturing systems having machines states with weak and strong interactions, Oper. Res. 39 (1991), 970–978.
Yu. Kabanov and S. Pergamenshchikov, Two-scale Stochastic Systems: Asymptotic Analysis and Control, Springer, New York, NY, 2003.
I.Ia. Kac and N.N. Krasovskii, On the stability of systems with random parameters, J. Appl. Math. Mech., 24 (1960), 1225–1246.
G. Kallianpur, Stochastic Filtering Theory, Springer-Verlag, New York, 1980.
D. Kannan, An Introduction to Stochastic Processes, North-Holland, New York, 1980.
S. Karlin and J. McGregor, The classification of birth and death processes, Trans. Amer. Math. Soc. 85 (1957), 489–546.
S. Karlin and H.M. Taylor, A First Course in Stochastic Processes, 2nd Ed., Academic Press, New York, 1975.
S. Karlin and H.M. Taylor, A Second Course in Stochastic Processes, Academic Press, New York, 1981.
J. Keilson, Green’s Function Methods in Probability Theory, Griffin, London, 1965.
J. Kevorkian and J.D. Cole, Perturbation Methods in Applied Mathematics, Springer-Verlag, New York, 1981.
J. Kevorkian and J.D. Cole, Multiple Scale and Singular Perturbation Methods, Springer-Verlag, New York, 1996.
H. Kesten and G.C. Papanicolaou, A limit theorem for stochastic acceleration, Comm. Math. Phys. 78 (1980), 19–63.
R.Z. Khasminskii, On diffusion processes with a small parameter, Izv. Akad. Nauk U.S.S.R. Ser. Mat. 27 (1963), 1281–1300.
R.Z. Khasminskii, On stochastic processes defined by differential equations with a small parameter, Theory Probab. Appl. 11 (1966), 211–228.
R.Z. Khasminskii, On an averaging principle for Ito stochastic differential equations, Kybernetika 4 (1968), 260-279.
R.Z. Khasminskii, Stochastic Stability of Differential Equations, 2nd Ed., Springer, New York, 2012.
R.Z. Khasminskii and G. Yin, Asymptotic series for singularly perturbed Kolmogorov-Fokker-Planck equations, SIAM J. Appl. Math. 56 (1996), 1766–1793.
R.Z. Khasminskii and G. Yin, On transition densities of singularly perturbed diffusions with fast and slow components, SIAM J. Appl. Math. 56 (1996), 1794–1819.
R.Z. Khasminskii and G. Yin, On averaging principles: An asymptotic expansion approach, SIAM J. Math. Anal., 35 (2004), 1534–1560.
R.Z. Khasminskii and G. Yin, Limit behavior of two-time-scale diffusions revisited, J. Differential Eqs., 212 (2005) 85–113.
R.Z. Khasminskii, G. Yin, and Q. Zhang, Asymptotic expansions of singularly perturbed systems involving rapidly fluctuating Markov chains, SIAM J. Appl. Math. 56 (1996), 277–293.
R.Z. Khasminskii, G. Yin, and Q. Zhang, Constructing asymptotic series for probability distribution of Markov chains with weak and strong interactions, Quart. Appl. Math. LV (1997), 177–200.
J.G. Kimemia and S.B. Gershwin, An algorithm for the computer control production in flexible manufacturing systems, IIE Trans. 15 (1983), 353–362.
J.F.C. Kingman, Poisson Processes, Oxford Univ. Press, Oxford, 1993.
S. Kirkpatrick, C. Gebatt, and M. Vecchi, Optimization by simulated annealing, Science 220 (1983), 671–680.
C. Knessel, On finite capacity processor-shared queues, SIAM J. Appl. Math. 50 (1990), 264–287.
C. Knessel and J.A. Morrison, Heavy traffic analysis of a data handling system with multiple sources, SIAM J. Appl. Math. 51 (1991), 187–213.
P.V. Kokotovic, Application of singular perturbation techniques to control problems, SIAM Rev. 26 (1984), 501–550.
P.V. Kokotovic, A. Bensoussan, and G. Blankenship (Eds.), Singular Perturbations and Asymptotic Analysis in Control Systems, Lecture Notes in Control and Inform. Sci. 90, Springer-Verlag, Berlin, 1987.
P.V. Kokotovic and H.K. Khalil (Eds.), Singular Perturbations in Systems and Control, IEEE Press, New York, 1986.
P.V. Kokotovic, H.K. Khalil, and J. O’Reilly, Singular Perturbation Methods in Control, Academic Press, London, 1986.
V. Korolykuk and A. Swishchuk, Evolution of Systems in Random Media, CRC Press, Boca Raton, 1995.
V.S. Korolyuk and N. Limnios, Diffusion approximation with equilibrium of evolutionary systems switched by semi-Markov processes, translation in Ukrainian Math. J. 57 (2005), 1466–1476.
V.S. Korolyuk and N. Limnios, Stochastic systems in merging phase space, World Sci., Hackensack, NJ, 2005.
N.M. Krylov and N.N. Bogoliubov, Introduction to Nonlinear Mechanics, Princeton Univ. Press, Princeton, 1947.
H. Kunita and S. Watanabe, On square integrable martingales, Nagoya Math. J. 30 (1967), 209–245.
T.G. Kurtz, A limit theorem for perturbed operator semigroups with applications to random evolutions, J. Functional Anal. 12 (1973), 55–67.
T.G. Kurtz, Approximation of Population Processes, SIAM, Philadelphia, PA, 1981.
T.G. Kurtz, Averaging for martingale problems and stochastic approximation, in Proc. US-French Workshop on Appl. Stochastic Anal., Lecture Notes in Control and Inform. Sci., 177, I. Karatzas and D. Ocone (Eds.), 186–209, Springer-Verlag, New York, 1991.
H.J. Kushner, Probability Methods for Approximation in Stochastic Control and for Elliptic Equations, Academic Press, New York, 1977.
H.J. Kushner, Approximation and Weak Convergence Methods for Random Processes, with Applications to Stochastic Systems Theory, MIT Press, Cambridge, MA, 1984.
H.J. Kushner, Weak Convergence Methods and Singularly Perturbed Stochastic Control and Filtering Problems, Birkhäuser, Boston, 1990.
H.J. Kushner and P.G. Dupuis, Numerical Methods for Stochastic Control Problems in Continuous Time, Springer-Verlag, New York, 1992.
H.J. Kushner and W. Runggaldier, Nearly optimal state feedback controls for stochastic systems with wideband noise disturbances, SIAM J Control Optim. 25 (1987), 289–315.
H.J. Kushner and F.J. Vázquez-Abad, Stochastic approximation algorithms for systems over an infinite horizon, SIAM J. Control Optim. 34 (1996), 712–756.
H.J. Kushner and G. Yin, Asymptotic properties of distributed and communicating stochastic approximation algorithms, SIAM J. Control Optim. 25 (1987), 1266–1290.
H.J. Kushner and G. Yin, Stochastic Approximation and Recursive Algorithms and Applications, 2nd Edition, Springer-Verlag, New York, 2003.
X.R. Li, Hybrid estimation techniques, in Control and Dynamic Systems, Vol. 76, C.T. Leondes (Ed.), Academic Press, New York, 1996.
Yu. V. Linnik, On the theory of nonhomogeneous Markov chains, Izv. Akad. Nauk. USSR Ser. Mat. 13 (1949), 65–94.
Y.J. Liu, G. Yin, and X.Y. Zhou, Near-optimal controls of random-switching LQ problems with indefinite control weight costs, Automatica, 41 (2005) 1063–1070.
P. Lochak and C. Meunier, Multiphase Averaging for Classical Systems, Springer-Verlag, New York, 1988.
J. Lehoczky, S.P. Sethi, H.M. Soner, and M. Taksar, An asymptotic analysis of hierarchical control of manufacturing systems under uncertainty, Math. Oper. Res. 16 (1992), 596–608.
G. Lerman and Z. Schuss, Asymptotic theory of large deviations for Markov chains, SIAM J. Appl. Math., 58 (1998), 1862–1877.
D. Ludwig, Persistence of dynamical systems under random perturbations, SIAM Rev. 17 (1975), 605–640.
X. Mao and C. Yuan, Stochastic Differential Equations with Markovian Switching, Imperial College Press, London, UK, 2006.
M. Mariton, Robust jump linear quadratic control: A mode stabilizing solution, IEEE Trans. Automat. Control, AC-30 (1985), 1145–1147.
M. Mariton, Jump Linear Systems in Automatic Control, Marcel Dekker, Inc., New York, 1990.
L.F. Martins and H.J. Kushner, Routing and singular control for queueing networks in heavy traffic, SIAM J. Control Optim. 28 (1990), 1209–1233.
W.A. Massey and W. Whitt, Uniform acceleration expansions for Markov chains with time-varying rates, Ann. Appl. Probab., 8 (1998), 1130–1155.
B.J. Matkowsky and Z. Schuss, The exit problem for randomly perturbed dynamical systems, SIAM J. Appl. Math. 33 (1977), 365–382.
S.P. Meyn and R.L. Tweedie, Markov Chains and Stochastic Stability, Springer-Verlag, London, 1993.
T. Naeh, M.M. Klosek, B.J. Matkowski, and Z. Schuss, A direct approach to the exit problem, SIAM J. Appl. Math. 50 (1990), 595–627.
A.H. Nayfeh, Introduction to Perturbation Techniques, J. Wiley, New York, 1981.
M.F. Neuts, Matrix-Geometric Solutions in Stochastic Models, Johns Hopkins Univ. Press, Baltimore, 1981.
R.E. O’Malley, Jr., Singular Perturbation Methods for Ordinary Differential Equations, Springer-Verlag, New York, 1991.
Z.G. Pan and T. Başar, H ∞-control of Markovian jump linear systems and solutions to associated piecewise-deterministic differential games, in New Trends in Dynamic Games and Applications, G.J. Olsder Ed., 61–94, Birkhäuser, Boston, MA, 1995.
Z.G. Pan and T. Başar, H ∞ control of large scale jump linear systems via averaging and aggregation, in Proc. 34th IEEE Conf. Decision Control, 2574-2579, New Orleans, LA, 1995.
Z.G. Pan and T. Başar, Random evolutionary time-scale decomposition in robust control of jump linear systems, in Proc. 35th IEEE Conf. Decision Control, Kobe, Japan, 1996.
Z.G. Pan and T. Basar, H ∞ control of large-scale jump linear systems via averaging and aggregation, Internat. J. Control, 72 (1999), 866–881.
G.C. Papanicolaou, D. Stroock, and S.R.S. Varadhan, Martingale approach to some limit theorems, in Proc. 1976 Duke Univ. Conf. on Turbulence, Durham, NC, 1976.
G.C. Papanicolaou, Introduction to the asymptotic analysis of stochastic equations, in Lectures in Applied Mathematics, Amer. Math. Soc., Vol. 16, 1977, 109-147.
G.C. Papanicolaou, Asymptotic analysis of stochastic equations, Studies in Probability Theory, M. Rosenblatt (Ed.), Vol. 18, MAA, 1978, 111–179.
E. Pardoux and S. Peng, Adapted solution of backward stochastic equation, Syst. Control Lett., 14 (1990), 55–61.
A. Pazy, Semigroups of Linear Operators and Applications to Partial Differential Equations, Springer-Verlag, New York, 1983.
L. Perko, Differential Equations and Dynamical Systems, Springer, 3rd Ed., New York, 2001.
A.A. Pervozvanskii and V.G. Gaitsgori, Theory of Suboptimal Decisions: Decomposition and Aggregation, Kluwer, Dordrecht, 1988.
R.G. Phillips and P.V. Kokotovic, A singular perturbation approach to modelling and control of Markov chains, IEEE Trans. Automat. Control 26 (1981), 1087–1094.
M.A. Pinsky, Differential equations with a small parameter and the central limit theorem for functions defined on a Markov chain, Z. Wahrsch. verw. Gebiete 9 (1968), 101–111.
M.A. Pinsky, Multiplicative operator functionals and their asymptotic properties, in Advances in Probability Vol. 3, P. Ney and S. Port (Eds.), Marcel Dekker, New York, 1974.
L. Prandtl, Über Flüssigkeits – bewegung bei kleiner Reibung, Verhandlungen, in III. Internat. Math. Kongresses, (1905), 484–491.
M.L. Puterman, Markov Decision Processes: Discrete Stochastic Dynamic Programming, J. Wiley, New York, 1994.
D. Revuz, Markov Chains, Revised Ed., North-Holland, Amsterdam, 1975.
R. Rishel, Controlled wear process: Modeling optimal control, IEEE Trans. Automat. Control 36 (1991), 1100–1102.
H. Risken, The Fokker-Planck Equation: Methods of Solution and Applications, 2nd Ed., Springer-Verlag, London, 1989.
M. Rosenblatt, Markov Processes: Structure and Asymptotic Behavior, Springer-Verlag, Berlin, 1971.
S. Ross, Introduction to Stochastic Dynamic Programming, Academic Press, New York, 1983.
E. Roxin, The existence of optimal controls, Mich. Math. J. 9 (1962), 109–119.
V.R. Saksena, J. O’Reilly, and P.V. Kokotovic, Singular perturbations and time-scale methods in control theory: Survey 1976-1983, Automatica 20 (1984), 273–293.
Z. Schuss, Singularly perturbation methods in stochastic differential equations of mathematical physics, SIAM Rev. 22 (1980), 119–155.
Z. Schuss, Theory and Applications of Stochastic Differential Equations, J. Wiley, New York, 1980.
E. Seneta, Non-negative Matrices and Markov Chains, Springer-Verlag, New York, 1981.
R. Serfozo, Introduction to Stochastic Networks, Springer, New York, 1999.
S.P. Sethi and G.L. Thompson, Applied Optimal Control: Applications to Management Science, Martinus Nijhoff, Boston, MA, 1981.
S.P. Sethi and Q. Zhang, Hierarchical Decision Making in Stochastic Manufacturing Systems, Birkhäuser, Boston, 1994.
S.P. Sethi and Q. Zhang, Multilevel hierarchical decision making in stochastic marketing-production systems, SIAM J. Control Optim. 33 (1995), 528–553.
O.P. Sharma, Markov Queues, Ellis Horwood, New York, 1990.
H.A. Simon, Models of Discovery and Other Topics in the Methods of Science, D. Reidel Publ. Co., Boston, MA, 1977.
H.A. Simon and A. Ando, Aggregation of variables in dynamic systems, Econometrica 29 (1961), 111–138.
A.V. Skorohod, Studies in the Theory of Random Processes, Dover, New York, 1982.
A.V. Skorohod, Asymptotic Methods of the Theory of Stochastic Differential Equations, Trans. Math. Monographs, Vol. 78, Amer. Math. Soc., Providence, 1989.
D.R. Smith, Singular Perturbation Theory, Cambridge Univ. Press, New York, 1985.
D. Snyder, Random Point Processes, Wiley, New York, 1975.
H.M. Soner, Optimal control with state space constraints II, SIAM J. Control Optim. 24 (1986), 1110–1122.
H.M. Soner, Singular perturbations in manufacturing systems, SIAM J. Control Optim. 31 (1993), 132–146.
D.W. Stroock and S.R.S. Varadhan, Multidimensional Diffusion Processes, Springer-Verlag, Berlin, 1979.
H.M. Taylor and S. Karlin, An Introduction to Stochastic Modeling, Academic Press, Boston, 1994.
W.A. Thompson, Jr., Point Process Models with Applications to Safety and Reliability, Chapman and Hall, New York, 1988.
D.N.C. Tse, R.G. Gallager, and J.N. Tsitsiklis, Statistical multiplexing of multiple time-scale Markov streams, IEEE J. Selected Areas Comm. 13 (1995), 1028–1038.
S.R.S. Varadhan, Large Deviations and Applications, SIAM, Philadelphia, 1984.
N.G. van Kampen, Stochastic Processes in Physics and Chemistry, North-Holland, Amsterdam, 1992.
A.B. Vasil’eava and V.F. Butuzov, Asymptotic Expansions of the Solutions of Singularly Perturbed Equations, Nauka, Moscow, 1973.
A.B. Vasil’eava and V.F. Butuzov, Asymptotic Methods in Singular Perturbations Theory (in Russian), Vysshaya Shkola, Moscow, 1990.
D. Vermes, Optimal control of piecewise deterministic Markov processes, Stochastics, 14 (1985), 165–207.
L.Y. Wang, P.P. Khargonekar, and A. Beydoun, Robust control of hybrid systems: Performance guided strategies, in Hybrid Systems V, P. Antsaklis, W. Kohn, M. Lemmon, A. Nerode, amd S. Sastry Eds., Lecuture Notes in Computer Sci., 1567, 356–389, Berlin, 1999.
Z. Wang and X. Yang, Birth and Death Processes and Markov Chains, Springer-Verlag, Science Press, Beijing, 1992.
J. Warga, Relaxed variational problems, J Math. Anal. Appl. 4 (1962), 111–128.
W. Wasow, Asymptotic Expansions for Ordinary Differential Equations, Interscience, New York, 1965.
W. Wasow, Linear Turning Point Theory, Springer-Verlag, New York, 1985.
A.D. Wentzel, On the asymptotics of eigenvalues of matrices with elements of order exp( − V ij ∕ 2ε2), Dokl. Akad. Nauk SSSR 222 (1972), 263–265.
D.J. White, Markov Decision Processes, Wiley, New York, 1992.
J.H. Wilkinson, The Algebraic Eigenvalue Problem, Oxford University Press, New York, 1988.
H. Yan, G. Yin, and S. X. C. Lou, Using stochastic optimization to determine threshold values for control of unreliable manufacturing systems, J. Optim. Theory Appl. 83 (1994), 511–539.
H. Yan and Q. Zhang, A numerical method in optimal production and setup scheduling in stochastic manufacturing systems, IEEE Trans. Automat. Control, 42 (1997), 1452–1455.
G. Yin, Asymptotic properties of an adaptive beam former algorithm, IEEE Trans. Information Theory IT-35 (1989), 859-867.
G. Yin, Asymptotic expansions of option price under regime-switching diffusions with a fast-varying switching process, Asymptotic Anal., 65 (2009), 203–222.
G. Yin and I. Gupta, On a continuous time stochastic approximation problem, Acta Appl. Math. 33 (1993), 3–20.
G. Yin, V. Krishnamurthy, and C. Ion, Regime switching stochastic approximation algorithms with application to adaptive discrete stochastic optimization, SIAM J. Optim., 14 (2004), 1187–1215.
G. Yin and D.T. Nguyen, Asymptotic expansions of backward equations for two-time-scale Markov chains in continuous time, Acta Math. Appl. Sinica, 25 (2009), 457–476.
G. Yin and K.M. Ramachandran, A differential delay equation with wideband noise perturbation, Stochastic Process Appl. 35 (1990), 231–249.
G. Yin, H. Yan, and S.X.C. Lou, On a class of stochastic optimization algorithms with applications to manufacturing models, in Model-Oriented Data Analysis, W.G. Müller, H.P. Wynn and A.A. Zhigljavsky (Eds.), 213–226, Physica-Verlag, Heidelberg, 1993.
G. Yin and H.L. Yang, Two-time-scale jump-diffusion models with Markovian switching regimes, Stochastics Stochastics Rep., 76 (2004), 77–99.
G. Yin and H. Zhang, Two-time-scale markov chains and applications to quasi-birth-death queues, SIAM J. Appl. Math., 65 (2005), 567–586.
G. Yin and H. Zhang, Singularly perturbed markov chains: Limit results and applications, Ann. Appl. Probab., 17 (2007), 207–229.
G. Yin, H. Zhang, and Q. Zhang, Applications of Two-time-scale Markovian Systems, Preprint, 2012.
G. Yin and Q. Zhang, Near optimality of stochastic control in systems with unknown parameter processes, Appl. Math. Optim. 29 (1994), 263–284.
G. Yin and Q. Zhang, Control of dynamic systems under the influence of singularly perturbed Markov chains, J. Math. Anal. Appl., 216 (1997), 343–367.
G. Yin and Q. Zhang (Eds.), Recent Advances in Control and Optimization of Manufacturing Systems, Lecture Notes in Control and Information Sciences (LNCIS) series, Vol. 214, Springer-Verlag, New York, 1996.
G. Yin and Q. Zhang (Eds.), Mathematics of Stochastic Manufacturing Systems, Proc. 1996 AMS-SIAM Summer Seminar in Applied Mathematics, Lectures in Applied Mathematics, Amer. Math. Soc., Providence, RI, 1997.
G. Yin and Q. Zhang, Continuous-Time Markov Chains and Applications: A Singular Perturbation Approach, 1st Ed., Springer-Verlag, New York, 1998.
G. Yin and Q. Zhang, Discrete-time Markov Chains: Two-time-scale Methods and Applications, Springer, New York. 2005.
G. Yin, Q. Zhang, and G. Badowski, Asymptotic properties of a singularly perturbed Markov chain with inclusion of transient states, Ann. Appl. Probab., 10 (2000), 549–572.
G. Yin, Q. Zhang, and G. Badowski, Singularly perturbed Markov chains: Convergence and aggregation, J. Multivariate Anal., 72 (2000), 208–229.
G. Yin, Q. Zhang, and G. Badowski, Occupation measures of singularly perturbed Markov chains with absorbing states, Acta Math. Sinica, 16 (2000), 161–180.
G. Yin, Q. Zhang, and G. Badowski, Discrete-time singularly perturbed Markov chains: Aggregation, occupation measures, and switching diffusion limit, Adv. in Appl. Probab., 35 (2003), 449–476.
G. Yin and X.Y. Zhou, Markowitz’s mean-variance portfolio selection with regime switching: From discrete-time models to their continuous-time limits, IEEE Trans. Automat. Control, 49 (2004), 349–360.
G. Yin and C. Zhu, Hybrid Switching Diffusions: Properties and Applications, Springer, New York, 2010.
K. Yosida, Functional Analysis, 6th Ed., Springer-Verlag, New York, NY, 1980.
J. Yong and X.Y. Zhou, Stochastic Controls: Hamiltonian Systems and HJB Equations, Springer, New York, 1999.
Q. Zhang, An asymptotic analysis of controlled diffusions with rapidly oscillating parameters, Stochastics Stochastics Rep. 42 (1993), 67–92.
Q. Zhang, Risk sensitive production planning of stochastic manufacturing systems: A singular perturbation approach, SIAM J. Control Optim. 33 (1995), 498–527.
Q. Zhang, Finite state Markovian decision processes with weak and strong interactions, Stochastics Stochastics Rep. 59 (1996), 283–304.
Q. Zhang, Nonlinear filtering and control of a switching diffusion with small observation noise, SIAM J. Control Optim., 36 (1998), 1738–1768.
Q. Zhang and G. Yin, Turnpike sets in stochastic manufacturing systems with finite time horizon, Stochastics Stochastics Rep. 51 (1994), 11–40.
Q. Zhang and G. Yin, Central limit theorems for singular perturbations of nonstationary finite state Markov chains, Ann. Appl. Probab. 6 (1996), 650–670.
Q. Zhang and G. Yin, Structural properties of Markov chains with weak and strong interactions, Stochastic Process Appl., 70 (1997), 181–197.
Q. Zhang and G. Yin, On nearly optimal controls of hybrid LQG problems, IEEE Trans. Automat. Control, 44 (1999), 2271–2282.
Q. Zhang and G. Yin, Nearly optimal asset allocation in hybrid stock-investment models, J. Optim. Theory Appl., 121 (2004), 197–222.
Q. Zhang, G. Yin, and E.K. Boukas, Controlled Markov chains with weak and strong interactions: Asymptotic optimality and application in manufacturing, J. Optim. Theory Appl., 94 (1997), 169–194.
Q. Zhang, G. Yin, and R.H. Liu, A near-optimal selling rule for a two-time-scale market model, SIAM J. Multiscale Modeling Simulation, 4 (2005), 172–193.
X.Y. Zhou, Verification theorem within the framework of viscosity solutions, J. Math. Anal. Appl. 177 (1993), 208–225.
X.Y. Zhou and G. Yin, Markowitz mean-variance portfolio selection with regime switching: A continuous-time model, SIAM J. Control Optim., 42 (2003), 1466–1482.
C. Zhu, G. Yin, and Q.S. Song, Stability of random-switching systems of differential equations, Quart. Appl. Math., 67 (2009), 201–220.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer Science+Business Media, LLC
About this chapter
Cite this chapter
Yin, G.G., Zhang, Q. (2013). Occupation Measures: Asymptotic Properties and Ramification. In: Continuous-Time Markov Chains and Applications. Stochastic Modelling and Applied Probability, vol 37. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-4346-9_5
Download citation
DOI: https://doi.org/10.1007/978-1-4614-4346-9_5
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4614-4345-2
Online ISBN: 978-1-4614-4346-9
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)