Abstract
The aim of this paper is to study a stochastic partially observed optimal control problem, for systems of forward backward stochastic differential equations (FBSDE for short), which are driven by both a family of Teugels martingales and an independent Brownian motion. By using Girsavov’s theorem and a standard spike variational technique, we prove necessary conditions to characterize an optimal control under a partial observation, where the control domain is supposed to be convex. Moreover, under some additional convexity conditions, we prove that these partially observed necessary conditions are sufficient. In fact, compared to the existing methods, we get the last achievement in two different cases according to the linearity or the nonlinearity of the terminal condition for the backward component. As an illustration of the general theory, an application to linear quadratic control problems is also investigated.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
In this paper, we are interested in partially observed optimal control of systems driven by a forward backward stochastic differential equation of the type
where W t is a brownian motion and \((H_{t}^{\left (i\right ) } )_{i=1}^{\infty }\) is a family of pairwise orthogonal martingales associated with some Lévy process L t , which is independent from W t . These martingales are called Teugels martingales. The control problem consists in minimizing the following cost functional
over a partially observed class of admissible controls to be specified later.
Since 1990, more precisely since the work of Pardoux and Peng [14], the theory of backward stochastic differential equations (BSDEs) and FBSDEs has found important applications and has become a powerful tool in many fields. For instance, the financial mathematics, optimal control, stochastic games, partial differential equations and homogenization, see e.g. [1, 7,8,9].
In the literature, optimal control problems are handled in two different approaches. One is the Bellman dynamic programming principle, and the second is the maximum principle. Our purpose in this framework is precisely to deal with the second approach in the case where the full range of information available to the controller is assumed to be partially observed. It is well-known that a partially observed stochastic optimal control of BSDEs and FBSDEs driven only by a brownian motion has been studied by many authors through several articles, such as [5, 15, 16, 18, 20] and the references therein. The case of FBSDEs driven by both a brownian motion and a Lévy process has been considered in [21], by using certain classical convex variational techniques.
A worthy and powerful motivation for studying SDEs and BSDEs driven by a brownian motion and Teugels martingales is due to the very useful representation theorem provided by Nualart and Schoutens [12]. This theorem asserts that every square integrable martingale adapted to the natural filtration of a brownian motion and an independent Lévy process, can be written as the sum of a stochastic integral with respect to the brownian motion and the sum of stochastic integrals with respect to the Teugels martingales associated to the Lévy process. In other words, this representation formula put brownian motion and Lévy processes in a unified theory of square integrable martingales. See the excellent accounts by Davis [6], Schoutens [19]. In another paper [13], the authors have proved an existence and uniqueness result for BSDEs driven by Teugels martingales, under Lipschitz conditions. Moreover, an application to Clark–Ocone and Feynman–Kac formulas for Lévy processes is presented. Their result has been extended to the locally Lipschitz property in [3].
It then becomes quite natural to investigate control problems for systems driven by this kind of equations. Let us point out that the first work in this direction has been carried out by Mitsui-Tabata [11], for the case of a linear quadratic problem. Then Meng-Tang [10] studied the stochastic maximum principle for systems driven by an Itô forward SDE, by using convex perturbations technique. Optimal control of BSDEs driven by Teugels martingales has been addressed in Tang and Zhang [18], where necessary and sufficient conditions have been established. It is worth noting that in all previous control problems the information of the control problem is assumed to be completely observed. In return, this is not always reasonable in the real world applications because the controllers can only get partial information at most cases. This gives us a motivation to study this kind of control problems.
However, up to now, there is only one literature (see Bahlali et al. [5]) dealing with a partial information control problem for a system governed by SDEs driven by a both Teugels martingales and an independent brownian motion. In this work, the control variable is allowed to enter into the both coefficient and is assumed to be adapted to subfiltration which is possibly less than the whole one. The authors investigated a partial information necessary as well as sufficient conditions by using certain classical convex variational techniques.
The main contribution of our present paper is to investigate a partially observed necessary as well as sufficient conditions satisfied by an optimal control. To obtain the optimality necessary conditions, we use the convex perturbation method and differentiate the perturbed both the state equations and the cost functional, in order to get the adjoint process, which is a solution of a backward forward SDE, driven by both a brownian motion and a family of Teugels martingales, on top of the variational inequality between the Hamiltonians. Moreover, an additional technical assumptions are required to prove that these partially observed necessary conditions are in fact sufficient.
The rest of the paper is structured as follows. A brief introduction to Teugels martingales and a precise formulation of the control problem are presented in Section 2. Section 3 consists of the proof of partially observed necessary conditions of optimality in term of classical convex variational techniques. Under some additional convexity conditions, we show that these partially observed necessary conditions of optimality are also sufficient in Section 4. Finally in Section 5, we illustrate the general results by solving an example.
2 Preliminaries and Problem Formulation
2.1 Preliminaries and Assumptions
Let \(\left ({\Omega },\mathcal {F},P\right ) \) be a complete filtered probability space equipped with two mutually independent standard brownian motions W and Y valued in \(\mathbb {R}^{d}\) and \(\mathbb {R}^{r}\), respectively and an independent \(\mathbb {R}^{m}\)-valued Lévy process {L t , t ∈ [0, T]} of the form L t = b t + l t , where l t is a pure jump process. Assume further that the Lévy measure ν(d z) corresponding to the Lévy process L t satisfies
-
i)
\({\int }_{\mathbb {R}}\left (1\wedge z^{2}\right ) \nu \left (dz\right ) <\infty \),
-
ii)
\({\int }_{\left (\left ] -\varepsilon ,\varepsilon \right [ \right )^{c} }e^{\alpha \left \vert z\right \vert }\nu \left (dz\right ) <\infty \), for every ε > 0 and some α > 0.
The two above settings imply that the random variable L t have moments in all orders. We also assume that
where \(\mathcal {N}\) denotes the totality of the P-null set and \(\mathcal {F}_{t}^{W}\), \(\mathcal {F}_{t}^{Y}\) and \(\mathcal {F}_{t}^{L}\) denotes the P-completed natural filtration generated by W, Y and L respectively.
Let us recall briefly the L 2 theory of Lévy processes as it is investigated in Nualart-Schoutens [12] and Schoutens [19]. A convenient basis for martingale representation is provided by the so-called Teugels martingales. This means that this family has the predictable representation property.
Denote by ΔL t = L t − L t− and define the power jump processes \(L_{t}^{\left (1\right ) }=L_{t}\) and \(L_{t}^{\left (i\right ) } =\sum \limits _{0<s\leq t}\left (\bigtriangleup L_{s}\right )^{(i)}\) for i ≥ 2. If we denote \(\mathbf {Y}_{t}^{\left (i\right ) }=L_{t}^{\left (i\right ) }-\mathbb {E}\left [ L_{t}^{\left (i\right ) }\right ] ,\) i ≥ 1, then the family of Teugels martingales \(\left (H_{t}^{\left (i\right ) }\right )_{i=1}^{\infty }\), is defined by \(H_{t}^{\left (i\right ) } =\sum \limits _{j=1}^{j=i}a_{ij}\mathbf {Y}_{t}^{\left (j\right ) }.\) The coefficients a i j correspond to the orthonormalization of the polynomials 1, x, x 2, ... with respect to the measure μ(d x) = x 2 ν(d x).
Then \(\left (H_{t}^{\left (i\right ) }\left (t\right ) \right )_{i=1}^{\infty }\) is a family of strongly orthogonal martingales such that \(\left \langle H_{t}^{\left (i\right ) },H^{(j)}\right \rangle _{t}\) = δ i j .t and that [H (i), H (j)] t −〈H (i), H (j)〉 t is an \(\mathcal {F}_{t}\)-martingale, see [17]. We refer the reader to [3, 6, 12] for the detailed proofs.
Throughout what follows, we shall assume the following notations
l 2 : the Hilbert space of real-valued sequences x = (x n ) n≥0 such that
For any integer m ≥ 1, we define
\(l^{2}\left (\mathbb {R}^{m}\right ) :\) the space of \(\mathbb {R}^{m}\)-valued sequences (x i ) i≥1 such that
(a, b) : the inner product in \(\mathbb {R}^{n},\) \(\forall a,b\in \mathbb {R}^{n}.\)
\(\left \vert a\right \vert =\sqrt {\left (a,a\right ) }:\) the norm of \(\mathbb {R}^{n},\) \(\forall a\in \mathbb {R}^{n}.\)
(A, B) : the inner product in \(\mathbb {R}^{n\times d},\) \(\left \vert A\right \vert =\sqrt {\left (A,A\right ) }:\) the norm of \(\mathbb {R}^{n\times n}.\)
\(l_{\mathcal {F}}^{2}\left (0,T,\mathbb {R}^{m}\right ) :\) the Banach space of \(l^{2}\left (\mathbb {R}^{m}\right ) \)- valued \(\mathcal {F}_{t}\)- predictable processes such that
\(\mathcal { \mathcal {L} }_{\mathcal {F}}^{2}\left (0,T,\mathbb {R}^{m}\right ) :\) the Banach space of \(\mathbb {R}^{m}\)-valued \(\mathcal {F}_{t}\)-adapted process such that
\(\mathcal {S}_{\mathcal {F}}^{2}\left (0,T,\mathbb {R}^{m}\right ) :\) the Banach space of \(\mathbb {R}^{m}\)-valued \(\mathcal {F}_{t}\)-adapted and càdlàg process such that
\(L^{2}\left ({\Omega },\mathcal {F},P,\mathbb {R}^{m}\right ) :\) the Banach space of \(\mathbb {R}^{m}\)-valued square integrable random variables on \(\left ({\Omega },\mathcal {F},P\right ) .\)
2.2 Formulation of the Control Problem
Let T be a strictly positive real number. An admissible control is an \(\mathcal {F}_{t}^{Y}\)- predictable process u = (u t ) with values in some convex subset U of \(\mathbb {R}^{k}\) and satisfies \(\mathbb {E}\left [ \underset {0\leq t\leq T}{\sup }\left \vert u_{t}\right \vert ^{2}\right ] <\infty .\) We denote the set of all admissible controls by \(\mathcal {U}\). The control u is called partially observable. Let us also assume that the coefficient of the controlled FBSDE (1) are defined as follows
We assume that the state processes (x, y, z, Z) cannot be observed directly, but the controllers can observe a related noisy process Y, which we call the observation process, via the following Itô process
where
and W υ is an \(\mathbb {R}^{r}\)-valued stochastic processes depending on the control υ. Define d P υ = Γυ d P, where
Obviously, Γυ is the unique \(\mathcal {F}_{t}^{Y}\)-adapted solution of
Then Girsanov’s theorem shows that
is an \(\mathbb {R}^{r}\)-valued brownian motion and \((H_{t}^{\left (i\right ) })_{i=1}^{\infty }\) is still a Teugels martingale under the probability measure P υ.
The objective is to characterize an admissible controls which minimize the following cost functional.
where \(\mathbb {E}^{u}\) denotes the expectation with respect to the probability measure space P u and
It is obvious that the cost functional (4) can be rewritten as the following
Now, we can state our partially observed control problem.
Problem A
Minimize (5) over \(u\in \mathcal {U} \), subject to Eqs. 1 and 3.
A control is said to be partially observed if the control is a non-anticipative functional of the observation Y. A set of controls is said to be partially observed if its every element is partially observed. Hence, the set of admissible controls \(\mathcal {U}\) is partially observed.
An admissible control \(\hat {u}\) is called a partially observed optimal if it attains the minimum of J(u) over \(\mathcal {U}.\) The Eqs. 1 and 2 are called respectively the state and the observation equations, and the solution \(\left (\hat {x},\hat {y},\hat {z},\hat {Z}\right ) \) corresponding to \(\hat {u}\) is called an optimal trajectory.
Throughout this paper, we shall make the following assumptions
(A 1)
-
The random mappings b, g, σ and φ are measurable with \(b(.,0,0)\in \mathcal { \mathcal {L} }_{\mathcal {F}}^{2}\left (0,T,\mathbb {R}^{n}\right ) \), \(g(.,0,0)\in \mathcal { \mathcal {L} }_{\mathcal {F}}^{2}\left (0,T,\mathbb {R}^{n}\right ) \), \(\sigma (.,0,0)\in l_{\mathcal {F}}^{2}\left (0,T,\mathbb {R}^{m}\right ) \) and \(\varphi (0)\in L^{2}\left ({\Omega },\mathcal {F},P,\mathbb {R}^{m}\right ) .\)
-
b, g, σ and φ are continuously differentiable in (x, u). They are bounded by (1 + |x| + |u|) and their derivatives in (x, u) are continuous and uniformly bounded.
-
The random mapping f is measurable with \(f(.,0,0,0,0)\in \mathcal { \mathcal {L} }_{\mathcal {F}}^{2}\left (0,T,\mathbb {R}^{m}\right ) \), f is continuous and continuously differentiable with respect to (x, y, z, Z, u). Moreover it is bounded by (1 + |x| + |y| + |z| + |Z| + |u|) and their derivatives are uniformly bounded.
(A 2)
-
l is continuously differentiable with respect to (x, y, z, Z, u) and bounded by (1 + |x|2 + |y|2 + |z|2 + |Z|2 + |u|2) . Furthermore, their derivatives are uniformly bounded.
-
M is continuously differentiable in x and h is continuously differentiable in y. Moreover, for almost all (t, ω) ∈ [0, T] × Ω, there exists a constant C, for all \((x,y)\in \mathbb {R}^{n}\times \mathbb {R}^{m},\)
$$\begin{array} {ccc} \left\vert M_{x}\right\vert \leq C\left( 1+\left\vert x\right\vert \right) & \text{and} & \left\vert h_{y}\right\vert \leq C\left( 1+\left\vert y\right\vert \right) . \end{array} $$
(A 3) ξ is continuously differentiable in (x, y, z, Z, u) and their derivatives in (x, y, z, Z, u) are uniformly bounded.
Following [13], it holds that under assumptions (A 1), there is a unique solution
which solves the state Eq. 1.
Let \(\left ({x_{t}^{1}},{y_{t}^{1}},{z_{t}^{1}},{Z_{t}^{1}}\right ) \) and \({\Gamma }_{t}^{1}\) be the solutions at time t of the following linear FBSDE and SDE, respectively,
and
where
Set 𝜗 t =Γ−1Γ1 satisfies the following dynamics
For any \(u\in \mathcal {U}\) and the corresponding state trajectory (x, y, z, Z), we introduce the following system of forward backward SDE, called the adjoint equations,
It is clear that (p, k, Q) is the adjoint process corresponding to the forward part of our system (1) and q is corresponding to the backward part. Manifestly, the above FBSDE admit a unique solution
under the assumptions (A 1) . We further introduce the following auxiliary BSDE, which also admit a unique solution under the assumptions (A 1),
Let us now, define the Hamiltonian function
by
3 A Partial Information Necessary Conditions for Optimality
In this section, we derive a partially observed necessary conditions for optimality for our control problem under the previous assumptions. The main objective is to solve the problem A.
3.1 Some Auxiliary Results
Let υ be an arbitrary element of \(\mathcal {U}\), then for a sufficiently small 𝜃 > 0 and for each t ∈ [0, T], we define a perturbed control as follows
Since the action space being convex, it is clear that \(u_{t}^{\theta }\) is an admissible control. Let us now, pointing out that we need the following two lemmas to state and prove the main result of this section. In fact, they play a crucial role in the sequel.
Lemma 4.1
If the assumptions (A 1) and (A 3) hold true, then we have the following estimates
Proof
We first prove (11). Applying Itô’s formula to \(\left \vert x_{t}^{\theta }-x_{t}\right \vert ^{2}\), taking expectations and using the relations 〈H (i), H (j)〉 s = δ i, j .t and [H (i), H (j)] t −〈H (i), H (j)〉 t is an \(\mathcal {F}_{t}\)-martingale together with the fact that b, σ, g are uniformly Lipschitz in (x, u), one can get
Thus (11) follows immediately, by using Gronwall’s lemma and letting 𝜃 go to 0.
Let us now prove (12). Applying Itô’s formula to \(\left \vert y_{t}^{\theta }-y_{t}\right \vert ^{2}\) and taking expectation to obtain
From Young’s inequality, for each ε > 0, we have
Then,
Due the fact that φ and f are uniformly Lipschitz with respect to x, y, z, Z and u, one can get
where \(\alpha _{t}^{\theta }\) is given by
By invoking (11) and sending 𝜃 to 0, we have
We now pick up \(\varepsilon = \frac {1}{2C} \), and replacing its value in Eq. 14 to obtain
Consequently, we obtain the desired result (12), by using Gronwall’s lemma and letting 𝜃 goes to 0. We now proceed to prove (13). Itô’s formula applied to \(\left \vert {\Gamma } _{t}^{\theta }-{{\Gamma }_{t}^{u}}\right \vert ^{2}\) yields
Here, \(\beta _{t}^{\theta }\) is given by the following equality
Keeping in mind that ξ is continuous in (x, y, z, Z, u), it is not difficult to see that
Hence, we obtain (13) by using Gronwall’s lemma and by sending 𝜃 to 0. Before we state and prove the next lemma, let us introduce the following short hand notations,
□
Lemma 4.2
Assume that (A 1) and (A 3) are in force. Then, we have the following convergence results
Proof
First, we start by giving the proof of Eq. 18. By the notation (17) and the first-order expansion, it is easy to check that \(\tilde {x}_{t}^{\theta }\) satisfies the following SDE
where
and
Since b x , b u , g x , g u and σ x , σ u are continuous in (x, u), it is not difficult to see that
Applying Itô’s formula to \(\left (\tilde {x}_{t}^{\theta }\right )^{2}\), we get
Using the inequality 2a b ≤ a 2 + b 2, seeing that \({b_{s}^{x}},\) \({g_{s}^{x}}\) and \({\sigma _{s}^{x}}\) are bounded, to obtain
Finally, by using Gronwall’s lemma and Eq. 22, we obtain (18).
We now turn out to prove (19). Again, in view of the notations (17), one can easily show that \(\left (\tilde {y}_{t}^{\theta },\tilde {z}_{t}^{\theta },\tilde {Z}_{t}^{\theta }\right ) \) satisfies the following BSDE
where \(\tilde {x}_{t}^{\theta }\) is the solution to the SDE (21) and
and
Due the fact that f x , f y , f z and f Z are continuous, we have
Again, Itô’s formula applied to \(\left \vert \tilde {y}_{t}^{\theta }\right \vert ^{2}\) leads to the following equality
By using Young’s inequality, for each ε > 0, we obtain
It follows that, in view of the boundedness of \({f_{t}^{x}},\) \({f_{t}^{y}},\) \({f_{t}^{z}}\) and \({f_{t}^{Z}}\),
where
Hence, in view of Eq. 18, the fact that φ x , \({f_{s}^{x}}\) are continuous and bounded, we get
and
Furthermore, From Eqs. 23 and 25, we deduce that
If we choose \(\varepsilon = \frac {1}{2C} \), it holds that,
The estimates (19) follow from an application of Gronwall’s lemma together with Eqs. 23 and 26.
Now we proceed to prove (20). From Eq. 17, it is plain to check that \(\tilde {\Gamma }^{\theta }\) satisfies the following equality,
where
and \(\bar {\chi }_{t}^{\theta }\) is given by
We deduce, taking into account the fact that ξ x , ξ y , ξ z and ξ Z are continuous,
Applying Itô’s formula to \(\left \vert \tilde {\Gamma }_{t}^{\theta }\right \vert ^{2}\), taking expectation, and using the fact that ξ, \({\xi _{t}^{x}}\), \({\xi _{t}^{y}}\), \({\xi _{t}^{z}}\) and \({\xi _{t}^{Z}}\) are bounded, to obtain
Keeping in mind the relations (18) and (27), we deduce that, using Gronwall’s inequality,
□
3.2 Variational Inequality and Optimality Necessary Conditions
Since u is an optimal control, then, with the fact that \(\theta ^{-1}\left [ J\left (u_{t}^{\theta }\right ) -J\left (u_{t}\right ) \right ] \geq 0,\) we have the following lemma.
Lemma 4.3
Suppose that the assumptions (A 1), (A 2) and (A 3) are satisfied. Then the following variational inequality holds
where l ρ (t) = l ρ (t, x t , y t , z t , Z t , u t )for ρ = x, y, z, Z.
Proof
From the definition of the cost functional and by using the first order development, one can get
Finally by using Eqs. 18, 19, 20 and letting 𝜃 goes to 0, we obtain (28).
In view of Eq. 8, the variational inequality (28) can be rewritten as
□
The main result of this section can be stated us follows.
Theorem 4.4 (Partial information maximum principle)
Suppose (A 1), (A 2), and (A 3) hold. Let (x, y, z, Z, u) be an optimal solution of the control problem A. There are 4-typle (p, q, k, Q) and a pair (P, Ξ) of \(\mathcal {F}_{t}\) -adapted processes which satisfy (9) and (10) respectively, such that the following maximum principle holds true,
Proof
By applying Itô’s formula to \(\left (p_{t},{x_{t}^{1}}\right ) \) and \(\left (q_{t},{y_{t}^{1}}\right ) \) and using the fact that q 0 = h y (y 0) and p T = M x (x T ) + φ x (x T )q T , we have
and
On the other hand, Itô’s formula applied to (𝜗 t , P t ), gives us
Consequently, From Eqs. 31, 32, and 33, we infer that
thus
This together with the variational inequality (29) imply (30), which achieve the proof. □
4 Partial Information Sufficient Conditions of Optimality
In this section, we will prove that the partial information maximum principle condition for the Hamiltonian function is in fact sufficient under additional convexity assumptions. It should be noted that we shall prove our result in two different cases. In the first case, we are going to prove the sufficient condition without assuming the linearity of the terminal condition for the backward part of the state equation. To this end, we restrict ourselves to the one dimensional case n = m = 1 and we state now the main result of this section.
Theorem 5.5
Suppose (A 1), (A 2), and (A 3) hold. Assume further that the functions φ , M and H (t, ., ., ., ., p t , q t , k t , Q t , Ξ t ) are convex, h is convex function and increasing. If the following maximum condition holds
\(\forall \upsilon _{t}\in \mathcal {U},\) a.e, a.s, then u is an optimal control in the sense that \(J\left (u\right ) \leq \inf \limits _{\upsilon \in \mathcal {U}}J\left (\upsilon \right ) .\)
Proof
Let u be an arbitrary element of \(\mathcal {U}\) (candidate to be optimal) and (x u, y u, z u, Z u) is the corresponding trajectory. For any \(\upsilon \in \mathcal {U}\) and its corresponding trajectory (x υ, y υ, z υ, Z υ), by the definition of the cost function (5), one can write
Since h and M are convex
And
Thus
Noting that
then, we have
by using the fact that h is convex function and increasing, we can write
On other hand, by applying Ito’s formula respectively to \({p_{t}^{u}}\left (x_{t}^{\upsilon }-{x_{t}^{u}}\right ) \), \({q_{t}^{u}}\left (y_{t}^{\upsilon } -{y_{t}^{u}}\right ) \) and \({P_{t}^{u}}\left ({\Gamma }_{t}^{\upsilon }-{\Gamma } _{t}^{u}\right ) \), and by taking expectations to the previous inequality, we get
By using the fact H is convex in (x, y, z, Z, u), we get
Substituting (38) into (37), we have
and thus
in view of the condition (34) above and keeping in mind that \({\Gamma }_{t}^{\upsilon }>0,\) one can get J(u) − J(υ) ≤ 0, which achieve the proof. □
Before we treat the second result of this section, it is worth to pointing out that we can prove a partial observed sufficient conditions of optimality without assuming neither that x and y need to be in the dimension one, nor that the function φ needs to be negative and decreasing.
Assume that φ(x) = N x, where N is a nonzero constant matrix with order m × n. Then, by using similar arguments developed above, we can easily state and prove the following theorem which illustrate the second case.
Theorem 5.6
Assume that (A 1), (A 2) and (A 3) are in force. Assume that the functions h(.), M(.) and H (t, ., ., ., ., p t , q t , k t , Q t , Ξ t ) are convex with φ(x) = N x . If further the maximum condition (34) holds true, then u is an optimal control in the sense that
Proof
Let υ be an arbitrary element of \(\mathcal {U}\) and (x υ, y υ, z υ, Z υ) is its corresponding trajectory. By using the definition of cost functional (5), taking under consideration the convexity property of h and M, a simple computation gives us
On the other hand, in view of φ(x) = N x, we apply Ito’s formula to \({p_{t}^{u}}\left (x_{t}^{\upsilon }-{x_{t}^{u}}\right ) \), \(q_{t}^{u}\left (y_{t}^{\upsilon }-{y_{t}^{u}}\right ) \) and \({P_{t}^{u}}\left ({\Gamma }_{t}^{\upsilon }-{{\Gamma }_{t}^{u}}\right ) \), respectively, then by combining their results together with the above inequality one can get
Since H is convex with respect to (x, y, z, Z, u) for almost all (t, w) ∈ [0, T] × Ω,
It turns out, using the condition (34) taking into account the fact that \({\Gamma }_{t}^{\upsilon }>0,\)
This means that u is an optimal partially observed control process and (x u, y u, z u, Z u) is an optimal 4-typle. The proof is complete. □
5 Application
In this section, we consider a partial observed linear quadratic control problem as a particular case of our control problem A. We find an explicit expression of the corresponding optimal control by applying the necessary and sufficient conditions of optimality. Consider the following control problem,
Minimize the expected quadratic cost function
subject to
where the observation state is given by the following SDE,
Define d P υ =Γυ d P and we denote by Γυ the unique \(\mathcal {F}_{t}^{Y}\) adapted solution of
Here, K(.) > 0, L(.) > 0, г (.) > 0, G(.) > 0, R(.) > 0, M 1 ≥ 0, M 2 ≥ 0, A i(.), B j(.) and D(.) are bounded and deterministic, for i = 1, ..., 6, and j = 1, ..., 5.
To overcome this problem, we first write down the Hamiltonian function
and the adjoint equations associated to the system (41)–(43) are given by
and
According to Theorem (4.4), if \(\hat {u}\) is a partial observed optimal control, then it satisfies
∙ Conversely, for the sufficient part, let \(\hat {u}\in \mathcal {U}\) be a candidate to be optimal control and let \(\left (\hat {x},\hat {y},\hat {z},\hat {Z}\right ) \) be the solution to the FBSDE (41) corresponding to \(\hat {u}\) and (p, k, Q, q), (P, Ξ) are the solution to the corresponding solution to Eqs. 45 and 46. It is straight forward to check that the functional H is convex in (x, y, z, Z, u). Thus, If \(\hat {u}\) satisfies (47) and the partially observed maximum principle condition (30) above. Then by applying Theorem (5.6), one can easily check that \(\hat {u}\) is an optimal control of our partially observed control problem.
References
Antonelli F. Backward-forward stochastic differential equations. Ann Appl Probab. 1993;3(3):777–93.
Bahlali K, Edahbi M, Essaki E. BSDE associated with Lévy processes and application to BDIE. Int J Stoch Anal. 2003;16(1):1–17.
Bahlali K, Khelfallah N, Mezerdi B. Optimality conditions for partial information stochastic control problemsdriven by Lévy processes. Syst Cont Lett. 2012;61(11):1079–84.
Davis MH. Martingale representation and all that. Advances in control, communication networks, and transportation systems. Boston: Birkhäuser; 2005. p. 57–68.
El Karoui N, Mazliak L. Backward stochastic differential equations. Longman: Addison Wesley; 1997.
El Karoui N, Peng S, Quenez MC. Backward stochastic differential equations in finance. Math Finance. 1997;7:1–70.
Ma J, Yong J. Forward backward stochastic differential equations and their applications. Lect. Notes in Math. Berlin: Springer; 1999. p. 1702.
Meng QX, Tang MN. Necessary and sufficient conditions for optimal control of stochastic systems associated with Lévy processes. Sci China Ser F-inf Sci. 2009;52:1982–92.
Mitsui K, Tabata MA. A stochastic linear quadratic problem with Lévy processes and its application to finance. Stochastic Proc Appl. 2008;118:120–52.
Nualart D, Schoutens W. Chaotic and predictable representation for Lévy processes with applications in finance. Stochastic Proc Appl. 2000;90:109–22.
Nualart D, Schoutens W. BSDE’s and Feynman-Kac formula for Lévy process with application in finance. Bernouli. 2001;7:761–76.
Pardoux E, Peng S. Adapted solutions of backward stochastic differential equations. Syst Cont Lett. 1990;14:55–61.
Peng S. Backward stochastic differential equations and application to optimal control. Appl Math Optim. 1993;27:125–44.
Peng S, Wu Z. Fully coupled forward backward stochastic differential equations and application to optimal control. SIAM J Control Optim. 1999;37:825–43.
Protter P. Stochastic integration and differential equations. Berlin: Springer Verlag; 1990.
Tang M, Zhang Q. Optimal variational principle for backward stochastic control systems associated with Lévy processes. Sci Chin Math. 2012;55:745–61.
Schoutens W. Stochastic processes and orthogonal polynomials. Springer Science & Business Media. 2012.
Wu Z. A maximum principle for partially observed optimal control of forward-backward stochastic systems. Sci Chin Inf Sci. 2010;53(11):2205–14.
Xiao H. The maximum principle for partially observed optimal control of forward-backward stochastic systems with random jumps. J Syst Sci Complex 2011;24:1083–99.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Bougherara, S., Khelfallah, N. The Maximum Principle for Partially Observed Optimal Control of FBSDE Driven by Teugels Martingales and Independent Brownian Motion. J Dyn Control Syst 24, 201–222 (2018). https://doi.org/10.1007/s10883-017-9377-4
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10883-017-9377-4
Keywords
- Forward-backward stochastic differential equation
- Optimal control
- Maximum principle
- Partially observed optimal control
- Teugels martingale
- Lévy process