Abstract
We investigate Euler discretization for a class of optimal control problems with a nonlinear cost functional of Mayer type, a nonlinear system equation with control appearing linearly and constraints defined by lower and upper bounds for the controls. Under the assumption that the cost functional satisfies a growth condition we prove for the discrete solutions Hölder type error estimates w.r.t. the mesh size of the discretization. If a stronger second-order optimality condition is satisfied the order of convergence can be improved. Numerical experiments confirm the theoretical findings.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Discretization methods like Euler discretization are used for the numerical solution of optimal control problems. The accuracy of the approximate solutions obtained in this way are often satisfactory from a practical point of view. However, if the optimal control has a special structure, a discretization method may help to detect this structure and then methods based on structural assumptions can be used to determine the optimal control more accurately. Especially for bang–bang controls Euler discretization can be used to compute approximations of the switching points, and then efficient numerical approaches, such as switching time parameterization, can be employed to determine the switching times more accurately (see e.g. Kaya et al. [23], Maurer et al. [29], Osmolovskii and Maurer [34] and the papers cited therein). It is well-known that, in particular when the solution controls are of bang–bang or bang-singular type, many difficulties are encountered in getting an approximate solution. Therefore, it is of practical interest to have conditions implying error estimates for approximate solutions which ensure that the approximate controls (optimal controls for the discretized problems) converge to the optimal control of the original, continuous-time control problem. Such error estimates are closely related to estimates for solutions of perturbed optimal control problems.
Discretization and perturbation of nonlinear optimal control problems governed by ordinary differential equations are well studied for the case that the optimal control is sufficiently smooth, and the results are usually based on strong second-order optimality conditions which require coercivity of the second derivative of the Lagrangian function with respect to the control variables (see e.g. Dontchev and Hager [10, 11], Dontchev et al. [12], Malanowski [24,25,26], Malanowski et al. [27], Alt [1,2,3]). For control problems with control appearing linearly such conditions are not satisfied and the optimal control may be discontinuous. Therefore, there have been only a few papers on discretization of such problems (see e.g. Alt and Mackenroth [5], Dhamo and Tröltzsch [9], Veliov [41] and the papers cited therein).
New second-order optimality conditions for optimal control problems with control appearing linearly have been developed during the last 10–15 years (see e.g. Felgenhauer [15,16,17,18, 20], Maurer et al. [29], Osmolovskii and Maurer [32,33,34] and the papers cited therein). In case of bang–bang controls these conditions have been used in Alt et al. [4], Alt and Seydenschwanz [7], and in Seydenschwanz [40] to obtain error estimates for Euler discretization of linear-quadratic optimal control problems governed by ordinary differential equations and in Deckelnick and Hinze [8] for discretizations of elliptic control problems. For convex control problems of Mayer type with a linear system equation and bang–bang solutions Veliov [41] has shown convergence of order 1 for Euler discretization. These results have been extended in Haunschmied et al. [21] under more general conditions based on a result on stability of optimal control problems under strong bi-metric regularity of Quincampoix and Veliov [36]. Pietrus et al. [35] investigate high order discrete approximations to Mayer type problems based on second order Volterra-Fliess approximations. Felgenhauer [19] shows convergence of order 1 for a class of nonlinear optimal control problems, where the linear term in the system equation does not depend on the state variables and the solution has bang-singular-bang structure. Alt et al. [6] prove convergence of order 1 for implicit Euler discretization of a general class of convex, linear-quadratic control problems with bang–bang solutions.
In the present paper we investigate a class of optimal control with a nonlinear cost functional of Mayer type, a nonlinear system equation with control appearing linearly and constraints defined by lower and upper bounds for the controls. Under the assumption that the cost functional satisfies a growth condition of order \(\kappa \ge 1\) we prove for the discrete solutions Hölder type error estimates of order \(1/\kappa \) w.r.t. the mesh size of the discretization. If a stronger second-order condition for the derivative of the Lagrangian w.r.t. the control and a weakened coercivity condition for the second derivative of the Lagrangian are satisfied, the order of convergence can be improved to 1.
We use the following notations: \(\mathbb {R}^n\) is the n-dimensional Euclidean space with the inner product denoted by \(\langle x, y\rangle \) and the norm \(|x| = \langle x,x\rangle ^{1/2}\). For an \(m\times n\)-matrix M we denote the spectral norm by \(\Vert M\Vert = \sup _{|z| \le 1} |Mz|\). Let \(t_0,t_f\in \mathbb {R}\), \(t_0<t_f\). We denote by \(L^1(t_0,t_f;\mathbb {R}^m)\) the Banach space of integrable, measurable functions \(u:[t_0,t_f]\rightarrow \mathbb {R}^m\) with
by \(L^\infty (t_0,t_f;\mathbb {R}^m)\) the Banach space of essentially bounded functions \(u:[t_0,t_f]\rightarrow \mathbb {R}^m\) with the norm
and \(C(t_0,t_f;\mathbb {R}^m)\) is the Banach space of continuous functions \(u:[t_0,t_f]\rightarrow \mathbb {R}^m\) with the norm
For \(p\in \{1,\infty \}\) we denote by \(W^1_p(t_0,t_f;\mathbb {R}^n)\) the spaces of absolutely continuous functions on \([t_0,t_f]\) with derivative in \(L^p(t_0,t_f;\mathbb {R}^n)\), i.e.
with
Let \(X= X_1\times X_2\), where \(X_1 = W^1_1(t_0,t_f;\mathbb {R}^n)\), \(X_2 = L^1(t_0,t_f;\mathbb {R}^m)\). We consider the following optimal control problem:
where g is defined by
Here, \(u(t)\in \mathbb {R}^m\) is the control, and \(x(t)\in \mathbb {R}^n\) is the state of a system at time \(t\in [t_0,t_f]\). Further \(f:\mathbb {R}^n\rightarrow \mathbb {R}\), \(g^{(1)}:\mathbb {R}^n\times [t_0,t_f]\rightarrow \mathbb {R}^n\), \(g^{(2)}:\mathbb {R}^n\times [t_0,t_f]\rightarrow \mathbb {R}^{n\times m}\), and the set \(U\subset \mathbb {R}^m\) is defined by lower and upper bounds, i.e.,
with \(b_{\ell }, b_u\in \mathbb {R}^m\), \(b_{\ell }<b_u\), where all inequalities are to be understood componentwise.
The organization of the paper is as follows. In Sect. 2 we state some basic results. Section 3 introduces the Euler discretization for Problem (OC). In Sect. 4 Hölder type error estimates are derived assuming a growth condition for the cost functional. Under stronger second-order conditions we prove in Sect. 5 convergence of order 1. Section 6 discusses some numerical results.
2 Basic results
We denote by
the set of admissible controls, and by
the feasible set of Problem (OC). For \(\varepsilon >0\) and \((x^*,u^*)\in X\)
is the open ball around \((x^*,u^*)\) with radius \(\varepsilon \).
Definition 1
A pair \((x^*,u^*)\in {{\mathscr {F}}}\) is called a local minimizer of f on \({\mathscr {F}}\) or a local solution of Problem (OC), if there exists \(\varepsilon >0\) such that \(f(x^*(t_f))\le f(x(t_f))\) for all \((x,u)\in {{\mathscr {F}}}\cap {{\mathscr {B}}}_\varepsilon (x^*,u^*)\). \(\Diamond \)
Note that we allow discontinuous optimal controls, especially bang–bang controls. Therefore, we consider local solutions w.r.t. the \(L^1\)-norm for control functions. We suppose in the following:
-
(2.1)
There exists \(\bar{\varepsilon }>0\) and \((x^*,u^*)\in {{\mathscr {F}}}\) such that \(f(x^*(t_f))\le f(x(t_f))\) for all \((x,u)\in {{\mathscr {F}}}\cap {\mathscr {B}}_{\bar{\varepsilon }}(x^*,u^*)\), i.e., \((x^*,u^*)\) is a local solution of (OC).
Since U is bounded, there exists a constant \(K_u\) such that for all \(u\in {{\mathscr {U}}}\)
Let
and let \({{\mathscr {B}}}\subset \mathbb {R}^n\) be a convex and open set such that
For given numbers \(n_1,n_2\in \mathbb {N}\), \(n_1\le n_2\), we define
We suppose that the following assumptions are satisfied:
-
(2.3)
The functions f, \(g^{(1)}\), and \(g^{(2)}\) are continuously differentiable w.r.t. x on \({{\mathscr {B}}}\).
-
(2.4)
The functions f, \(g^{(1)}\), and \(g^{(2)}\) are Lipschitz continuous, i.e., there are constants \(L_f\), \(\tilde{L}_f\) and \(L_g\) such that
$$\begin{aligned}&|f(x)-f(z)|\le L_f\,|x-z|,\\&|g^{(1)}(x,t)-g^{(1)}(z,s)|\le L_g\,(|x-z|+|t-s|),\\&\Vert g^{(2)}(x,t)-g^{(2)}(z,s)\Vert \le L_g\,(|x-z|+|t-s|), \end{aligned}$$for all \(s,t\in [t_0,t_f]\) and all \(x,z\in {{\mathscr {B}}}\)
-
(2.5)
The functions \(f_x\), \(g^{(1)}_x\), and \(g^{(2)}_x\) are Lipschitz continuous, i.e., there are constants \(L_f^{(1)}\) and \(L_g^{(1)}\) such that
$$\begin{aligned}&|f_x(x)-f_x(z)|\le L_f^{(1)}\,|x-z|,\\&|g^{(1)}_{j,x}(x,t)-g^{(1)}_{j,x}(z,s)| \le L_g^{(1)}\,(|x-z|+|t-s|),\; j\in J_1^n,\\&|g^{(2)}_{ji,x}(x,t)-g^{(2)}_{ji,x}(z,s)| \le L_g^{(1)}\,(|x-z|+|t-s|),\; j\in J_1^n,\;i\in J_1^m, \end{aligned}$$for all \(s,t\in [t_0,t_f]\) and all \(x,z\in {{\mathscr {B}}}\).
For \((x,u)\in X\) with \(\Vert x-x^*\Vert \le \bar{\varepsilon }\) it follows from (2.2) and (2.5) that
This implies
with some constant \(K_g\) independent of \((x,u)\in X\) with \(\Vert x-x^*\Vert \le \bar{\varepsilon }\). Moreover, for \((x,u)\in {\mathscr {F}}\) with \(x\in {{\mathscr {N}}}_{\bar{\varepsilon }}(x^*)\) and \(t,s\in [t_0,t_f]\) we have
By (2.2), Assumption (2.4), and (2.6) this implies
This further implies that with some constant \(L_x\)
for all \((x,u)\in {{\mathscr {F}}}\) with \(x\in {\mathscr {N}}_{\bar{\varepsilon }}(x^*)\), which shows that the feasible trajectories \(x\in {{\mathscr {N}}}_{\bar{\varepsilon }}(x^*)\) are uniformly Lipschitz with Lipschitz modulus \(L_x\).
The Hamiltonian \({{\mathscr {H}}}:\mathbb {R}^n\times \mathbb {R}^m\times \mathbb {R}^n \times [t_0,t_f]\rightarrow \mathbb {R}\) for Problem (OC) is defined by
We denote by
the i-th column vector of \(g^{(2)}(x,t)\) and by
the j-th row of \(g^{(2)}(x,t)\). Then
Optimality conditions for Problem (OC) are well-known. Let \((x^*,u^*)\in {{\mathscr {F}}}\) be a local solution of (OC). Then there exists a function \(\lambda ^*\in W_1^1(t_0,t_f;\mathbb {R}^n)\) such that the adjoint equation
is satisfied for a.a. \(t\in [t_0,t_f]\) with terminal condition \(\lambda ^*(t_f)=f_x(x^*(t_f))^\mathsf {T}\), and the local minimum principle
holds for a.a. \(t\in [t_0,t_f]\) and all \(u\in U\).
We denote by \(\sigma ^*:[t_0,t_f]\rightarrow \mathbb {R}^m\) the switching function defined by
For a strong local solution \((x^*,u^*)\in {{\mathscr {F}}}\) of Problem (OC) with associated adjoint function \(\lambda ^*\in X_1\), (2.10) implies for \(i\in \{1,\ldots ,m\}\)
Therefore, the optimal control \(u^*\) is of bang–bang type or may have singular arcs.
3 Euler Approximation
Let \(N\in \mathbb {N}\), \(h=h_N=(t_f-t_0)/N\) be the mesh size and \(t_j=t_0+jh\), \(j\in J_0^{N}\), the grid points of the discretization. We approximate the space \(X_2\) of controls by functions in the subspace \(X_{2,N}\subset X_2\) of piecewise constant functions \(u_h\) represented by their values \(u_h(t_j) = u_{h,j}\) at the grid points \(t_j\), \(j\in J_0^{N-1}\). Further, we approximate state and adjoint state variables by functions \(x_h\), resp. \(\lambda _h\), in the subspace \(X_{1,N}\subset X_1\) of continuous, piecewise linear functions represented by their values \(x_h(t_j)=x_{h,j}\), resp. \(\lambda _h(t_j)=\lambda _{h,j}\), at the grid points \(t_j\), \(j\in J_0^N\). Then based on Euler’s method for the discretization of the system equation we obtain the discrete optimal control problem
By \({{\mathscr {F}}}_N\) we denote the feasible set of \(\text{(OC) }_N\).
Definition 2
A pair \((x^*_h,u^*_h)\in {{\mathscr {F}}}_N\) is called a local minimizer of f on \({{\mathscr {F}}}_N\) or a local solution of Problem \(\text{(OC) }_N\), if there exists \(\varepsilon >0\) such that \(f(x^*_{h,N})\le f(x_{h,N})\) for all \((x_h,u_h)\in {{\mathscr {F}}}_N\cap {{\mathscr {B}}}_\varepsilon (x^*_h,u^*_h)\). \(\Diamond \)
Since \({{\mathscr {F}}}_N\) is nonempty and bounded, Problem \(\text{(OC) }_N\) has a (global) solution. Optimality conditions can be derived in the same way as in Ioffe and Tihomirov [22, Section 6.4]. For any local solution \((x^*_h,u^*_h)\in {{\mathscr {F}}}_N\) of Problem \(\text{(OC) }_N\) there exists a multiplier \(\lambda ^*_h\) such that the discrete adjoint equation
for \(j\in J_0^{N-1}\) with terminal condition \(\lambda ^*_{h,N} = f_x(x^*_{h,N})^{\mathsf {T}}\), and the discrete minimum principle
for \(j\in J_0^{N-1}\) and all \(u\in U\) are satisfied.
By \(\lambda ^*_h\) we denote the continuous, piecewise linear function defined by the values \(\lambda _h(t_j)=\lambda _{h,j}\), \(i=0,\ldots ,N\), and by \(\sigma ^*_h(t)\) we denote the continuous, piecewise constant function defined by the values
the discrete analogue of the switching function (2.11). From (3.2) we obtain for \(i=1,\ldots ,m\), \(j\in J_0^{N-1}\),
4 Error estimates for local minimizers
We first prove some auxiliary results. For a function \(z:[t_0,t_f]\rightarrow \mathbb {R}^k\) of bounded variation and \(s_1,s_2\in [t_0,t_f]\), \(s_1<s_2\), we denote by \(\displaystyle \mathsf {V}_{s_1}^{s_2}z\) the total variation of z on \([s_1,s_2]\).
Lemma 1
Suppose that \(u\in X_2\) has bounded variation, and let \(u_h\in X_{2,N}\) be the piecewise constant function defined by the values \(u_{h,j}=u(t_j)\), \(j\in J_0^{N-1}\). Then
Proof
Since for \(s\in [t_j,t_{j+1}]\)
we have
which proves (4.1). \(\square \)
Remark 1
In many applications the optimal control \(u^*\) is a piecewise Lipschitz continuous function. In this case \(u^*\) has bounded variation. \(\Diamond \)
The following result is a special case of Sendov and Popov [39, Theorem 6.1].
Lemma 2
If Assumptions (2.3) and (2.4) are satisfied, \((x,u)\in {{\mathscr {F}}} \cap {{\mathscr {B}}}_\varepsilon (x^*,u^*)\), \(\dot{x}\) has bounded variation, and \(x_h\) is the solution of the discrete system equation
then
where \(c_1= e^{(t_f-t_0)L_g(1+K_u)}\) is a constant independent of N.
Lemma 3
Suppose that Assumptions (2.1), (2.3), and (2.4) are satisfied, and that \(u^*\) has bounded variation. Then for \((x,u)\in {{\mathscr {F}}}\cap {{\mathscr {B}}}_\varepsilon (x^*,u^*)\) we have
where \(c_2\) is a constant independent of N.
Proof
The variation of \(\dot{x}\) can be estimated by the variation of the right hand side of the system equation. For \(t,s\in [t_0,t_f]\) we have by (2.7)
Hence, by (2.8) we obtain
which proves the assertion. \(\square \)
For \(\rho >0\) we consider the auxiliary problem
which is Problem \(\text{(OC) }_{N}\) with the additional constraints
For \(\rho >0\) we denote by \({{\mathscr {F}}}_{N,\rho }\) the feasible set of Problem \(\text{(OC) }_{N,\rho }\), i.e.
Lemma 4
Suppose that Assumptions (2.1), (2.3), and (2.4) are satisfied, and that \(u^*\) has bounded variation. Further let \(\rho >0\) be arbitrary but fixed. Then for sufficiently large N Problem \(\text{(OC) }_{N,\rho }\) has a solution.
Proof
Let \({\hat{u}}_h\in X_{2,N}\) be defined by the values \({\hat{u}}_{h,j}=u^*(t_j)\), \(j\in J_0^{N-1}\). Then \({\hat{u}}_h\in {{\mathscr {U}}}\), and by Lemma 1 we have
Let \({\hat{x}}_h\) be the solution of the discrete system Eq. (4.2) for \(u=u^*\). By Lemmas 2 and 3 we have
This shows that \(({\hat{x}}_h,{\hat{u}}_h)\in {{\mathscr {F}}}_{N,\rho }\), and hence \({{\mathscr {F}}}_{N,\rho }\not =\emptyset \) for sufficiently large N. Since the cost functional is continuous and the feasible set is closed and bounded, a solution exists. \(\square \)
The following result on the dependence of solutions of the system equation on parameters is well-known.
Lemma 5
If Assumptions (2.1), (2.3), and (2.5) are satisfied, there exists \(\bar{\rho }\in \,]0,\bar{\varepsilon }]\) such that for each \(u\in X_2\) with \(\Vert u-u^*\Vert _1<\bar{\rho }\), and each \(\eta \in L^1(t_0,t_f;\mathbb {R}^n)\) with \(\Vert \eta \Vert _1<\bar{\rho }\) the perturbed system equation
has a unique solution \(x = x(u,\eta )\), and for \(x_i=x(u_i,\eta _i)\), \(i=1,2\), we have
where the constant \(c_s\) is independent of u and \(\eta \).
We further need the following auxiliary result.
Lemma 6
Suppose that Assumptions (2.1), (2.3), (2.4), and (2.5) are satisfied, and let \(\bar{\rho }>0\) be given by Lemma 5. Then there is a number \(\bar{N}\in \mathbb {N}\) such that for \(N\ge {\bar{N}}\) and any \((x_h,u_h)\in {\mathscr {F}}_{N,\bar{\rho }}\) there exists a function \(z\in X_1\) such that \((z,u_h)\in {{\mathscr {F}}}\) and
with a constant c independent of N and \((x_h,u_h)\in {\mathscr {F}}_{N,\bar{\rho }}\).
Proof
Let \((x_h,u_h)\in {{\mathscr {F}}}_{N,\bar{\rho }}\) and \(N\in \mathbb {N}\) be given. Since \(\Vert x_h-x^*\Vert \le \bar{\rho }\le \bar{\varepsilon }\) it follows from (2.6) that
By Lemma 5 the system equation of (OC) for \(u=u_h\), i.e.,
has a unique solution z, i.e. \((z,u_h)\in {{\mathscr {F}}}\). We define the piecewise constant function \({\bar{x}}_h:[t_0,t_f]\rightarrow \mathbb {R}^n\) by \({\bar{x}}_h(t_j) = x_h(t_j)\) for \(j\in J_0^{N-1}\). Then the discrete system Eq. (4.2) for \(x_h\) can be written in the form
where
Since \(x_h\) solves the discrete system Eq. (4.2) we have by (4.7) for \(t\in [t_j,t_{j+1}[\),
and by (2.2), Assumption (2.4), and (2.8) it follows that for \(t\in [t_0,t_f]\),
We choose \({\bar{N}}\in \mathbb {N}\) such that
for \(N\ge {\bar{N}}\). Then
for \(N\ge {\bar{N}}\). By Lemma 5 this implies
which proves (4.6). \(\square \)
In order to obtain error estimates for local solutions we proceed similarly to Alt [1, 2] (compare also Alt et al. [6]). In addition to (2.1) we use the following growth condition for the cost functional:
-
(4.8)
There exist \(\alpha >0\), \(\kappa \ge 1\) such that
$$\begin{aligned} f(x(t_f))-f(x^*(t_f)) \ge \alpha \,\Vert u-u^*\Vert _1^\kappa \end{aligned}$$for all \((x,u)\in {{\mathscr {F}}}\cap {\mathscr {B}}_{\bar{\varepsilon }}(x^*,u^*)\).
Remark 2
The growth condition required here implies that \((x^*,u^*)\) is a strict local solution of (OC). Such conditions are closely related to second-order optimality conditions (see e.g. Ioffe and Tihomirov [22, Chapter 7] or Maurer and Zowe [31, Theorem 5.6]). In the following section we use the stronger second-order optimality condition (5.7) implying (4.8) with \(\kappa =2\) (see Theorem 3).
Theorem 1
Let Assumptions (2.1), (2.3), (2.4), (2.5), and (4.8) be satisfied and suppose that \(u^*\) has bounded variation. Then for each \(0<\rho <\bar{\rho }\), where \(\bar{\rho }>0\) is given by Lemma 5, Problem \(\text{(OC) }_{N,\rho }\) has a global solution for sufficiently large N. Further for each such solution \((x^*_h,u^*_h)\) the estimates
hold with constants \(c_u\), \(c_x\) independent of N and the solution \((x^*_h,u^*_h)\).
Proof
We choose \(N\ge {\bar{N}}\) sufficiently large, where \({\bar{N}}\) is defined by Lemma 6. Then by Lemma 4 Problem \(\text{(OC) }_{N,\rho }\) has a (global) solution. Let \((x^*_h,u^*_h)\) be any such solution. By Lemma 6 there exists a function \(z^*\in X_1\), such that \((z^*,u^*_h)\in {{\mathscr {F}}}\) and
with a constant \(c_1\) independent of N and \((x^*_h,u^*_h)\). Further, since \((x^*_h,u^*_h)\in {{\mathscr {F}}}_{N,\rho }\) we have \(\Vert x^*_h-x^*\Vert _\infty \le \rho \). Together with (4.10) and the fact that \(\Vert z^*-x^*_h\Vert _\infty \le \Vert z^*-x^*_h\Vert _{1,1}\) this implies
and therefore
for sufficiently large N. Since \((x^*_h,u^*_h)\in {\mathscr {F}}_{N,\rho }\) we have \((z^*,u^*_h)\in {{\mathscr {F}}}\cap {{\mathscr {B}}}_{\bar{\varepsilon }}(x^*,u^*)\). By (4.8) we therefore have
Further, by (2.4) and (4.10) we have
and therefore
for N sufficiently large.
Let \({\hat{u}}_h\in X_{2,N}\) be defined as in the proof of Lemma 4. Then for sufficiently large N we have \(({\hat{x}}_h,{\hat{u}}_h)\in {{\mathscr {F}}}_{N,\rho }\) (see proof of Lemma 4) and therefore \(f({\hat{x}}_h(t_f)) \ge f(x^*_h(t_f))\). Further we have
with a constant \(c_2\) independent of N. By (4.12), (2.4) this implies
In the proof of Lemma 6 we have shown that the discrete system Eq. (4.2) for \(x_h=x^*_h\) can be written in the form
where \(|\eta (t)|\le c_3h_N\) with a constant \(c_3\) independent of N. By Lemma 5 we therefore obtain
where the constant \(c_4\) is independent of u and N. \(\square \)
Remark 3
Note that Theorem 1 assumes that \((x^*_h,u^*_h)\) is a global solution of Problem \(\text{(OC) }_{N,\rho }\). For such a solution we have \(\Vert u^*_h-u^*\Vert _1<\rho \) and \(\Vert x^*_h-x^*\Vert _{1,1}<\rho \) for sufficiently large N, i.e. the additional constraints (4.5) are not active, and \((x^*_h,u^*_h)\) is a local minimizer of Problem \(\text{(OC) }_N\). Similar results on the existence of approximate local minimizers for control problems obtained by Euler discretization and error estimates for the discrete solutions are well-known in case that the optimal control is continuous (see e.g. Malanowski et al. [27], Dontchev and Hager [11], Dontchev et al. [12, 13]). In these papers a strong second-order sufficient optimality condition is used which also implies local uniqueness of the discrete solutions. This can not be shown under the weaker condition (4.8) used here. \(\Diamond \)
If \((x^*_h,u^*_h)\) is a global solution of Problem \(\text{(OC) }_{N,\rho }\), then by Remark 3 \((x^*_h,u^*_h)\) is a local minimizer of Problem \(\text{(OC) }_N\). Therefore a multiplier \(\lambda ^*_h\) exists satisfying the discrete adjoint Eq. (3.1). In order to derive an error estimate for this multiplier we need some auxiliary results. Since the adjoint equation is a linear differential equation one easily obtains the following result.
Lemma 7
Suppose that Assumptions (2.1), (2.3), (2.4), and (2.5) are satisfied. Let \(\bar{\rho }>0\) be given by Lemma 5 and \(0<\rho \le \bar{\rho }\). Then if N is sufficiently large we have for any solution \((x^*_h,u^*_h)\) of Problem \(\text{(OC) }_{N,\rho }\) and the associated adjoint function \(\lambda ^*_h\) the estimate
with a constant \(K_\lambda \) independent of N and the solution \((x^*_h,u^*_h)\).
In the proof of Lemma 6 we have shown that the discrete state variables can be viewed as the solution of a perturbation of the system equation of Problem (OC). In the same way one can show that the discrete adjoint variables \(\lambda ^*_h\) can be viewed as the solution of a perturbation of the adjoint Eq. (2.9).
Lemma 8
Suppose that Assumptions (2.1), (2.3), (2.4), and (2.5) are satisfied. Let \(\bar{\rho }>0\) be given by Lemma 5 and \(0<\rho \le \bar{\rho }\). Then, if N is sufficiently large, we can write the discrete adjoint equation (3.1) in the form
for a.a. \(t\in [t_0,t_f]\), where the function \(\xi _h:[t_0,t_f]\rightarrow \mathbb {R}^n\) can be estimated by
for a.a. \(t\in [t_0,t_f]\) with a constant \(c_\xi \) independent of N and the solution \((x^*_h,u^*_h)\).
Now we can derive an error estimate for the discrete adjoint functions.
Theorem 2
Let the assumptions of Theorem 1 be satisfied and suppose that \(u^*\) has bounded variation. Then for each \(0<\rho <\bar{\rho }\), where \(\bar{\rho }>0\) is given by Lemma 5, Problem \(\text{(OC) }_{N,\rho }\) has a (global) solution for sufficiently large N. Further for each such solution \((x^*_h,u^*_h)\) and the associated adjoint function \(\lambda ^*_h\) the estimate
holds with a constant \(c_\lambda \) independent of N and the solution \((x^*_h,u^*_h)\).
Proof
We define \(\lambda := \lambda ^*_h-\lambda ^*\). By (2.9) and (4.17) we have
with terminal condition \(\lambda (t_f) = f_x(x^*_{h,N})^{\mathsf {T}} - f_x(x^*(t_f))^\mathsf {T}\). Since this is a linear differential equation it follows that
with some constant \({\bar{c}}_{\lambda ,1}\) independent of N and \((x^*_h,u^*_h))\). Finally, together with (4.15) we obtain
with a constant \(c_{\lambda ,2}\) independent of N and the solution \((x^*_h, u^*_h)\). By Theorem 1 this implies (4.19). \(\square \)
5 Improved error estimates
We can improve the error estimates of the last section to order 1, if we replace condition (4.8) by a stronger second-order sufficient optimality condition. To this end we require in addition to Assumptions (2.1), (2.3), (2.4), and (2.5):
-
(5.1)
The functions f, \(g^{(1)}\), and \(g^{(2)}\) are twice continuously differentiable w.r.t. x on \({{\mathscr {B}}}\).
-
(5.2)
The functions \(f_{xx}\), \(g^{(1)}_{xx}\), and \(g^{(2)}_{xx}\) are Lipschitz continuous, i.e., there are constants \(L_f^{(2)}\) and \(L_g^{(1)}\) such that for all \(s,t\in [t_0,t_f]\) and all \(x,z\in {{\mathscr {B}}}\)
$$\begin{aligned}&|f_{xx}(x)-f_{xx}(z)| \le L_f^{(2)}\,|x-z|,\\&|g^{(1)}_{j,xx}(x,t)-g^{(1)}_{j,xx}(z,s)| \le L_g^{(2)}\,(|x-z|+|t-s|),\; j\in J_1^n,\\&|g^{(2)}_{ji,xx}(x,t)-g^{(2)}_{ji,xx}(z,s)| \le L_g^{(2)}\,(|x-z|+|t-s|),\; j\in J_1^n,\;i\in J_1^m. \end{aligned}$$
Remark 4
Assumptions (5.1), (5.2) imply Assumptions (2.3), (2.4), and (2.5). \(\Diamond \)
As in Alt [2, Section 6] we can formulate Problem (OC) as an abstract optimization problem of type
where \(z=(x,u)\in X\), \(F:X\rightarrow \mathbb {R}\) is defined by
\(G:X\rightarrow Y:= L^1(t_0,t_f;\mathbb {R}^n)\times \mathbb {R}^n\) is defined by
and \(C=X_1\times {{\mathscr {U}}}\), \(K=\{0_Y\}\). As shown in Alt [2] it then follows by results of Robinson [37, 38] that the set
approximates the feasible set of Problem (OC) in the sense of Maurer and Zowe [31, Definition 4.1]. From Alt [2, Lemma 2.1] and Lemma 5 we therefore get the following result:
Lemma 9
Let Assumptions (2.1), (2.3), (2.4), and (2.5) be satisfied. Then for each \(\gamma >0\) there exists \(\rho (\gamma )>0\) such that for each \((x,u)\in {{\mathscr {F}}}\) with \(\Vert u-u^*\Vert _1 < \rho (\gamma )\) there exists \(({\bar{x}},{\bar{u}})\in T(x^*,u^*)\) with
For \(\lambda \in L^\infty (t_0,t_f;\mathbb {R}^n)\) we define the Lagrange function by
It follows from the adjoint Eq. (2.9) that
for all \(x\in X_1\) with \(x(t_0)=0\), and by the local minimum principle (2.10) we have
for all \(u\in {\mathscr {U}}\), where \(\sigma ^*\) is the switching function defined by (2.11).
By \({{\mathscr {L}}}''\) we denote the second derivate of \({{\mathscr {L}}}\) w.r.t. (x, u). Since the control u appears only linearly in Problem (OC), we have
and therefore
for all \(({\bar{x}},{\bar{u}},\bar{\lambda }), (x,u,\lambda )\in X\times X_1\).
If (5.1) is satisfied, then there exists a constant \(C_{\!{\mathscr {L}}}\) such that
for all \(({\bar{x}},{\bar{u}},\bar{\lambda })\in X\times X_1\) with \(\Vert \bar{x}-x^*\Vert _\infty <\bar{\rho }\), \(\Vert \bar{\lambda }-\lambda ^*\Vert _\infty <\bar{\rho }\), \({\bar{u}}\in {\mathscr {U}}\), and all \((x_1,u_1)\), \((x_2,u_2)\in X\).
In case of a continuous optimal control convergence of order 1 can be shown for Euler approximation if a strong second-order optimality condition is satisfied which especially requires that the bilinear form \({{\mathscr {L}}}''(x^*,u^*,\lambda ^*)\) is positive definite w.r.t. to the control function (compare e.g. Dontchev et al. [13]). By (5.5) this condition cannot be satisfied for the class of control problems considered here. We use instead a second-order condition for the switching function \(\sigma ^*\) defined by (2.11). This condition has been introduced by Felgenhauer [15] (see also Maurer and Osmolovskii [30], Maurer et al. [29]) and has been used e.g. in Alt et al. [6], Alt and Seydenschwanz [7], Seydenschwanz [40] to investigate Euler discretization of linear quadratic control problems:
-
(5.7)
There exists \(\bar{\alpha }> 0\) such that
$$\begin{aligned} \int _{t_0}^{t_f} \sigma ^*(t)^{{\mathsf {T}}}\left( u(t)-u^*(t)\right) \,{\mathrm {d}}t= {{\mathscr {L}}}_u(x^*,u^*,\lambda ^*)(u-u^*) \ge \bar{\alpha }\,\Vert u-u^*\Vert _1^2, \end{aligned}$$for all \(u\in {\mathscr {U}}\).
Remark 5
One should note that Assumption (5.7) excludes singular arcs of the optimal control, i.e., the optimal control \(u^*\) must be of bang–bang type. As shown in Alt et al. [6, Lemma 4] the assumption is satisfied if the optimal control is of bang–bang type with finitely many boundary arcs and if an additional growth condition for the switching function around its zeros holds. \(\Diamond \)
In Alt and Seydenschwanz [7] and Alt et al. [6] we used an additional assumption ensuring convexity of the linear-quadratic control problems considered there. Here we use the somewhat weaker assumption:
-
(5.8)
There exists \(\beta > 0\) such that \(\alpha :=\bar{\alpha }-\beta >0\) and
$$\begin{aligned}&z(t_f)^{{\mathsf {T}}} f_{xx}\left( x^*(t_f)\right) z(t_f) + \int _{t_0}^{t_f}\!z(t)^{{\mathsf {T}}} {{\mathscr {H}}}_{xx}\left( x^*(t),u^*(t),\lambda ^*(t),t\right) z(t)\,{\mathrm {d}}t\\&\quad +2\int _{t_0}^{t_f}\!z(t)^{{\mathsf {T}}} {{\mathscr {H}}}_{xu}\left( x^*(t),u^*(t),\lambda ^*(t),t\right) v(t)\,{\mathrm {d}}t\\&\quad = {\mathscr {L}}''(x^*,u^*,\lambda ^*)\left( (z,v),(z,v)\right) \ge -\beta \,\Vert v\Vert _1^2 \end{aligned}$$for all \((z,v)=(x,u)-(x^*,u^*)\) with \((x,u)\in T(x^*,u^*)\).
Remark 6
The condition \((z,v)=(x,u)-(x^*,u^*)\) with \((x,u)\in T(x^*,u^*)\) is equivalent to \(u\in {{\mathscr {U}}}\), \(z(t_0)=0\) and
for a.a. \(t\in [t_0,t_f]\). Therefore, \(\Vert z\Vert _{1,1}\le c\,\Vert v\Vert _1\) with a constant c independent of v. \(\Diamond \)
Example 1
It can easily be seen that Assumption (6) used in Alt et al. [6] for a class of linear quadratic control problems is equivalent to (5.8).
If the system equation is linear then the coercivity condition in (5.8) reads
for all \((z,v)=(x,u)-(x^*,u^*)\) with \((x,u)\in T(x^*,u^*)\). \(\Diamond \)
We now show, that Assumptions (5.7) and (5.8) imply the growth condition (4.8) with \(\kappa =2\). The proof is based on a result of Ioffe and Tihomirov [22, Chapter 7] concerning a general second-order sufficient optimality condition for equality constrained optimization problems. More general results can be found in Maurer [28], and Maurer and Zowe [31] (see also Alt [2]). A general result on sufficient optimality conditions for optimal control problems can be found in Felgenhauer [15]. We need some auxiliary results which are modifications of corresponding results in Sect. 3 of Alt [2].
Lemma 10
Let Assumptions (2.1), (5.1), (5.2), (5.7), and (5.8) be satisfied. Then there exists \(0<\delta _1\le \bar{\rho }\) such that
for all \((z,v)=(x,u)-(x^*,u^*)\) with \((x,u)\in T(x^*,u^*)\) and all \(({\bar{x}},{\bar{u}},\bar{\lambda })\in X\times X_1\) with \(\Vert {\bar{x}}-x^*\Vert _\infty +\Vert {\bar{u}}-u^*\Vert _1 +\Vert \bar{\lambda }-\lambda ^*\Vert _\infty < \delta _1\).
Proof
Let \((z,v)=(x,u)-(x^*,u^*)\) with \((x,u)\in T(x^*,u^*)\) and \(({\bar{x}},{\bar{u}},\bar{\lambda })\in X\times X_1\) with \(\Vert \bar{x}-x^*\Vert _\infty +\Vert {\bar{u}}-u^*\Vert _1 +\Vert \bar{\lambda }-\lambda ^*\Vert _\infty < \bar{\varepsilon }\). By Assumption (5.8) we have
i.e.,
By (5.2) the absolute value of the first term on the right hand side of this inequality can be estimated by
with some constant \(c_1\) independent of z, v, \({\bar{x}}\), \(\bar{u}\), and \(\bar{\lambda }\). Using
the absolute value of the second term on the right hand side of (5.9) can be estimated by
with some constant \(c_2\) independent of z, v, \({\bar{x}}\), \(\bar{u}\), and \(\bar{\lambda }\). In the same way it can be shown that the absolute value of the third term on the right hand side of (5.9) can be estimated by
with some constant \(c_3\) independent of z, v, \({\bar{x}}\), \(\bar{u}\), and \(\bar{\lambda }\). Since z satisfies the linear differential equation
with initial condition \(z(t_0)=0\) we have
with a constant \(c_4\) independent of x and u. Now combining (5.10)–(5.13) the absolute value of the right hand side of (5.9) can be estimated by
The assertion then follows if we choose \(\delta _1>0\) small enough. \(\square \)
Lemma 11
Let Assumptions (2.1), (5.1), (5.2), (5.7), and (5.8) be satisfied. Then there exists \(0<\delta _2\le \delta _1\) such that
for all \((x,u)\in {{\mathscr {F}}}\) with \(\Vert u-u^*\Vert _1 < \delta _2\) and all \(({\bar{x}},{\bar{u}},\bar{\lambda })\in X\times X_1\) with \(\Vert \bar{x}-x^*\Vert _\infty + \Vert {\bar{u}}-u^*\Vert _1 + \Vert \bar{\lambda }-\lambda ^*\Vert _\infty < \delta _2\).
Proof
We choose \(\gamma >0\) such that
where \(C_{\!{\mathscr {L}}}\) is defined by (5.6) and \(c_s\) is the constant defined by Lemma 5. Let \(\rho (\gamma )\) be defined by Lemma 9. We choose \(0<\delta _2\le \delta _1\) such that
Then for \((x,u)\in {{\mathscr {F}}}\) with \(\Vert u-u^*\Vert _1 < \delta _2\) we have by Lemma 5
Hence by Lemma 9 there exists \(({\bar{z}},{\bar{v}})\in T(x^*,u^*)\) with
Together with (5.15) this implies
and therefore
Further using \((x-x^*,u-u^*)=({\bar{z}}-x^*+x-{\bar{z}},\bar{v}-u^*+u-{\bar{v}})\) we obtain
Next we estimate the terms on the ride hand side of (5.18). By Lemma 10 and (5.17) we have
By (5.17) this implies
Again by (5.6) and (5.16) we obtain
Now inserting the estmates (5.19), (5.20), (5.21) into (5.18) the assertion follows from (5.14). \(\square \)
We can now show, that Assumptions (5.7) and (5.8) imply the growth condition (4.8) with \(\kappa =2\).
Theorem 3
Let Assumptions (2.1), (5.1), (5.2), (5.7), and (5.8) be satisfied. Then
for all \((x,u)\in {{\mathscr {F}}}\) with \(\Vert u-u^*\Vert _1 < \delta _2\), where \(\delta _2\) is defined by Lemma 11. Moreover, condition (4.8) is satisfied with \(\kappa =2\).
Proof
For arbitrary \((x,u)\in {{\mathscr {F}}}\) with \(\Vert u-u^*\Vert _1 < \delta _2\) we have
Using \(x(t_0)-x^*(t_0)=0\) and (5.3) we get by Taylor expansion
where \((z,v)=(1-\tau )(x^*,u^*)+\tau (x,u)\) with \(\tau \in \,]0,1[\). By (5.7) and Lemma 11 this implies
which proves the first part of the assertion. For arbitrary \((x,u)\in {{\mathscr {F}}}_{\delta _2}\) we have \((x,u)\in {{\mathscr {F}}}\) and \(\Vert u-u^*\Vert _1<\delta _2\) which implies that condition (4.8) is satisfied with \(\kappa =2\), \(\alpha \) replaced by \(\frac{3}{4}\alpha \), and \(\bar{\varepsilon }\) replaced by \(\delta _2\). \(\square \)
Remark 7
Note that for the proof of Theorem 3 we only need the fact, that \((x^*,u^*)\) is feasible and satisfies together with the unique solution \(\lambda ^*\) of the adjoint equation the minimum principle and Assumptions (2.3), (2.4), (2.5), (5.1), (5.2), (5.7), and (5.8). Theorem 3 then shows, that \((x^*,u^*)\) is a strict local solution, i.e., Assumptions (5.8), (5.7) can be viewed as sufficient optimality condition. In case of linear-quadratic control problems as considered in Alt et al. [6] the second derivatives of the Hamiltonian and the Lagrange function do not depend on (x, u). This allows to use a more general version of condition (5.7). \(\Diamond \)
For the derivation of error estimates of order 1 for the discrete solutions we proceed similarly to Dontchev and Veliov [14], Haunschmied et al. [21] and use the fact that the discrete solutions can be interpreted as solution of a perturbation of Problem (OC). This approach has also been used in Alt et al. [6] for linear-quadratic control problems. For the more general class of nonlinear control problems considered here we adapt results of Alt [2, Section 3], where Lipschitz continuity of perturbed solutions of nonlinear optimization problems has been studied.
Lemma 12
Let Assumptions (2.1), (5.1), (5.2), (5.7), and (5.8) be satisfied. Further let \(0<\rho <\bar{\rho }\), where \(\bar{\rho }>0\) is given by Lemma 5, and let \((x^*_h,u^*_h)\) be a (global) solution of Problem \(\text{(OC) }_{N,\rho }\). Then there is a function \(\zeta _h:[t_0,t_f]\rightarrow \mathbb {R}^m\) satisfying
such that
Proof
By the discrete minimum principle (3.2) we have
We define piecewise constant functions \({\bar{x}}_h:[t_0,t_f]\rightarrow \mathbb {R}^n\), \(B_h:[t_0,t_f]\rightarrow \mathbb {R}^{n\times m}\), and \(\bar{\lambda }_h:[t_0,t_f]\rightarrow \mathbb {R}^n\) by
for \(t\in [t_j,t_{j+1}[\), \(j\in J_0^{N-1}\). Then we can write the discrete switching function \(\sigma ^*_h\) defined by (3.3) in the form
where \(\zeta _h\) is defined by
Further we can write the discrete minimum principle (5.24) in the form
for a.a. \(t\in [t_0,t_f]\), From (5.25) we further obtain (5.23). In order to estimate \(|\zeta _h(t)|\) we use (2.4). For \(t\in [t_j,t_{j+1}[\), \(j\in J_0^{N-1}\), we have
with some constant \(c_g\) independent of \((x^*_h,u^*_h)\in {{\mathscr {F}}}_{N,\rho }\). By (2.8) we have
and from the discrete adjoint equation we obtain
which implies (5.22). \(\square \)
Theorem 4
Let Assumptions (2.1), (5.1), (5.2), (5.7), and (5.8) be satisfied and suppose that \(u^*\) has bounded variation. Then for each \(0<\rho \le \delta _2\) Problem \(\text{(OC) }_{N,\rho }\) has a (global) solution for sufficiently large N. Further for each such solution \((x^*_h,u^*_h)\) and the associated adjoint function \(\lambda ^*_h\) the estimates
hold with constants \(c_u\), \(c_x\), and \(c_\lambda \) independent of N and the solution \((x^*_h,u^*_h)\).
Proof
Let \(0<\rho \le \delta _2\) be given. By Theorem 3 condition (4.8) is satisfied. Therefore, by Theorem 1 Problem \(\text{(OC) }_{N,\rho }\) has a (global) solution for sufficiently large N. Further for each such solution \((x^*_h,u^*_h)\) the estimates
hold with constants \(c_u\), \(c_x\) independent of N, and by (4.19) we have
with some constant \(c_{\lambda ,2}\) independent of N and \(x^*_h\), \(x^*\), \(u^*_h\), and \(u^*\). For suffiently large N we therefore have
and \((x^*_h,u^*_h)\in {{\mathscr {F}}}_{N,\bar{\rho }}\). By Lemma 6 there exists a function \(z^*_h\in X_1\), such that \((z^*_h,u^*_h)\in {{\mathscr {F}}}\) and
with a constant \(c_z\) independent of N, which implies
for sufficiently large N. As in the proof of Theorem 3, using \(z^*_h(t_0)-x^*(t_0)=0\) and (5.3), we get by Taylor expansion around \((x^*,u^*)\)
where \((z,v)=(1-\tau )(x^*,u^*)+\tau (z^*_h,u^*_h)\) with \(\tau \in \,]0,1[\). By (5.7), (5.8), and Lemma 11 this implies
Similarly we obtain by Taylor expansion around \((z^*_h,u^*_h)\)
where \((z,v)=(1-\tau )(z^*_h,u^*_h)+\tau (x^*,u^*)\) with \(\tau \in \,]0,1[\,\). By Lemma 11 this implies
Combining this estimate with (5.29) we obtain
We define \(z=x^*-z^*_h\). Using integration by parts we obtain for the first term on the right hand side of (5.30)
By Lemma 8 we can write the discrete adjoint Eq. (3.1) in the form (4.17). Using this and the terminal condition \(\lambda ^*_h(t_f) = f_x(x^*_h(t_f))^{\mathsf {T}}\) we further obtain
By (2.4), (4.16), (4.18), and (5.27) this implies
It follows from Lemma 5 that
and
By Lemma 12 there is a function \(\zeta _h:[t_0,t_f]\rightarrow \mathbb {R}^m\) satisfying (5.22) such that the discrete minimum principle can be written in the form (5.23). Therefore, defining
we obtain for the second term on the right hand side of (5.30)
Further by (2.4), (4.16), (5.22), and (5.27) we have
which implies
Finally (5.30) and the estimates (5.32) and (5.33) show that with some constant \({\tilde{c}}_u\) independent of N and the solution \((x^*_h,u^*_h)\),
which immediately implies the estimate for \(\Vert u^*_h-u^*\Vert _1\) in (5.26). The estimate for \(\Vert x^*_h-x^*\Vert _{1,1}\) then follows from (5.27) and (5.31), and the estimate for \(\Vert \lambda ^*_h-\lambda ^*\Vert _{1,1}\) follows from (4.20). \(\square \)
Remark 8
(compare Remark 3) Theorem 4 also assumes that \((x^*_h,u^*_h)\) is a global solution of Problem \(\text{(OC) }_{N,\rho }\) and we have not shown uniqueness of the discrete solutions. The reason is that for the class of control problems considered here condition (5.7) does in general not hold for the discrete control problems. The linear-quadratic control problems considered in Alt et al. [6] are convex optimization problems. Therefore, the solutions of the discrete problems are global solutions. In this case Theorem 4 implies Alt et al. [6, Theorem 14] for \(\kappa =1\). \(\Diamond \)
6 Numerical results
Example 2
We consider the following modification of the rocket car problem discussed in Alt et al. [4, Example 6.1] with a nonlinear and non convex cost functional and a nonlinear state equation:
Here the function g is defined by
and
The Hamiltonian is defined by
and we have
and
Therefore, the condition for the quadratic form in Assumption (5.8) is here equivalent to
for all \((z,v)=(x,u)-(x^*,u^*)\) with \(u\in {{\mathscr {U}}}\), \(z(t_0)=0\) and
for a.a. \(t\in [0,5]\). From the numerical results we have \(x_1^*(5)\approx 1.5475>0\). Therefore, this condition is satisfied for arbitrarily small \(\beta \) if \(\varepsilon \) is sufficiently small.
For \(\varepsilon = \frac{1}{2}\) the optimal control is of bang–bang type with one switching point \(s_1\approx 4.49848\). The discretization errors depicted in Table 1 indicate convergence of order 1 w.r.t. the mesh size h as expected by Theorem 4. Figure 1 shows the computed optimal control and states, and the switching function for \(N=100\). \(\Diamond \)
7 Conclusions
In this paper we derived error estimates for Euler approximation of a class of nonlinear optimal control problems of Mayer type with control appearing linearly. Such estimates were previously known only in case of continuous controls, for linear-quadratic problems affine w.r.t. the control, and for some special classes of control problems with a nonlinear cost functional but a linear or semilinear state equation. The results were obtained under the growth condition (4.8) for the cost functional or under the stronger second-order sufficient optimality condition (5.7) excluding singular arcs of the optimal control (see Remark 5). Felgenhauer [18] shows for scalar bang–bang controls that a second-order optimality condition for the so-called “induced finite-dimensional problem” of optimizing the switching times implies (4.8) with \(\kappa =2\). It is an open question whether (4.8) can be satisfied if the optimal control has singular arcs. But it should be noted that in this case other second-order conditions may be useful (see e.g. Felgenhauer [19], where a second-order condition in connection with the Goh transformation has been used).
A nonlinear control problem may have many local solutions and Example 2 shows that the analytical verification of conditions (4.8) and (5.7) may be difficult. Therefore, another important topic, not treated in this paper, is the numerical verification of such conditions. For the numerical verification of (4.8) in case of bang–bang controls one can use known results. First the control problem is solved by Euler discretization in order to obtain a good approximation for the switching times. If this works, in a second step the induced finite-dimensional problem mentioned above can be solved in order to compute the switching times more accurately. Then the test for the numerical verification of a second-order sufficient optimality condition for the induced finite-dimensional problem discussed in Maurer et al. [29] can be applied. If the test is successful, then the results of Felgenhauer [18] show that (4.8) holds with \(\kappa =2\). While the results of Maurer et al. [29] are stated for general nonlinear control problems, the results of Felgenhauer [18] so far are restricted to scalar control problems. As an alternative, tests based on Riccati differential equations can be used (see Felgenhauer [18, Section 4], Osmolovskii and Maurer [34] and the papers cited therein).
References
Alt, W.: On the approximation of infinite optimization problems with an application to optimal control problems. Appl. Math. Optim. 12, 15–27 (1984)
Alt, W.: Local stability of solutions to differentiable optimization problems in Banach spaces. J. Optim. Theory Appl. 70, 443–466 (1991)
Alt, W.: Discretization and mesh-independence of Newton’s method for generalized equations. In: Fiacco, A.V. (ed.) Mathematical Programming with Data Perturbations V. Lecture Notes in Pure and Applied Mathematics, vol. 195, pp. 1–30. Marcel Dekker, New York (1997)
Alt, W., Baier, R., Gerdts, M., Lempio, F.: Error bounds for Euler approximation of linear-quadratic control problems with bang–bang solutions. Numer. Algebra Control Optim. 2(3), 547–570 (2012)
Alt, W., Mackenroth, U.: Convergence of finite element approximations to state constrained convex parabolic boundary control problems. SIAM J. Control Optim. 27, 718–736 (1989)
Alt, W., Schneider, C., Seydenschwanz, M.: Regularization and implicit Euler discretization of linear-quadratic optimal control problems with bang-bang solutions. Appl. Math. Comput. (2016). https://doi.org/10.1016/j.amc.2016.04.028
Alt, W., Seydenschwanz, M.: An implicit discretization scheme for linear-quadratic control problems with bang–bang solutions. Optim. Methods Softw. 29(3), 535–560 (2014). https://doi.org/10.1080/10556788.2013.821612
Deckelnick, K., Hinze, M.: A note on the approximation of elliptic control problems with bang–bang controls. Comput. Optim. Appl. 51(2), 931–939 (2012)
Dhamo, V., Tröltzsch, F.: Some aspects of reachability for parabolic boundary control problems with control constraints. Comput. Optim. Appl. 50(1), 75–110 (2011)
Dontchev, A.L., Hager, W.W.: Lipschitzian stability in nonlinear control and optimization. SIAM J. Control Optim. 31(3), 569–603 (1993)
Dontchev, A.L., Hager, W.W.: The Euler approximation in state constrained optimal control. Math. Comput. 70, 173–203 (2000)
Dontchev, A.L., Hager, W.W., Malanowski, K.: Error bounds for Euler approximation of a state and control constrained optimal control problem. Numer. Funct. Anal. Optim. 21, 653–682 (2000)
Dontchev, A.L., Hager, W.W., Veliov, V.M.: Second-order Runge–Kutta approximations in control constrained optimal control. SIAM J. Numer. Anal. 38, 202–226 (2000)
Dontchev, A.L., Veliov, V.M.: Metric regularity under approximations. Control Cybern. 38(4B), 1283–1303 (2009)
Felgenhauer, U.: On stability of bang–bang type controls. SIAM J. Control Optim. 41(6), 1843–1867 (2003)
Felgenhauer, U.: Optimality properties of controls with bang–bang components in problems with semilinear state equation. Control Cybern. 34(3), 764–785 (2005)
Felgenhauer, U.: The shooting approach in analyzing bang–bang extremals with simultaneous control switches. Control Cybern. 37(2), 307–327 (2008)
Felgenhauer, U.: Note on local quadratic growth estimates in bang–bang optimal control problems. Optimization 64(3), 521–537 (2015). https://doi.org/10.1080/02331934.2013.773000
Felgenhauer, U.: Discretization of semilinear bang-singular-bang control problems. Comput. Optim. Appl. 64, 295–326 (2016). https://doi.org/10.1007/s10589-015-9800-2
Felgenhauer, U., Poggiolini, L., Stefani, G.: Optimality and stability result for bang–bang optimal controls with simple and double switch behaviour. Control Cybern. 38, 1305–1325 (2009)
Haunschmied, J.L., Pietrus, A., Veliov, V.M.: The Euler method for linear control systems revisited. In: Large-Scale Scientific Computing—9th International Conference, LSSC 2013, Sozopol, Bulgaria, pp. 90–97, 3–7 June 2013. Revised Selected Papers (2013)
Ioffe, A., Tihomirov, V.M.: Theorie der Extremalwertaufgaben. VEB Deutscher Verlag der Wissenschaften, Berlin (1979)
Kaya, C.Y., Lucas, S.K., Simakov, S.T.: Computations for bang–bang constrained optimal control using a mathematical programming formulation. Optim. Control Appl. Methods 25, 295–308 (2004). https://doi.org/10.1002/oca.749
Malanowski, K.: Stability and sensitivity of solutions to optimal control problems for systems with control appearing linearly. Appl. Math. Optim. 16, 73–91 (1987)
Malanowski, K.: Stability and sensitivity of solutions to nonlinear optimal control problems. Appl. Math. Optim. 32, 111–141 (1995)
Malanowski, K.: Stability and sensitivity analysis for optimal control problems. A survey. Trudy Inst. Mat. i Mekh. UrO RAN 16(5), 278–288 (2010)
Malanowski, K., Büskens, C., Maurer, H.: Convergence of approximations to nonlinear optimal control problems. In: Fiacco, A.V. (ed.) Mathematical Programming with Data Perturbations V. Lecture Notes in Pure and Applied Mathematics, vol. 195, pp. 253–284. Marcel Dekker, New York (1997)
Maurer, H.: First- and second order sufficient optimality conditions in mathematimacal programming and optimal control. Math. Program. Study 14, 163–177 (1981)
Maurer, H., Büskens, C., Kim, J.H.R., Kaya, C.Y.: Optimization methods for the verification of second-order sufficient conditions for bang–bang controls. Optim. Control Appl. Methods 26(3), 129–156 (2005)
Maurer, H., Osmolovskii, N.P.: Second order sufficient conditions for time optimal bang–bang controls. SIAM J. Control Optim. 42(6), 2239–2263 (2004)
Maurer, H., Zowe, J.: First- and second-order necessary and sufficient optimality conditions for infinite-dimensional programming problems. Math. Program. 16, 98–110 (1979)
Osmolovskii, N.P., Maurer, H.: Equivalence of second order optimality conditions for bang–bang control problems. Part 1: main results. Control Cybern. 34(3), 927–950 (2005)
Osmolovskii, N.P., Maurer, H.: Equivalence of second order optimality conditions for bang–bang control problems. Part 2: proofs, variational derivatives and representations. Control Cybern. 36(1), 5–45 (2007)
Osmolovskii, N.P., Maurer, H.: Applications to Regular and Bang–Bang Control: Second-Order Necessary and Sufficient Optimality Conditions in Calculus of Variations and Optimal Control. SIAM, Philadelphia (2012)
Pietrus, A., Scarinci, T., Veliov, V.M.: High Order Discrete Approximations to Mayer’s Problems for Linear Systems. Technical Report, TU Wien, ORCOS (2016)
Quincampoix, M., Veliov, V.M.: Metric regularity and stability of optimal control problems for linear systems. SIAM J. Control Optim. 51(5), 4118–4137 (2013)
Robinson, S.M.: Regularity and stability for convex multivalued functions. Math. Oper. Res. 1, 130–143 (1976)
Robinson, S.M.: Stability theory for systems of inequalities, Part II: differentiable nonlinear systems. SIAM J. Numer. Anal. 13(4), 497–513 (1976)
Sendov, B., Popov, V.A.: The averaged Moduli of Smoothness. Wiley, Chichester (1988)
Seydenschwanz, M.: Convergence results for the discrete regularization of linear-quadratic control problems with bang–bang solutions. Comput. Optim. Appl. 61, 731–760 (2015). https://doi.org/10.1007/s10589-015-9730-z
Veliov, V.M.: Error analysis of discrete approximations to bang–bang optimal control problems: the linear case. Control Cybern. 34(3), 967–982 (2005)
Acknowledgements
The authors would like to thank the anonymous referees for their careful reading of the manuscript and their constructive and valuable suggestions and comments.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Alt, W., Felgenhauer, U. & Seydenschwanz, M. Euler discretization for a class of nonlinear optimal control problems with control appearing linearly. Comput Optim Appl 69, 825–856 (2018). https://doi.org/10.1007/s10589-017-9969-7
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10589-017-9969-7