Abstract
In this paper, we prove a local null controllability result for the three-dimensional Navier–Stokes equations on a (smooth) bounded domain of \({\mathbb R}^3\) with null Dirichlet boundary conditions. The control is distributed in an arbitrarily small nonempty open subset and has two vanishing components. Lions and Zuazua proved that the linearized system is not necessarily null controllable even if the control is distributed on the entire domain, hence the standard linearization method fails. We use the return method together with a new algebraic method inspired by the works of Gromov and previous results by Gueye.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
1.1 Notations and statement of the theorem
Let \(T>0\), let \(\varOmega \) be a nonempty bounded domain of \({\mathbb {R}}^3\) of class \(C^\infty \) and let \(\omega \) be a nonempty open subset of \(\varOmega \). We define \(Q\subset {\mathbb {R}}\times {\mathbb {R}}^3\) by
and we call
The current point \(x\in {\mathbb {R}}^3\) is \(x=(x_1,x_2,x_3)\). The \(i\)-th component of a vector (or a vector field) \(f\) is denoted \(f^i\). The control is \(u=(u^1,u^2,u^3)\in L^2 (Q)^3\). We require that the support of \(u\) is included in \(\omega \), which is our control domain. We impose that two components of \(u\) vanish, for example the first two:
so that \(u\) will be written under the form \((0,0,1_\omega v)\) with \(v\in L^2(Q)\) from now on, where \(1_\omega : \varOmega \rightarrow {\mathbb {R}}\) is the characteristic function of \(\omega \):
Let us define
The space \(V\) is equipped with the \(H^1_0\)-norm. Let us denote by \(H\) the closure of \(V\) in \(L^2(\varOmega )^3\). The space \(H\) is equipped with the \(L^2\)-norm.
We are interested in the following Navier–Stokes control system:
From [28, Theorem 3.1, p. 282], we have the following existence result: For every \(y^0\in H\), there exist \(y\in L^2((0,T),V)\cap L^\infty ((0,T),H)\) and \(p\in L^2(Q)\) satisfying
such that (1.2) holds.
Our main result is the following theorem, which expresses the small-time local null-controllability of (1.2):
Theorem 1
For every \(T>0\) and for every \(r >0\), there exists \(\eta >0\) such that, for every \(y^0\in V\) verifying \(||y^0||_{H_0^1(\varOmega )^3}\leqslant \eta \), there exist \(v\in L^2(Q)\) and a solution \((y,p)\in L^2((0,T),H^2(\varOmega )^3\cap V)\cap L^\infty ((0,T),H^1(\varOmega )^3\cap V)\times L^2(Q)\) of (1.2)–(1.3) such that
Remark 1
Once a control \(v\in L^2(Q)\) is given, the corresponding solution \((y,p)\) of (1.2), (1.3) and (1.5) given by Theorem 1 is unique (recall that for the Navier–Stokes system, the uniqueness of \((y,p)\) means that \(y\) is unique and \(p\) is unique up to a constant depending on the time). This comes from the uniqueness result given in [28, Theorem 3.4, p. 297]: One has
thanks to a classical Sobolev embedding, and there is at most one solution \((y,p)\) of (1.2) and (1.3) in the space
Remark 2
One observes that in Theorem 1 the initial condition \(y^0\) is more regular than usual (\(y^0\in H\)). In fact, using the same arguments as in [15] and [16] (see also [6, Remark 1]), one can easily extend the previous theorem to small initial data in \(H\cap L^4(\varOmega )^3\) with a solution \((y,p)\in L^2((0,T),V)\cap L^\infty ((0,T),H)\times L^2(Q)\). In this case, Remark 1 is no longer true and there might possibly exist many solutions \((y,p)\) verifying (1.2) and (1.3) once \(v\) is given.
1.2 Some previous results
The controllability of the two or three-dimensional Navier–Stokes equations with a distributed control has been studied in numerous papers. In general, for Navier–Stokes equations, it is relevant to consider the approximate controllability, the null controllability or the exact controllability to the trajectories, the second one being a particular case of the third one.
In [22], a first result of local exact controllability to the trajectories was established under technical conditions: \(\varOmega \) had to be homeomorphic to a ball, the control had to be supported in a nonempty open subset whose closure is included in \(\varOmega \), and the target trajectory had to be a stationary solution of the Navier–Stokes equation. Moreover, there were some technical regularity conditions for these stationary solutions. A similar result for the linearized Navier–Stokes equations was established but with the same strong conditions. Many of these hypotheses were removed in [23].
Then, it was proved in [15] the local exact controllability to the trajectories with regularity conditions that were weaker and more suitable for the study of the Navier–Stokes equations. In this article, the authors also proved some exact controllability results for linearized Navier–Stokes systems, with very weak regularity conditions. The same authors proved in [16] the local exact controllability to the trajectories with a control having one vanishing component, provided that \(\omega \) “touches” the boundary of the domain \(\varOmega \) in some sense. Later on it was proved in [13] a local null controllability result for the Stokes system with a control having a vanishing component without the geometrical condition on \(\omega \), but the authors were not able to extend it to the nonlinear Navier–Stokes system. A recent work [6] improved the previous one and proved the local null controllability of the Stokes system with an additional source member by means of a control having a vanishing component, which enabled the authors to prove the local null controllability of the Navier–Stokes system for a control having a vanishing component. In all these articles, the main points of the proof were to establish first the controllability of the linearized control system around the target trajectory thanks to Carleman estimates on the adjoint of the linearized equation, and then to use an inverse mapping theorem or a fixed-point theorem to deal with the nonlinear system.
The natural question is then: Can we remove another component of the control, which would be an optimal result with respect to the number of controls? Reducing the number of components of the control is important for applications, and have already been studied many times for linear or parabolic systems of second order (that are quite similar to linearized Navier–Stokes systems), see for example [4], where a necessary and sufficient condition to control a system of coupled parabolic equations with constant coefficients and with less controls than equations is given, or [3, 26] for time-dependent coefficients. If the coefficients depend on the time and the space, there are no general results, in particular if we consider two coupled parabolic systems where the coupling region and the control region do not intersect (a partial result under the Geometric Control Condition is given in [1]). For a recent survey on the controllability of coupled linear parabolic equations, see [5].
1.3 The linear test
To obtain Theorem 1, the first natural idea is to linearize the system around 0, i.e. to consider the Stokes control system
It is well-known (see for example [16], or [23]) that if this linear system were null controllable (with, in addition, an arbitrary source term in a suitable space), then applying an inverse mapping theorem (for example the one presented in [2]) in some relevant weighted spaces, we would obtain that (1.2) is locally null controllable around 0. However, the linear control system (1.7) is in general not null controllable and not approximately controllable: In [25], it is proved that this is for example the case if \(\varOmega \) is a cylinder with a circular generating set and with an axis parallel to \(e_3\), even if we control on the entire cylinder (the approximate controllability property holds “generically” with respect to the generating set of the cylinder as explained in [25] though).
Since linearizing around 0 is not relevant, we are going to use the return method, which consists in linearizing system (1.2) around a particular trajectory \((\overline{y},\overline{p},\overline{u})\) (that we construct explicitly) verifying \(\overline{y}(0,\cdot )=\overline{y}(T,\cdot )= 0\), proving that the linearized system (with a source term \(f\) verifying an exponential decrease condition at time \(t=T\)) is null controllable, and then concluding by a usual inverse mapping argument that our system is locally null controllable. This method was introduced in [7] for a stabilization problem concerning nonlinear ordinary differential equations and first used in the context of partial differential equations in [8]. The return method was already successfully used in [9, 11, 18] to obtain global controllability results for the Navier–Stokes equations and in [12] to prove the local null controllability for the Navier–Stokes equations on the torus \({\mathbb {T}}_2\) when the control has one vanishing component. For more explanations about the return method and other examples of applications, see [10, Chapter 6].
1.4 Structure of the article and sketch of the proof of Theorem 1
The paper is organized as follows.
-
In Sect. 2, according to what was explained at the end of Sect. 1.3, we construct a family of explicit particular trajectories \((\overline{y},\overline{p},\overline{u})\) of the controlled Navier–Stokes system (1.2) going from 0 at time \(t=0\) to 0 at time \(t=T\). These trajectories are compactly supported in \([T/4,T]\times \omega \) and vanish exponentially at time \(t=T\). Moreover, they are polynomials in space on some subcylinder of \(\omega \) denoted \({\mathcal C}_2\), and they can be arbitrarily small. We then linearize (1.2) around \((\overline{y},\overline{p},\overline{u})\) and study the linearized equation
$$\begin{aligned} \left\{ \begin{array}{l@{\quad }l} y^1_t-\Delta y^1+(\overline{y}\cdot \nabla )y^1 +(y\cdot \nabla ){\overline{y}}^1+ \partial _{x_1}p=f^1&{} \text{ in } \ Q, \\ y^2_t-\Delta y^2+(\overline{y}\cdot \nabla )y^2 +(y\cdot \nabla ){\overline{y}}^2+ \partial _{x_2}p=f^2&{} \text{ in } \ Q, \\ y^3_t-\Delta y^3+(\overline{y}\cdot \nabla )y^3 +(y\cdot \nabla ){\overline{y}}^3+ \partial _{x_3}p=1_\omega v+f^3&{} \text{ in } \ Q, \\ \nabla \cdot y=0&{} \text{ in } \ Q, \\ y= 0&{} \text{ on } \ \varSigma , \end{array}\right. \nonumber \\ \end{aligned}$$(1.8)where \(f\) is some source term in an appropriate space.
-
Section 3 is devoted to proving that (1.8) is indeed null controllable (Proposition 5). Section 3.1 is dedicated to introducing some useful notations and the crucial Proposition 1. This proposition explains that we can split up our proof of the null controllability of the linearized equations with a scalar control into two parts:
-
Firstly, we control the following linearized Navier–Stokes system:
$$\begin{aligned} \left\{ \begin{array}{l@{\quad }l} y^{*1}_t\!-\!\Delta y^{*1}\!+\!(\overline{y}\cdot \nabla )y^{*1} \!+\!(y^*\cdot \nabla ){\overline{y}}^1\!+\! \partial _{x_1}p^*={\mathcal {B}}^1u^*\!+\!f^1&{} \text{ in } Q, \\ y^{*2}_t\!-\!\Delta y^{*2}\!+\!(\overline{y}\cdot \nabla )y^{*2} \!+\!(y^*\cdot \nabla ){\overline{y}}^2\!+\! \partial _{x_2}p^*\!=\!{\mathcal {B}}^2u^*\!+\!f^2&{} \text{ in } Q, \\ y^{*3}_t\!-\!\Delta y^{*3}\!+\!(\overline{y}\cdot \nabla )y^{*3} \!+\!(y^*\cdot \nabla ){\overline{y}}^3\!+\! \partial _{x_3}p^*={\mathcal {B}}^3u^*+f^3&{} \text{ in } Q, \\ \nabla \cdot y^*=0&{} \text{ in } Q, \\ y^*= 0&{} \text{ on } \varSigma , \end{array}\right. \nonumber \\ \end{aligned}$$(1.9)where \({\mathcal B}\) is some suitable local control operator that acts on each equation. This is the purpose of Sect. 3.3, where the controllability of System (1.9) is proved thanks to the usual HUM method. More precisely, we prove an appropriate Carleman estimate with observation \({\mathcal B}^*\) on the adjoint equation of (1.9) (Lemma 3), so that we create controls in the image of \({\mathcal B}\) thanks to the study of an appropriate Lax–Milgram type problem, which also enables us to obtain controls that are very regular in the sense that they are in weighted Sobolev spaces of high order in space and time (Proposition 4). Let \((y^*,p^*,u^*)\) be a trajectory of (1.9) that brings the initial condition \(y^0\) to 0 at time \(T\), with a very regular \(u^*\) compactly supported in space at each time in some open subset \(\omega _0\) of \({\mathcal C}_2\) to be chosen later, and that decreases exponentially at time \(t=T\). We emphasize that \((y^*,p^*)\) is less regular than \({\mathcal B}u^*\) (however, it is in some weighted Sobolev space of small order) because the source term \(f\) is not as regular as \({\mathcal B}u^*\).
-
Secondly, we study in Sect. 3.2 the following system locally on \(Q_0:=[T/2,T]\times \omega _0\):
$$\begin{aligned} \left\{ \begin{array}{l@{\quad }l} \tilde{y}^1_t-\Delta \tilde{y}^1+(\overline{y}\cdot \nabla )\tilde{y}^1 +(\tilde{y}\cdot \nabla ){\overline{y}}^1+ \partial _{x_1}\tilde{p}=-{\mathcal {B}}^1u^*&{} \text{ in } Q_0, \\ \tilde{y}^2_t-\Delta \tilde{y}^2+(\overline{y}\cdot \nabla )\tilde{y}^2 +(\tilde{y}\cdot \nabla ){\overline{y}}^2+ \partial _{x_2}\tilde{p}=-{\mathcal {B}}^2u^*&{} \text{ in } Q_0, \\ \tilde{y}^3_t-\Delta \tilde{y}^3+(\overline{y}\cdot \nabla )\tilde{y}^3 +(\tilde{y}\cdot \nabla ){\overline{y}}^3+ \partial _{x_3}\tilde{p}=-{\mathcal {B}}^3u^*+\tilde{v}&{} \text{ in } Q_0, \\ \nabla \cdot \tilde{y}=0&{} \text{ in } Q_0, \end{array}\right. \nonumber \\ \end{aligned}$$(1.10)where \(u^*\) has been introduced above, and where the unknowns are \((\tilde{y},\tilde{p},\tilde{v})\). We want to prove that there exists a solution \((\tilde{y},\tilde{p},\tilde{v})\) of (1.10) (extended by 0 on \([T/2,T]\times \varOmega \)) which has the same support as \(u^*\). This seems reasonable because System (1.10) is analytically underdetermined: we have five unknowns (the 3 components of \(\tilde{y}\), the pressure \(\tilde{p}\) and the scalar control \(\tilde{v}\)) and only four equations. In fact, we prove in Proposition 2 that is possible to find such a \((\tilde{y},\tilde{p},\tilde{v})\) which can moreover be expressed as a linear combination of \(u^*\) and some of its derivatives up to a certain order. This explains why we need \(u^*\) to be very regular. Since \(u^*\) decreases exponentially at time \(T\), this is also the case for \((\tilde{y},\tilde{p},\tilde{v})\). The main idea behind the proof of the existence of such a \((\tilde{y},\tilde{p},\tilde{v})\) is to consider the adjoint system of (1.10) and to differentiate the equations appearing in this system until we get more equations than “unknowns”, the “unknowns” being there the functions and all their derivatives appearing in the equations of the adjoint system. Since Sect. 3.2 is the most innovative, important, and difficult part of the article, we give some further details.
-
1.
In Sect. 3.2.1, we make a choice for operator \({\mathcal B}\) and we prove that the existence of \((\tilde{y},\tilde{p},\tilde{v})\) can be reduced to proving the following property: There exists some \(\omega _0\) and a linear partial differential operator \({\mathcal {N}}:C^{\infty }(Q_0)^4 \rightarrow C^{\infty }(Q_0)^6\) such that for every \(\varphi =(\varphi ^1,\varphi ^2,\varphi ^3,\varphi ^4)\in C^{\infty }(Q_0)^4\), if \((z^1,z^2,\pi )\in C^{\infty }(Q_0)^3\) is a solution of
$$\begin{aligned} \left\{ \begin{array}{l@{\quad }l} -2\partial _{x_3}\overline{y}^1 \partial _{x_1}z^1-\partial _{x_3}\overline{y}^2\partial _{x_2}z^1+(\partial _{x_1}\overline{y}^1-\partial _{x_3}\overline{y}^3)\partial _{x_3}z^1-\overline{y}^1 \partial ^2_{x_1x_3}z^1\\ -\overline{y}^2\partial ^2_{x_2x_3}z^1 -\overline{y}^3\partial ^2_{x_3x_3}z^1-\partial ^2_{x_3t}z^1-\Delta \partial _{x_3}z^1 -\partial _{x_3}\overline{y}^2\partial _{x_1}z^2 \\ +\partial _{x_1} \overline{y}^2\partial _{x_3}z^2=\partial _{x_3}\varphi ^1-\partial _{x_1}\varphi ^3,\\ -\partial _{x_3}\overline{y}^1\partial _{x_2}z^1\!+\!\partial _{x_2}\overline{y}^1 \partial _{x_3}z^1\!-\!\partial _{x_3}\overline{y}^1\partial _{x_1}z^2\!-\!\overline{y}^1\partial ^2_{x_1x_3}z^2\!-\!2\partial _{x_3}\overline{y}^2\partial _{x_2} z^2\\ -\overline{y}^2 \partial ^2_{x_2x_3}z^2\!+v(\partial _{x_2}\overline{y}^2\!-\!\partial _{x_3}\overline{y}^3)\partial _{x_3}z^2 \!-\!\overline{y}^3\partial ^2_{x_3x_3}z^2\!-\!\partial ^2_{x_3t}z^2\!-\!\Delta \partial _{x_3}z^2\\ =\partial _{x_3}\varphi ^2-\partial _{x_2}\varphi ^3, \\ -\partial _{x_1}z^1-\partial _{x_2}z^2 =\varphi ^4. \end{array}\right. \nonumber \\ \end{aligned}$$(1.11)then \((-\partial _{x_1}z^1,-\partial _{x_2}z^1,-\partial _{x_3}z^1,-\partial _{x_1}z^2,-\partial _{x_2}z^2,-\partial _{x_3}z^2)={\mathcal {N}} \varphi \).
-
2.
In Sect. 3.2.2 we study the overdetermined system (1.11). If we consider \(z^1\), \(z^2\), and all their derivatives at every order as independent algebraic unknowns (i.e. we forget that \(\partial _{x_1}z^1,\ldots \) are derivatives of \(z^1\) and consider them as unknowns of System (1.11)), we obtain a system of 3 equations with \(20\) unknowns. However, we can prove that if we differentiate the equations of System (1.11) enough times, one can obtain more equations than unknowns. In particular, if the two first equations of (1.11) are differentiated 19 times and if the last equation of (1.11) is differentiated 21 times, then we get 30360 equations and 29900 unknowns. We can write the big system describing these equations as follows:
$$\begin{aligned} L_0(t,x)Z= \varPhi , \end{aligned}$$where \(L_0\in C^\infty (Q_0;{\mathcal {M}}_{30360\times 29900}({\mathbb {R}}))\), \(Z\in {\mathbb {R}}^{29900}\) contains the derivatives of \(z^1\) and \(z^2\) up to the order \(22\) and \(\varPhi \in {\mathbb {R}}^{30360}\) contains the derivatives of \(\varphi \) up to the order \(21\). If we are able to find a suitable submatrix of \(L_0\) (denoted \(P\)) that is invertible, then roughly the matrix \(P^{-1}\) (seen as a differential operator) will be a good candidate for \({\mathcal N}\).
-
3.
In Sect. 3.2.3, we describe how we created a program that enables us to differentiate the equations of system (1.11) and that finds a proper matrix \(P\). Of course it cannot be done by hand, we have to use a computer. Let us point out that in the computer part of the proof, we only use symbolic computations, so that no approximations are made by the computer. We first prove (cf. Lemma 2) that it is enough to find a suitable matrix \(P\) which is invertible at some precise point \(\xi ^0\), i.e. it is enough to consider \(L_0(\xi ^0)\) for some well-chosen \(\xi ^0\). We explain in Lemma 1 how we found \(P(\xi ^0)\) thanks to a suitable reordering of matrix \(L_0(\xi ^0)\) given by the Dulmage–Mendelsohn decomposition of \(L_0(\xi ^0)\).
-
1.
-
We now remark that \((y^*+\tilde{y},p^*+\tilde{p},\tilde{v})\) is a trajectory of (1.8) (see (1.9) and (1.10)) that brings the initial condition \(y^0\) to 0 at time \(T\). We then prove in Sect. 3.4 that \((y^*+\tilde{y},p^*+\tilde{p}, \tilde{v})\) is in some appropriate weighted Sobolev space (Proposition 5).
-
To conclude, in Sect. 4, we explain how the suitable functional setting we obtained for the solutions \((y,p,v)\) of System (1.8) enables us to go back to the local null controllability of (1.2) thanks to a usual argument of inverse mapping theorem.
2 Constructing a relevant trajectory
In this subsection, we construct explicit particular trajectories \((\overline{y},\overline{p},\overline{u})\) going from 0 to 0 so that, as it will be shown in Sect. 3, the linearized control system around them is null controllable.
Without loss of generality we may assume that \(0\in \omega \). Let \(g\in C^\infty ({\mathbb {R}}^3)\),
and \(h\in C^\infty ({\mathbb {R}}^3)\),
For \((x_1,x_2,x_3)\in {\mathbb {R}}^3\), let \(r:=\sqrt{x_1^2+x_2^2}\). We define \(\overline{y} \in C^\infty ({\mathbb {R}}^4;{\mathbb {R}}^3)\) by
Let \(r_1>0\) be small enough so that
On the functions \(g\) and \(h\), we also require that
In (2.3), (2.3) and in the following, \(\text {Supp}(f)\) denotes the support of the function \(f\). From (2.1), (2.2), (2.3) and (2.3), one obtains
which implies in particular that \(\overline{y}\) has null trace on \(\varSigma \). Let \(\widehat{p} \in C^\infty ({\mathbb {R}}^3)\) be defined by
Let \( \overline{p} \in C^\infty ({\mathbb {R}}^4)\) be defined by
From (2.2), (2.3), (2.3), (2.6) and (2.7), it follows that
From (2.1), (2.6) and (2.7), one obtains
Let \(\overline{u} \in C^\infty ({\mathbb {R}}^4)^3\) be defined by
From (2.11), one obtains (1.1). From (2.5), (2.8) and (2.11), we have
From (2.9), (2.10), (2.11) and (2.12), we have
Finally, in order to have
it suffices to impose
Let \(\nu \) be a positive numerical constant which will be chosen later. Let \(a\in C^\infty ({\mathbb {R}})\), \(b\in C^\infty ({\mathbb {R}})\) and \(c\in C^\infty ({\mathbb {R}})\) be such that
We then set
and
where \(\epsilon >0\) (which will be chosen small enough later). From (2.19) and (2.20), one obtains (2.15).
In the next section, we prove that, for every small enough \(T\), for every small enough \(\varepsilon >0\) and for a well-chosen \(\nu \), the linearized control system around the trajectory \((\overline{y},\overline{p},\overline{u})\) is controllable.
3 A controllability result on the linearized system
3.1 Definitions and notations
The linearized control system around the trajectory \((\overline{y}, \overline{p}, \overline{u})\) is the linear control system
where the state is \(y:Q\rightarrow {\mathbb {R}}^3, f:Q\rightarrow {\mathbb {R}}^3\) is a source term (it will be specified later in which space exactly it shall be) and the control is \(v:Q\rightarrow {\mathbb {R}}\). In all what follows, in order to lighten the notations, we will write \(\overline{y}\) as a function of \(t\) and \(x\) only, but one has to remember that \(\overline{y}\) also depends on \(\varepsilon \) and \(\nu \). Let \(\omega _0\) be a nonempty open subset of
which will be chosen more precisely in the next section. Let \(Q_0:=(T/2,T)\times \omega _0.\) The following figure summarizes the different roles of each open subset of \(\varOmega \) we introduced up to now (Fig. 1).
Let \({\mathcal {L}}: C^{\infty }(Q_0)^5\rightarrow C^{\infty }(Q_0)^4\) be defined by
for every \(y=(y^1,y^2,y^3)\in C^{\infty }(Q_0)^3\), for every \(p \in C^{\infty }(Q_0) \) and for every \(v\in C^{\infty }(Q_0)\). Let us denote by
the current point in \(Q_0\). For \(\alpha =(\alpha _0, \alpha _1,\alpha _2,\alpha _3)\in {\mathbb {N}}^4\) and \(\varphi :Q_0\rightarrow {\mathbb {R}}^k, \partial ^\alpha \varphi \), denotes, as usual,
Let \({\mathcal {L}}({\mathbb {R}}^k;{\mathbb {R}}^l)\) be the set of linear maps from \({\mathbb {R}}^k\) into \({\mathbb {R}}^l\) and \({\mathcal {M}}_{k,l}({\mathcal R})\) be the set of matrices of size \(k\times l\) with values in the ring \({\mathcal R}\).
As usual, in the inequalities written in this article \(C\) denotes a constant (depending in general only on \(\omega \), \(\varOmega \), \(T\)) that may change from one line to another.
Let us give some other definitions.
Definition 1
A linear map \({\mathcal {M}}: C^{\infty }(Q_0)^k\rightarrow C^{\infty }(Q_0)^l\) is called a linear partial differential operator of order \(m\) if, for every \(\alpha = (\alpha _0, \alpha _1,\alpha _2,\alpha _3)\in {\mathbb {N}}^4\) with \(|\alpha |:=\alpha _0+\alpha _1+\alpha _2+\alpha _3\leqslant m\), there exists \(A_\alpha \in C^\infty (Q_0; {\mathcal {L}}({\mathbb {R}}^k;{\mathbb {R}}^l))\) such that
A linear map \({\mathcal {M}}:C^{\infty }(Q_0)^k\rightarrow C^{\infty }(Q_0)^l\) is called a linear partial differential operator if there exists \(m\in {\mathbb {N}}\) such that \({\mathcal {M}}\) is a linear partial differential operator of order \(m\).
Let \(k\) be a positive integer and let \({\mathcal {B}}:=({\mathcal {B}}^1,{\mathcal {B}}^2,{\mathcal {B}}^3): C^{\infty }(Q_0)^k\rightarrow C^{\infty }(Q_0)^3\) be a linear partial differential operator. Let us consider the linear equation
where the data is \(u\in C^{\infty }(Q_0)^k\) and the unknown is \((y,p,v)\in C^{\infty }(Q_0)^5\). Following [19, p. 148], we adopt the following definition.
Definition 2
The linear equation (3.4) is algebraically solvable if there exists a linear partial differential operator \({\mathcal {M}}:C^{\infty }(Q_0)^k\rightarrow C^{\infty }(Q_0)^5\) such that, for every \(u\in C^{\infty }(Q_0)^k\), \({\mathcal {M}}u\) is a solution of (3.4), i.e. such that
In the following, every function \(\varphi \in C^{\infty }(Q_0)^l\) with a compact support included in \(Q_0\) is extended by 0 in \(Q{\setminus }Q_0\) and we still denote this extension by \(\varphi \).
The next proposition explains how the notion of “algebraic solvability” can be useful to reduce the number of controls as soon as a controllability result is already known for a large number of controls. In fact, the question of the null-controllability of (3.1) can be split up into two distinct problems: One “algebraic” part (solving system (3.5)) and one “analytic” part (finding controls which are in the image of \({\mathcal B}\), the control acting possibly on all the equations and not only on the third one). This proposition has a very general scope and could be formulated for more general control systems. It is inspired by techniques used in the control of ordinary differential equations (see, in particular, [10, Chapter 1, pages 13–15]).
Proposition 1
Let us consider the linear control system
where the state is \(y:Q\rightarrow {\mathbb {R}}^3\), the control is \(u\in C^{\infty }(Q)^k\), which is required to have a support in \(Q_0\), and \(f:=(f^1,f^2,f^3)\in C^\infty (Q)^3\) is a source term. Let us assume that:
-
\({\mathcal {A}}_1\) The linear control system (3.6) is null controllable during the interval of time \([0,T]\) in the sense that for every \(y^0\in V\) and for every \(f\in C^\infty (Q)\) such that
$$\begin{aligned} \text {there exists } \delta >0 \hbox { such that } f=0 \quad \hbox { on } [T-\delta ,T]\times \varOmega , \end{aligned}$$(3.7)there exists \(u\in C^{\infty }(Q)^k\) with a compact support included in \(Q_0\) such that the solution \((y,p)\) of (3.6) with initial condition \(y(0,\cdot )=y^0\) satisfies \(y(T,\cdot )=0\).
-
\({\mathcal {A}}_2\)
$$\begin{aligned} (3.4) \hbox { is algebraically solvable.} \end{aligned}$$
Then, the linear control system (3.1) is null controllable during the interval of time \([0,T]\): For every \(y^0\in V\) and for every \(f\in C^\infty (Q)\) satisfying (3.7), there exists \(v\in C^{\infty }(Q)\) with a compact support included in \(Q_0\) such that the solution \((y,p)\) of (3.1) with initial condition \(y(0,\cdot )=y^0\) satisfies \(y(T,\cdot )=0\).
Proof of Proposition 1
First of all, we use the null-controllability of (3.6) with controls in the image of \({\mathcal B}\) (Assumption \({\mathcal {A}}_1\)) with source term \(f\): let \(y^0\in V\) and let \(u^*\in C^{\infty }(Q_0)^k\) with compact supported included in some open subset \(Q^*\subset \subset Q_0\) included in \(Q_0\) such that the solution \((y^*,p^*)\) of the following equation:
with initial condition \(y^*(0,\cdot )=y^0\) satisfies \(y^*(T,\cdot )=0\). Let us remark that \({\mathcal B}\) is a local operator, which implies that \({\mathcal {B}}u^*_{|Q_0}\) still has a compact support included in \(Q^*\). Now we use the algebraic solvability of (3.4) (Assumption \({\mathcal {A}}_2\)). Let \({\mathcal {M}}\) be as in Definition 2. For a map \(h\in C^\infty (Q)^k\) with a support included in \(Q_0\), we denote by \({\mathcal {M}}\) the map from \(Q\) into \({\mathbb {R}}^5\) defined by
We shall use this slight abuse of notation until the end of the paper. Note that, for every \(h\in C^\infty (Q)^k\) with a support included in \(Q_0, {\mathcal {M}}h \in C^\infty (Q)^5\) and has a support included in \(Q_0\) (because \({\mathcal M}\) is a local operator). Let us call
so that \((\tilde{y},\tilde{p},\tilde{v})\) verifies the following linearized Navier–Stokes equation:
One observes that the support of \((\tilde{y},\tilde{p}, \tilde{v})\) is still included in \(Q^*\) (which is strongly included in \(Q_0\)). In particular \(\tilde{y}(0,\cdot )=0\) and \(\tilde{y}(T,\cdot )= 0\). Let
Note that \((y,p)\) is different from \((y^*,p^*)\) only on \(Q^*\). In particular one has \(y(0,\cdot )=y^0\) and \(y(T,\cdot )=0\). Moreover, from (3.8) and (3.9), we obtain that \((y,p,v)\) verifies the equation
which shows that the linear control system (3.1) is indeed null controllable during the interval of time \([0,T]\) and concludes the proof of Proposition 1. \(\square \)
Remark 3
For the sake of simplicity, we have formulated Proposition 1 in a \(C^\infty \) setting. Let us assume that the control \(u\) coming from Assumption \({\mathcal {A}}_1\) is not of class \(C^\infty \), but is less regular (one sees that the regularities of \(y^*\), \(p^*\) and \(f\) does not matter for the proof of Proposition 1 since only \(u^*\) is differentiated by the linear partial differential operator \({\mathcal M}\)). For example, assume that \(u^*\in H_1\) where \(H_1\) is a functional space (for example a weighted Sobolev space), and assume that \({\mathcal M}\) can be extended on \(H_1\), \({\mathcal M} u^*\) being then in another functional space \(H_2\) (for example another weighted Sobolev space of order less that \(H_1\) in order to take into account that \({\mathcal M}\) is a linear partial differential operator). Then one easily verifies that Proposition 1 remains true as soon as every function of \(H_1\) (and its derivatives until the order at least the order of \({\mathcal M}\)) vanishes at time \(t=T\), the first Assumption \({\mathcal {A}}_1\) being changed as the following: the linear control system (3.6) is null controllable during the interval of time \([0,T]\), i.e. for every \(y^0\in V\) and for every \(f\in L^2(Q)\) satisfying (3.7) there exists \(u\in H_1\) with support included in \(Q_0\) such that the solution \((y,p)\) of (3.6) satisfying the initial condition \(y(0,\cdot )=y^0\) satisfies \(y(T,\cdot )=0\). Note that the scalar control \(v\) is now only in \(H_2\). Similarly, we will need to relax property (3.7) by replacing it with a suitable decay rate near \(t=T\). This will be detailed in Sect. 3.4.
It remains to deal, for a suitable choice of \({\mathcal {B}}\), with Assumption \({\mathcal {A}}_2\) (we shall do it in Sect. 3.2) and with Assumption \({\mathcal {A}}_1\), i.e. with the null controllability of the linear control system (3.6) in suitable spaces (we shall do it in Sect. 3.3).
3.2 Algebraic solvability of (3.4)
We choose \(k=7\) and define \({\mathcal {B}}\) by
The main result of this subsection is the following proposition.
Proposition 2
There exists \(\varepsilon ^*>0\), there exists \(T^*>0\) such that, for every \(\varepsilon \in (0,\varepsilon ^*)\), there exists a nonempty open subset \(\omega _0\) of \({\mathcal C}_2\) such that Assumption \({\mathcal {A}}_2\) holds for every \(T<T^*\): There exists a linear partial differential operator \({\mathcal {M}}:C^{\infty }(Q_0)^7\rightarrow C^{\infty }(Q_0)^5\) such that (3.5) holds.
3.2.1 The adjoint problem
Let \({\mathcal {L}}_0: C^{\infty }(Q_0)^4\rightarrow C^{\infty }(Q_0)^3\) be the linear partial differential operator defined by
for every \(y=(y^1,y^2,y^3)\in C^{\infty }(Q_0)^3\), and every \(p\in C^{\infty }(Q_0)\).
Let \({\mathcal {B}}_0:C^{\infty }(Q_0)^6\rightarrow C^{\infty }(Q_0)^3\) be the linear partial differential operator defined by
Note that the third equation of (3.4) can be read as
Hence, one easily sees that Assumption \({\mathcal {A}}_2\) is equivalent to the existence of a linear partial differential operator \({\mathcal {M}}_0: C^{\infty }(Q_0)^6\rightarrow C^{\infty }(Q_0)^4\) such that
As in [19, p. 157], we study (3.13) by looking at the “adjoint equation”. For every linear partial differential operator \({\mathcal {M}}: C^{\infty }(Q_0)^k\rightarrow C^{\infty }(Q_0)^l\), \({\mathcal {M}}=\sum _{|\alpha |\leqslant m} A_\alpha \partial ^\alpha \), we associate its (formal) adjoint
defined by
where \(A_\alpha ^T (\xi )\) is the transpose of the matrix \(A_\alpha (\xi )\). Definition (3.14) makes sense since
implies that the \(A_\alpha \) are all equal to 0.) One has \({\mathcal {M}}^{**}={\mathcal {M}}\) and, if \({\mathcal {M}}: C^{\infty }(Q_0)^k\rightarrow C^{\infty }(Q_0)^l\) and \({\mathcal {N}}: C^{\infty }(Q_0)^l\rightarrow C^{\infty }(Q_0)^m\) are two linear partial differential operators, then \( ({\mathcal {N}} \circ {\mathcal {M}})^*={\mathcal {M}}^*\circ {\mathcal {N}}^*\).
Hence, (3.13) is equivalent to
Direct computations, together with (2.14), show that, for every \(z=(z^1,z^2)\in C^{\infty }(Q_0)^2\) and for every \(\pi \in C^{\infty }(Q_0)\),
Assumption \({\mathcal {A}}_2\) is now equivalent to the following property: There exists a linear partial differential operator \({\mathcal {N}} (={\mathcal {M}}_0^*):C^{\infty }(Q_0)^4 \rightarrow C^{\infty }(Q_0)^6\) such that for every \(\varphi =(\varphi ^1,\varphi ^2,\varphi ^3,\varphi ^4)\in C^{\infty }(Q_0)^4\), if \((z^1,z^2,\pi )\in C^{\infty }(Q_0)^3\) is a solution of
then \((-\partial _{x_1}z^1,-\partial _{x_2}z^1,-\partial _{x_3}z^1,-\partial _{x_1}z^2,-\partial _{x_2}z^2,-\partial _{x_3}z^2)={\mathcal {N}} \varphi \).
Remark 4
The most natural linear partial differential operator \({\mathcal {B}}\) to try first would have been \({\mathcal {B}} : C^{\infty }(Q_0)^3\rightarrow C^{\infty }(Q_0)^3\) defined by
Unfortunately, Proposition 2 does not hold with this \({\mathcal {B}}\). Indeed, in this case \({\mathcal {B}}_0^*: C^{\infty }(Q_0)^3\rightarrow C^{\infty }(Q_0)^2\) would be now (compare with (3.16)) such that, for every \(z=(z^1,z^2)\in C^{\infty }(Q_0)^2\) and for every \(\pi \in C^{\infty }(Q_0)\),
Let \(F_1\in C^\infty (T/2,T)\) and let \(F_2\in C^\infty (T/2,T)\). We define \(z=(z^1,z^2) \in C^\infty (Q_0;{\mathbb {R}}^2)\) and \(\pi \in C^{\infty }(Q_0)\) by
Then \({\mathcal {L}}_0^*(z,\pi )=0\). However, if \((F_1,F_2)\not = (0,0)\), then \({\mathcal {B}}_0^*(z,\pi )\not =0\). Hence, in this case, (3.15) does not hold whatever the linear partial differential operator \({\mathcal {M}}_0\) is and whatever the trajectory \((\overline{y}, \overline{p},\overline{u})\) is.
3.2.2 Number of variables and equations
Let us give some algebraic results about the number of derivatives of a certain order.
Definition 3
Consider a scalar PDE with a smooth (enough) variable \(z\) depending on 4 variables \(x_0,x_1,x_2,x_3\). We call equations of level \(n\) all the different equations we obtain by differentiating the PDE with respect to all the possible multi-integers of length \(n\). The number of “distinct” equations of level \(n\) is denoted \(E(n)\), and the number of “distinct” equations of a level less than or equal to \(n\) is denoted \(F(n)\).
Remark 5
Clearly, \(E(n)\) is also the distinct number of derivatives of order \(n\) for (smooth enough) functions having 4 variables, and \(F(n)\) is also the distinct number of derivatives of an order less or equal than \(n\) for (smooth enough) functions having 4 variables. Moreover, if we consider a scalar PDE with many variables \(z^1,\ldots ,z^k\) depending on \(x_0,x_1,x_2,x_3\) containing derivatives of \(z^1,\ldots z^k\) of order \(m\) at most, the maximum number of derivatives of \(z^1,\ldots z^k\) we may expect in the equations of a level less than or equal to \(n\) is \(kF(n+m)\).
We want to compute \(E\) and \(F\) precisely. One has
Indeed
defines a bijection between the set of \((\alpha _0,\alpha _1,\alpha _2,\alpha _3)\in {\mathbb {N}}^4\) such that \(\alpha _0+\alpha _1+\alpha _2+\alpha _3\leqslant n\) and the set of subsets of \(\{1,2,\ldots ,n+4\}\) having 4 elements. Hence, \(F(n)\) being the number of \((\alpha _0,\alpha _1,\alpha _2,\alpha _3)\in {\mathbb {N}}^4\) such that \(\alpha _0+\alpha _1+\alpha _2+\alpha _3\leqslant n\), we have (3.21). In order to obtain (3.21), it suffices to notice that
3.2.3 A related overdetermined system
Let us now study Eq. (3.18), where the data is \((\varphi ^1,\varphi ^2,\varphi ^3,\varphi ^4) \in C^{\infty }(Q_0)^4\) and the unknown is \((z^1,z^2,\pi )\in C^{\infty }(Q_0)^3\).
Let us explain the idea behind the reasoning we are going to develop in this subsection. Equation (3.18) is “analytically” overdetermined, since we have more equations (4) than unknowns (3). However, if we see (3.18) as a linear system of algebraic unknowns (the unknowns being \(z^1,z^2,\pi \) and their derivatives) the system is now “algebraically” underdetermined: We have 4 equations and \(19\) unknowns. But it is easy to obtain as many new equations as we want: It suffices to differentiate (3.18) enough times. Some new “algebraic unknowns” (the derivatives of \(z^1,z^2,\pi \)) appear, but since the system was “analytically” overdetermined, one can hope that they are not “too many” new unknowns appearing. Notably, one can hope that, after differentiating a sufficient number of times, we obtain more equations than “algebraic unknowns”. We would then deduce Assumption \({\mathcal {A}}_2\) by “inverting” in some sense this well-posed linear system (this will be explained in detail later).
We first eliminate \(\pi \) in our equation (3.18). To reach this goal, in (3.18), we apply \(\partial _3\) to the first and second lines, and use the third line. We obtain the following equations:
The first and second equation of (3.23) contain derivatives of \(z^1\) and \(z^2\) up to order 3 and the third equation derivatives up to order 1. We would like to have the same maximal order of derivatives appearing in the three equations in order to be sure that the derivatives of maximal order appearing in the first and second equation might also appear in the third one. Hence we are going to differentiate the last equation two more times than the others. If we count the maximum number of derivatives of \(z^1\) and \(z^2\) we create by differentiating \(n\) times the first and second equation and \(n+2\) times the third one, we obtain
different derivatives. The number \(G(n)\) of equations we obtain is then
From (3.24) and (3.25), one sees that \(G(n)-H(n)\) is increasing with respect to \(n\) and that
Hence, in order to have more equations than unknowns and as few equations as possible, we choose \(n=19\). We have \(G(19)=30360\) equations and \(H(19)=29900\) unknowns. We can see this system of \(30360\) partial differential equations as a linear system
where \(L_0\in C^\infty (Q_0;{\mathcal {M}}_{30360\times 29900}({\mathbb {R}}))\), \(Z\in {\mathbb {R}}^{29900}\) (\(Z\) contains the derivatives of \(z^1\) and \(z^2\) up to the order \(19\)) and \(\varPhi \in {\mathbb {R}}^{30360}\) (\(\varPhi \) contains the derivatives of \(\varphi \) up to the order \(19\)). Note that \(L_0\) also depends on \(\varepsilon \) and \(\nu \), but this does not need to be emphasized in what follows. Hence, in order to lighten the notations, we will only see \(L_0\) as a function of \(t\) and \(x\) (as for \(\tilde{L}_0,N,\ldots \) that are be introduced later). We order the \(29900\) lines of \(Z\) so that
Assumption \({\mathcal {A}}_2\) can then be written as follows: Prove the existence of a nonempty open subset \(\omega _0\) of \({\mathcal C}_2\) and of a map \(N\in C^\infty (Q_0; {\mathcal {M}}_{6\times 30360}({\mathbb {R}}))\) (\(N\) is the algebraic version of the linear partial differential operator \({\mathcal N}\) introduced in Sect. 3.2, every linear partial differential operator can be alternatively considered as a matrix acting on the derivatives of the input functions) such that
Since the size of the matrix \({L_0}(t,x)\) is very large, it is impossible to find some \(N\) verifying System (3.26) by hand and we will have to do computations on a computer. Notably, it would be more convenient to make \({L_0}\) be a sparse matrix in order to use relevant tools adapted to the study of big sparse linear systems. This is the reason for our simple choices for \(a\), \(b\) and \(c\) given in (2.16), (2.16) and (2.17) (polynomials of small order do not create to many non zero coefficients in \({L_0}\) when they are differentiated). Using (2.1), (2.16), (2.16), (2.17), (2.19) and (2.20), System (3.23) becomes simply, in \(Q_0\),
Let us consider the change of variables
and
(\(e\) appears when we differentiate \(t\mapsto a(t)\) on \(Q_0\)). Let \({\mathbb {R}}[E,S,X]\) be the set of polynomials in the variables \(e\), \(s\), \(x_1\), \(x_2\), \(x_3\), with real coefficients. The \(30360\times 29900\) entries of \({L_0}\) can alternatively be seen as functions depending on \((t,x_1,x_2,x_3,\varepsilon )\) or as elements of \({\mathbb {R}}[E,S,X]\) and, from now on, we consider \({L_0}\) as an element of \({\mathcal M}_{30360\times 29900}({\mathbb {R}}[E,S,X])\). As we will see after, it turns out that many of the entries of \({L_0}\) are the 0 polynomial.
For a positive integer \(k\), let us denote by \({\mathfrak {S}}_k\) the set of permutations of \(\{1,\ldots , k\}\). To each \(\sigma \in {\mathfrak {S}}_k\), we associate the matrix \(S_\sigma \in {\mathcal M}_{k,k}({\mathbb {R}})\) defined by
For two positive integers \(k\) and \(l\), let us denote by \(0_{k\times l}\) the null matrix of \({\mathcal M}_{k\times l}({\mathbb {R}})\) (which is included in \({\mathcal M}_{k\times l}({\mathbb {R}}[E,S,X])\)). The following lemma is a key step for the proof of Proposition 2.
Lemma 1
There exist
and
such that
Let us assume for the moment that this lemma holds and end the proof of Proposition 2. A consequence of Lemma 1 is the following:
Lemma 2
There exists a nonempty open subset \(\omega _0\) of \({\mathcal C}_2\), \(T^*>0\) and \(\varepsilon ^*>0\), such that
Proof of Lemma 2
Let us first point out that \(\text{ det } P\in {\mathbb {R}}[E,S,X]\) and, by (3.31), this polynomial is not the 0 polynomial. Hence there exist a nonnegative integer \(m\) and a polynomial \(\tilde{P}\in {\mathbb {R}}[E,S,X]\) such that
By (3.34), there exist \(\delta >0\), \(C'>0\) and a nonempty open subset \(\omega _0\) of \({\mathcal C}_2\) such that
By the mean value theorem, there exist a positive integer \(l\) and a positive real number \(C^*\) such that
By (2.16), there exists \(\varepsilon ^*\) such that
From (3.35), (3.36) and (3.37), we obtain that
which concludes the proof of Lemma 2. \(\square \)
Let us now go back to the proof of Proposition 2. For every positive integer \(l\), we denote by \(\text {Id}_{l}\) the identity matrix of \({\mathbb {R}}^{l}\). By (3.32), there exists \(U\in C^\infty (Q_0; {\mathcal {M}}_{7321\times 7321}({\mathbb {R}}))\) such that
Let \(\tilde{U} \in C^\infty (Q_0; {\mathcal {M}}_{7321\times 30360}({\mathbb {R}}))\) be defined by
From (3.30), (3.38) and (3.39), one has
Let \(K\in {\mathcal M}_{6,7321}({\mathbb {R}})\) be defined by
From (3.29), (3.40) and (3.41), one has
which shows that (3.26) holds with \(N(t,x):=K\tilde{U}(t,x)S_{{\tilde{\sigma }}}\), and ends the proof of Proposition 2. \(\square \)
To finish the proof of Proposition 2, it suffices to prove Lemma 1.
Proof of Lemma 1
The fact that the dependence of \(\overline{y}\) and its derivatives in the time variable is quite complicated (it is both exponential and fractional) compared to the dependence in the space variable (which is polynomial) is problematic, because it is not very convenient to use for computations on a computer. In the previous proof we have seen \(\text {det }P\) as a polynomial in \(s=\varepsilon a(t)\), \(e=\frac{1}{T-t}\) (which corresponds to terms appearing when we differentiate \(t\mapsto a(t)\)) and \(x\). Assume that we fix \(e=0\): This is equivalent to do “as if” the derivatives of \(a\) were all identically the null function, i.e. to do as if the function \(t\mapsto a(t)\) were replaced by a constant function, which is simpler than our original function \(a\). We will then impose \(e^0=0\) for our computations. Let us set \(\xi ^0:=(e^0,s^0,x^0)\) with \(e^0=0, s^0=1\) and \(x^0=(1.1,1.2,1.3)\).
First of all, let us prove that one can decompose \(M\) as in (3.30) at least at point \(\xi ^0\). We present in the Appendix A how we computed the matrix
thanks to a \(C^{++}\) program.
From now on we assume that we have matrix \(L^0_0\) at our disposal and we are going to explain how to exploit it in order to obtain Lemma 1.
We begin with reordering the columns so that the null columns of \({L_0^0}\) are moved to the last columns. One verifies for example thanks to Matlab that there are exactly \(140\) such columns. There exist \(\overline{\sigma } \in {\mathfrak {S}}_{29900}\) and \(N^0\in {\mathcal {M}}_{30360\times 29760}({\mathbb {R}})\) such that
One problem is that it could happen that some columns of \({L_0^0}\) are equal to 0 but the corresponding columns of \({L_0}\) are not identically null. However, we check that it is not the case (thanks to the evaluation function described in Appendix A).
Let us recall that our goal is to extract a well-chosen submatrix of \(L^0_0\) which is of maximal rank. A reasonable hope would have been that the matrix \(N^0\) (of size \(30360\times 29760\)) itself is of maximal rank 29760 (we would then have obtained something similar to Lemma 1 by choosing some squared extracted matrix of maximal rank \(P^0\) of \(N^0\), which is always possible, the matrix \(P^0\) would then have been of size \(29760\times 29760\) and the non-selected lines would be permuted to obtain matrices \(Q^0\) of size \(600\times 29760\) and \(R^0\) of size \(600\times 140\)). However it turns out to be false, as we will see later.
Since computing the rank of \({L_0^0}\) on a computer is too long because of its size, we introduce the notion of structural rank.
Definition 4
Let \(A\in {\mathcal M}_{n,m}({\mathbb {R}})\) and \(B\in {\mathcal M}_{n,m}({\mathbb {R}})\). We say that \(A\) and \(B\) are structurally equivalent if the following property is verified:
This is an equivalence relation on \({\mathcal M}_{n,m}({\mathbb {R}})\), and we call \(Cl(A)\) the equivalence class of \(A\). The structural rank of a matrix \(A\) (denoted sprank \((A)\) in the following) is the maximal rank of the elements of \(Cl(A)\). Equivalently, if we fill randomly the nonzero coefficients of \(A\), then, with probability 1, the rank of \(A\) is equal to the structural rank.
One sees that the structural rank does not depend on the coefficients of the matrix but only on the distribution of the zeros in the matrix and is never less than the rank. The advantage of the structural rank is that it can be computed fast (in a couple of seconds in our case), especially on sparse matrices. It corresponds to the function sprank in Matlab.
Computing the structural rank of \(N^0\) thanks to Matlab we find that
hence there is no hope that the rank of \(N^0\) is maximal.
To extract a submatrix of \(N^0\) which is of maximal rank, we can, for example, begin with extracting a submatrix of \(P^0\) which is of maximal structural rank, and verify that it is of maximal rank too. The right way to do this is to explore more carefully how the structural rank is computed. In fact the key point is the existence of a decomposition in block triangular form (which is related to the Dulmage–Mendelsohn decomposition for the bipartite graph associated to any matrix, see [14] and [27]) of a matrix.
Proposition 3
Let \(A\) be a matrix. Then one can permute the columns and the lines of \(A\) to obtain a matrix of the following form:
where:
-
1.
\((A_{11}, A_{12})\) is the underdetermined part of the matrix, it always has more columns than rows.
-
2.
\((A_{33}, A_{34})\) is the overdetermined part of the matrix, it always has more rows than columns.
-
3.
\(A_{12}, A_{23}, A_{34}\) are square matrices with nonzero diagonals (in particular these matrices are of maximal structural rank)
-
4.
\(A_{23}\) is the well-determined part of the matrix (if the matrix is square and non-singular, it is the entire matrix).
Moreover, one can permute rows and columns so that \(A_{23}\) is also block triangular. The decomposition obtained is called the block triangular form of matrix \(A\). The structural rank of \(A\) is given by the sum of the structural ranks of \(A_{12}, A_{23}, A_{34}\).
The block triangular form (3.43) (called the coarse decomposition) of the matrix is in fact given by the dmperm function in Matlab, which also gives the permutation that makes the matrix be in the form of (3.43) and a block triangular form for the well-determined part (which is called the fine decomposition). One easily understands how to obtain a matrix in the form of (3.30) thanks to this decomposition: One can (for instance) permute the blocks to obtain
from which we easily deduce decomposition (3.30).
To simplify the computations, we are not going to apply this block triangular decomposition directly to \({L_0^0}\) but to a well-chosen submatrix \(\tilde{L}^{0}_0\). First of all, we go back to \({L_0}\) and select some equations and unknowns: There exist \(\sigma ^0 \in {\mathfrak {S}}_{29900}, {\tilde{\sigma }}^0 \in {\mathfrak {S}}_{30360}, \tilde{Q}\in {\mathcal {M}}_{16623\times 14630}({\mathbb {R}})\) and \(\tilde{R}\in \mathcal {M}_{16623\times 15270}({\mathbb {R}})\) such that (see (3.44))
where \(\tilde{L}_{0}\) corresponds to the equations we obtain by differentiating the two first equations of (3.23) 15 times and the last equation \(17\) times, so that \(\tilde{L}_{0}\) is of size \((G(15),H(15))=(13737,14630)\) (here there are more unknowns than equations but we will see that this will not be a problem).
Let us call
One has
Thanks to Matlab, we find the Dulmage–Mendelsohn decomposition of \(\tilde{L}^{0}_0\) and observe that there exists some permutations matrices \(\sigma ^1\) and \({\tilde{\sigma }}^1\) such that (see (3.44))
with \(\overline{Q}^0\in {\mathcal {M}}_{4687\times 9050}({\mathbb {R}})\), \(\overline{R}^0\in {\mathcal {M}}_{4687\times 5578}({\mathbb {R}})\), and \(\overline{L}^{0}_0\in {\mathcal {M}}_{9050\times 9050}({\mathbb {R}})\) is of maximal structural rank and square (it corresponds to the block \(A_{34}\) in the block triangular decomposition). Applying the Dulmage–Mendelsohn algorithm now on \(\overline{L}^{0}_0\), we can write \(\overline{L}^{0}_0\) in an (upper) block triangular form with \(352\) diagonal blocks, the first \(351\) of them being of “small” size and the latter one being of size \(7321\). Let us call \({\overline{L}^{0}_0}_{(i,j)}\) (with \((i,j)\in [|1;352|]^2\)) the blocks of \(\overline{L}^{0}_0\).
Using this decomposition, one can see (using Matlab) that \(\overline{L}^{0}_0\) is not of maximal rank. However, by computing the rank of the block \({\overline{L}^{0}_0}_{352352}\) thanks to Matlab, one sees that it is of maximal rank 7321. Moreover, we verify that
by looking carefully on Matlab where the columns corresponding to these unknowns have been moved under the action of the permutation matrices \(S_{\sigma ^0}\) and \(S_{\sigma ^1}\). More precisely, \(\partial _{x_i} z^1\) corresponds to the \(i\)-th column of \({\overline{L}^{0}_0}_{(352352)}\) and \(\partial _{x_i} z^2\) to the \((3632+i)\)-th column of \(\overline{L^{0}}_{352352}\).
Let us call
There exist \(\sigma \in {\mathfrak {S}}_{29900}, {\tilde{\sigma }} \in {\mathfrak {S}}_{30360}, {Q^0}\in {\mathcal M}_{23039\times 7321}({\mathbb {R}}[E,S,X])\) and \({R^0}\in {\mathcal M}_{23039\times 22579}({\mathbb {R}}[E,S,X])\) such that
the rank of the first block \(P^0\) being maximal. The distribution of the nonzero coefficients of \(P^0\) is given in Fig. 2.
Now, we come back to the matrix \({L_0}\) and we consider the following matrix:
where \(S_{{\tilde{\sigma }}}\) and \(S_{{\sigma }}\) are introduced in (3.47).
Let us call \(\varTheta \) the set of all coefficients of \({L_0}\) (considered as a polynomial in the variables \(E,S,X\)) that are not identically zero, and \(\varTheta ^0\) the set of all coefficients of \({L^0_0}\) that are not equal to 0. Clearly \(\varTheta ^0\subset \varTheta \) (in fact thanks to Matlab one can see that \(\varTheta ^0\) is much smaller than \(\varTheta \)), moreover \(\varTheta {\setminus }\varTheta ^0\) corresponds to the nonzero coefficients of the matrix that become identically null when we change \(a(t)\) into the function identically equal to 1 and apply it at point \(\xi ^0\). What could happen is that \({\tilde{M}}^0\) is not of block triangular form as in (3.47) (the null block of the matrix \(\tilde{M}^0\) may contain some elements of \(\varTheta {\setminus }\varTheta ^0\)). However, since the only important thing is the location of the elements of \(\varTheta {\setminus }\varTheta ^0\) and not their value, one can verify easily on Matlab that the coefficients of \(\varTheta {\setminus }\varTheta ^0\) do not influence the block form (3.47) (by looking where the elements of \(\varTheta {\setminus }\varTheta ^0\) are moved under the action of \(S_{{\tilde{\sigma }}}\) and \(S_{\sigma }\)), i.e. the permutations \(S_{{\tilde{\sigma }}}\) and \(S_{\sigma }\) also give a decomposition as in (3.47) for \(\tilde{L}^0\): There exists
and
such that
with the relations \(P(\xi ^0)=P^0,Q(\xi ^0)=Q^0,R(\xi ^0)=R^0\).
Property (3.30) follows then directly from (3.45) and (3.48), (3.31) is a direct consequence of the above construction, and (3.29) can be easily deduced by permuting some lines and columns of \(S_{{\tilde{\sigma }}}\) and \(S_{\sigma }\) (thanks to Property (3.46)). Finally, Lemma 1 holds. \(\square \)
Consequently Proposition 2 holds. Moreover, one observes that the linear partial differential operator \({\mathcal {M}}_0\) that we have just created so that (3.13) holds is exactly \({\mathcal P}^*\), (where \({\mathcal P}\) is the differential version of the matrix \(P\) seen as a partial differential operator acting on \((\partial _{x_1}z^1,\partial _{x_2}z^1,\partial _{x_3}z^1,\partial _{x_1}z^2,\partial _{x_2}z^2,\partial _{x_3}z^2)\)) and is of order \(17\) (because we have differentiated the equations of (3.18) \(16\) times, and (3.23) was obtained by differentiating System (3.18) 1 time). Hence, the corresponding operator \({\mathcal M}\) in equality (3.5) is also of order 17.
This concludes the proof of Proposition 2. \(\square \)
3.3 Controllability of the linear control system (3.6)
In this subsection, we prove some technical lemmas that imply the null controllability of (3.6) with controls which are derivatives of smooth enough functions having a small support. This is needed to ensure that the controls are in the image of \({\mathcal B}\) (this is exactly Assumption \({\mathcal {A}}_1\)) and to take into account Remark 3.
The first lemma is the following one (remind that \(\overline{y}\) is one of the trajectories constructed in Sect. 2). It consists in a Carleman estimate with curl observation. We call \(D_{\overline{y}}\) the operator
\(D_{\overline{y}}\) is exactly the opposite of the adjoint operator of \(y\mapsto (\overline{y}.\nabla y)+(y.\nabla )\overline{y}\) (because \(\overline{y}\) is divergence-free).
Lemma 3
Let \(\theta : \varOmega \rightarrow [0,+\infty )\) be a lower semi-continuous function which is not identically 0 and let \(r\in (0,1)\). There exists \(C_1>0\) such that, for every \(K_1>C_1\), every \(\nu \geqslant K_1(1-r)/r\), there exists \(\varepsilon ^0\) such that for every \(\varepsilon \in (0,\varepsilon ^0)\), there exists \(C>0\) such that, for every \(g\in L^2((0,T)\times \varOmega )^3\) and for every solution \(z\in L^2((0,T),V)\cap L^\infty ((0,T),H)\) of the adjoint of the linearized Navier–Stokes system
one has
Proof of Lemma 3
In this proof, our system, which is initially defined on \((0,T)\), will only be considered on the interval of time \((T/2,T)\). In fact, in the following, (see, in particular, the proof of Proposition 4), we will not act on the system on the interval \((0,T/2)\), hence we only need a Carleman estimate on \((T/2,T)\). For our proof, we need to use the particular form of our \(\overline{y}\) in time, in particular that (see (2.16))
Without loss of generality, we may assume that there exists a nonempty open subset \(\omega ^*\) of \(\varOmega \) such that \(\theta = 1_{\omega ^*}\). Let us now give some other notations. Let \(\eta ^0\in C^2(\overline{\varOmega })\) such that \(\eta >0\) and \(|\nabla \eta ^0|>0\) in \(\overline{\varOmega }{\setminus }\omega ^*\) and \(\eta ^0=0\) on \(\partial \varOmega \). For the existence of \(\eta ^0\), see [17, Lemma1.1,p. 4]. Let us call
and
We call \(Q_{/2}:=(T/2,T)\times \varOmega \). Using (3.51) and [21, Proposition 3.1, p. 6] on the adjoint system (3.49) (where we see the first and zero order terms of this equation as a second member, because Proposition 3.1 of [21] concerns only the Stokes system), one has, for some \(C\) large enough, \(\lambda \geqslant C\) and \(s\geqslant C\),
In fact, looking carefully at the proof of Proposition 3.1 of [21], one remarks (just by changing the weight \(\rho (t):=e^{-s\alpha ^*}\) by \(\rho (t):=e^{-\mu s\alpha ^*}\) where \(\mu >1\) is a parameter that can be chosen as large as we wish) that the previous inequality can be improved in the following way, as soon as \(s\) is large enough, for every \(\mu >1\) (the constant \(C\) depends on \(\mu \)):
As usual, we now change our weights so that they do not vanish at time \(t=T/2\). Let us call \(l:[T/2,T]\rightarrow {\mathbb R}\) defined by \(l(t)=T^2/16\) on \([T/2,3T/4]\) and \(l(t)=(t-T/2)(T-t)\) on \([3T/4,T]\). Next, we define
Clearly, the functions \(\alpha \) and \(\beta \) coincide on \([3T/4,T]\), as well as the functions \(\xi \) and \(\gamma \). Using classical energy arguments, we deduce the existence of \(C\) (depending now on \(s,\lambda \), which are assumed to be large enough and fixed from now on, and \(\mu \)) such that
One remarks that, since \(\nabla \cdot z=0\) in \(Q\) and \(z=0\) on \((0,T)\times \partial \varOmega \), one has
Using Poincaré’s inequality, we also have
Putting this into (3.53), one obtains
Let us define, for \(\mu >1\),
From equality (3.57), one deduces the existence of \(C^*>0\) (depending on \(\mu >1\), \(\lambda >>1\) and \(s>>1\)) such that
Moreover, from equality (3.57), there exists \(\widehat{C}>0\) (depending on \(\mu >1\), \(\lambda >>1\) and \(s>>1\)) such that
Fixing \(s,\lambda \) and making \(\mu \rightarrow + \infty \), one easily sees that \(K_0/K_1=\frac{1+\mu +\sqrt{\mu }}{\mu -\sqrt{\mu }}\rightarrow 1^+\) so that for every \(r\in (0,1)\), we have for \(\mu \) large enough, \(K_0<K_1/r\). For \(\varepsilon >0\) small enough and for \(\nu \) large enough (\(\nu \geqslant K_0-K_1\)), one can absorb the undesired terms \(\varepsilon ^2\int _{Q_{/2}} e^{-2\mu s\beta ^*}e^{-\frac{2\nu }{(T-t)^5}}(|z|^2+|\nabla z|^2)\) from the right-hand side of (3.56). Then using some classical energy estimates together with (3.59) and (3.60), one obtains (3.50). \(\square \)
From now on, we set
Let us now derive from this Carleman inequality a result of null-controllability with controls which are derivatives of smooth functions. Let \(\widehat{1}_{\omega _0}: {\mathbb {R}}^3\rightarrow [0,1]\) be a function of class \(C^\infty \) which is not identically equal to 0 and having a support included in \(\omega _0\), where \(\omega _0\) was introduced in Lemma 2. We apply Lemma 3 with \(\theta = \widehat{1}_{\omega _0}\). One has the following proposition.
Proposition 4
With the notations of Lemma 3, let \(f\in L^2(Q)^3\) be such that \({\rho _r}^{-1/2}f\in L^2(Q)^3\) and let us consider the following linearized Navier–Stokes control system
where the control is \(v\). Then, for every \(y_0\in V\), there exists a solution \((y,p,v)\) of (3.61) such that \(y(0,\cdot )=y^0\) and for every \(\tilde{K}_1\) verifying \(0<\tilde{K}_1<K_1\),
Remark 6
What is important in the previous proposition is the fact the controls are very regular (which is quite new and interesting in itself) and that the controls are derivatives (in fact curls) of functions, as in [20]. In the following, it is enough to obtain a regularity \(L^2((0,T),H^{53}(\varOmega )^3)\cap H^{27}((0,T),H^{-1}(\varOmega )^3)\) for \(e^{\tilde{K}_1(2-1/r)/(2(T-t)^5)}(\nabla \wedge v){\widehat{1}_{\omega _0}}\) but the following proof can be easily adapted to deduce controls \(v\) with
for every given \(m\) as large as one wants.
Proof of Proposition 4
In the following, we only control on the interval of time \((T/2,T)\), i.e. we set \(v=0\) on \((0,T/2)\) and let the corresponding solution \((y,p)\) of (3.61) on \((0,T/2)\) evolve naturally until time \(T/2\). Let \(y^{T/2}=y(T/2, \cdot )\).
Let \(P:L^2(0,L)^3\rightarrow L^2(0,L)^3\) be the Leray projector \(P\varphi :=\varphi -\nabla p\), where \(\Delta p=\text { div } \varphi \) in \(\varOmega \) and \(\partial p/\partial n=\varphi \cdot n\) on \(\partial \varOmega \), (\(n\) is the unit outward normal vector on \(\partial \varOmega \)). Since \(P\Delta \varphi = \Delta P\varphi \) for every \(\varphi \in C^\infty _0(\varOmega )^3\), \(P\) can be extended as a continuous linear map from \(H^{-1}(\varOmega )^3\) to \(H^{-2}(\varOmega )^3\). We still denote by \(P\) this extension. Let \(S: {\mathcal {D}}'((T/2,T),H^1_0(\varOmega )^3)\rightarrow {\mathcal {D}}'((T/2,T), H^{-2}(\varOmega )^3)\) and \(S^*: {\mathcal {D}}'((T/2,T); H^1_0(\varOmega )^3)\rightarrow {\mathcal {D}}'((T/2,T), H^{-2}(\varOmega )^3)\) be defined by
(\(S^*\) corresponds to the linearized time-dependent Navier–Stokes operator and, formally, \(S\) is the adjoint of \(S^*\)).
Since \(y^{T/2}\) is regular enough, one can assume from now on without loss of generality that \(y^{T/2}=0\) by adding some suitable term in the source term \(f\) (that we still call \(f\)) that still satisfies \({\rho _r}^{-1/2}f\in L^2(Q)^3\), and one can always assume that \(P f=f\) by changing the pressure. We define a closed linear unbounded operator \({\mathcal {S}}: L^2(Q_{/2})^3\rightarrow L^2(Q_{/2})^3\) by
We call
and
where the pivot space is \(L^2(Q_{/2})^3\). For every \((k,l)\in {\mathbb Z}^2\) such that \(k\leqslant l\), one has
Moreover \(X_m\) is an Hilbert space for the scalar product
The associated norm is denoted \(||.||_{X_m}\). For \(m\in {\mathbb N}\), one can define \({\mathcal {S}}^*\) as an operator from \(X_{-m}\) into \(X_{-m-1}\) by setting, for every \(z_1\in X_{-m-1}\) and \(z_2\in X_{m+1}\),
(One easily checks that this definition is consistent: it gives the same image if \(z_1\) is also in \(X_{-m'}\) for some \(m'\in {\mathbb {N}})\). This implies in particular that, for every \(z_1\in L^2(Q_{/2})^3\) and for every \(z_2\in X_m\), one has, for every \(0\leqslant j\leqslant l\),
Let \({\mathcal {H}}_0\) be the set of \(z\in H^1((T/2,T), L^2(\varOmega )^3)\cap L^2((T/2,T),H^2(\varOmega )\cap V)\) such that
Let \(a\) be the following bilinear form defined on \(\mathcal {H}_0\):
From (3.50), we deduce that \(a\) is a scalar product on \({\mathcal {H}}_0\). Let \({\mathcal {H}}\) be the completion of \({\mathcal {H}}_0\) for this scalar product. Note that, still from (3.50) and also from the definition of \({\mathcal {H}}\), \({\mathcal {H}}\) is a subspace of \(L^2_{loc}([T/2,T),H^1_0(\varOmega )^3)\) and, for every \(z\in {\mathcal {H}}\), one has (3.70), (3.70) and
Let us now consider the linear form \(l\) defined on \({\mathcal {H}}\) by
The linear form \(l\) is well-defined and continuous on \({\mathcal {H}}\) since, by the Cauchy–Schwarz inequality together with (3.72), one has, for every \(w\in {\mathcal {H}}\),
Applying the Riesz representation theorem, there exists a unique
verifying, for every \(w\in {\mathcal {H}}\),
with
We then set
We want to gain regularity on \(\tilde{y}\) (by accepting to have a weaker exponential decay rate for \(\tilde{y}\) when \(t\) is close to \(T\)). Let \(\psi \in C^\infty ([T/2,T])\) and \(y\in X_{-1}\). One can define \(\psi y\in X_{-1}\) by the following way. Since \({\mathcal S}^*:X_{0}\rightarrow X_{-1}\) is onto, there exists \(h\in X_0\) such that \({\mathcal S}^*h=y\). We define \(\psi y \) by
This definition is compatible with the usual definition of \(\psi y\) if \(y\in X_0\). We can then define by induction on \(m\) \(\psi y\in X_{-m}\) for \(\psi \in C^\infty ([T/2,T])\) and \(y\in X_{-m}\) in the same way. Using (3.77), this allows us to define
From (3.75), (3.76), (3.77) and (3.79), one gets
Let \(\tilde{K}_1\in (0,K_1)\) and \(\tilde{\rho }_1:=e^{-\tilde{K}_1/(T-t)^5}\). Using (3.76), (3.77) and (3.80), one has
We want to deduce from (3.81) that \(\tilde{y}\) is more regular. This can be achieved thanks to the following lemma:
Lemma 4
Let \(m\in {\mathbb {N}}\). If \(y\in X_{-m}\) and \({\mathcal S}^*y\in X_{-m}\), then \(y\in X_{-m+1}\).
Proof of Lemma 4
If \(m=0\), Lemma 4 follows from usual estimates on usual regularity property of solutions of the linearized Navier–Stokes system. From now on, we assume that \(m\in {\mathbb {N}}^*\). Let \(h\in X_m\). Since \({\mathcal S}\) is an operator from \(X_{m+1}\) onto \(X_{m}\), there exists \(\alpha \in X_{m+1}\) such that \({\mathcal {S} }\alpha =h\). Thanks to (3.68), one has
the last equality coming from the fact that \({\mathcal {S}}^*y\in X_{-m}\). We deduce from (3.82) that there exists some constant \(C>0\) such that for every \(h\in X_m\),
which shows that \(y\in X_{1-m}\). This concludes the proof of Lemma 4. \(\square \)
From (3.81) and Lemma 4, one gets that
Using an easy induction argument together with Lemma 4 (and the fact that one can choose \(\tilde{K}_1<K_1\) arbitrarily close to \(K_1\)), we deduce that, for every \(\tilde{K}_1\in (0,K_1)\), \((\sqrt{\rho _1}/\sqrt{\tilde{\rho }_1})\tilde{y}\in X_{0}\).
Let us now focus on \(\widehat{u}\). Let us call \(v:=\rho _1 \widehat{z}\). Using (3.72), one gets that
Using (3.74) together with regularity results for \(S\) applied on \({\tilde{\rho }_1}^{-1}{\rho _r}^{1/2}v\in L^2(Q_{/2}) \) and, as above for the proof of (3.84), a bootstrap argument (together with the fact that one can choose \(\tilde{K}_1\in (0,K_1)\) arbitrarily close to \(K_1\)), one obtains that
From (3.85) and (3.80), we deduce (by looking the equation verified by \((1/\sqrt{\tilde{\rho }_1}) \widehat{y}\) and using usual regularity results for linearized Navier–Stokes system) that
Proposition 4 follows from (3.80), (3.86) and (3.85).\(\square \)
3.4 Null-controllability of (3.1)
To finish, one can gather the results of Sects. 3.2 and 3.3 in order to apply Proposition 1 and obtain a controllability result on (3.1). However, we cannot work in the \(C^\infty \) setting of Proposition 1, so we need to take into account Remark 3 and to be careful concerning the spaces we are working with.
Proposition 5
For every \(T>0\) small enough, for every \(\alpha \in (0,1)\), there exists \(r_0\in (0,1)\) such that for every \(r\in (r_0,1)\), there exists \(C_1>0\) such that for every \(K_1>C_1\), for every \(f\in L^2(Q)\) be such that \(e^{\frac{K_1}{2r(T-t)^5}}f\in L^2(Q)^3\) and for every \(y^0\in V\), if
there exists a solution \((y,p,v)\) of the following linearized Navier–Stokes control system
such that
Proof of Proposition 5
We want to apply Proposition 1. First of all, we deal with Assumption \({\mathcal {A}}_1\). We apply Proposition 4: There exists a solution \((y^*,p^*,v^*)\) of (3.61) such that \(y^*(0,\cdot )=y^0\) and, for every \(\tilde{K}_1\) verifying \(0<\tilde{K}_1<K_1\),
Using well-known interpolation results (see for example [24, Section 13.2, p. 96]) and setting \(n:=27\), we obtain that
Let us call \(w:=\widehat{1}_{\omega _0}(\nabla \wedge v^*)\), which is supported in \(Q_0\). One observes that
Hence in view of equality (3.10) and setting
one has \(\nabla \wedge w\in Im({\mathcal B})\) and Assumption \({\mathcal {A}}_1\) holds.
Now, we observe that Assumption \(\mathcal {A}_2\) follows from Proposition 2. Let \((\tilde{y},\tilde{p}, \tilde{v})\) be defined by
where \({\mathcal M}\) is as in (the proof of) Proposition 2. It makes sense to apply \({\mathcal M}\) to \(w\) because \({\mathcal M}\) is a partial differential operator of order \(17\) and \(w\in H^{17}(Q_0)\).
Using the fact that the operator \({\mathcal {M}}\) is a partial differential operator of order \(17\) and that the coefficients of \({\mathcal M}\) explode at time \(t=T\) at rate at most \(e^{\frac{7321 \nu }{(T-t)^5}}\), as it follows from the construction of \(\overline{y}\) given in Sect. 2 (see in particular (2.16)) and the construction of \(M\) given in the proof of Proposition 2, one has
for every \(K_2<K_1(2-1/r)-7321 \nu \). In order to be obtain \(\tilde{y}(T,.)=0\), it is enough to have an exponential decay for \(y\) at time \(T\), i.e. to impose
which, with (3.87), is equivalent to
which can be ensured since \(r\) can be arbitrarily chosen close to 1. Let \(\alpha \in (0,1)\). We set
Then, if \(r\in (r_0,1)\), one has
By (3.96), there exists \(K_2\) such that
Finally, one can apply (the proof of) Proposition 1 and we set
Thanks to (3.91), (3.92), (3.93) and (3.93), one has
Then, using usual regularity results for the linearized Navier–Stokes operators on \(e^{\frac{K_2}{2(T-t)^5}}y\) (now considered on the entire time interval \((0,T)\)), we obtain
as soon as \(y^0\in V\). The proof of Proposition 5 is completed. \(\square \)
4 Proof of Theorem 1
To conclude, we are going to apply an inverse mapping theorem to go back to the nonlinear system, which is the following (see [2, Chapter 2, Section 2.3]):
Proposition 6
Let \(E\) and \(F\) be two Banach spaces. Let \(e_0\in E\) and \({\mathcal F}:E\rightarrow F\) which is of class \(C^1\) in a neighborhood of \(e_0\). Assume that the operator \(d{\mathcal F}(e_0)\in {\mathcal L}_c(E,F)\) is onto. Then there exist \(\eta >0\) and \(C>0\) such that for every \(g\in F\) verifying \(||g-{\mathcal F}(e_0)||<\eta \), there exists \(e\in E\) such that
-
1.
\({\mathcal F}(e)=g\),
-
2.
\(||e-e_0||_E\leqslant C||g-{\mathcal F}(e_0)||_F\).
We are going to use the same techniques as in [12]. Let \(\alpha \in (0,1)\), and let us consider some \(r\in (r_0,1)\) where \(r_0\) verifies (3.95). We apply Proposition 6 with \(E\) and \(F\) defined in the following way. Let \(E\) be the space of the functions
such that
-
1.
\(e^{\frac{\alpha {K}_1}{2(T-t)^5}}y\in L^\infty ((0,T),V)^3\cap L^{2}((0,T),H^2(\varOmega )^3\cap V)\),
-
2.
\(\nabla p\in L^2(Q)\),
-
3.
\(e^{\frac{\alpha {K}_1}{2(T-t)^5}}v\in L^{2}(Q)^3\) and the support of \(v\) is included in \(Q_0\),
-
4.
\(e^{\frac{K_1}{2r(T-t)^5}}(y_t-\Delta y +(\overline{y}\cdot \nabla )y+(y\cdot \nabla )\overline{y} +\nabla p-(0,0,v))\in L^2(Q)^3\),
-
5.
\(y(0,\cdot )\in V\),
equipped with the following norm which makes it a Banach space:
Let \(F\) be the space of the functions \((h,y^0)\in L^2(Q)^3\times V\) such that
equipped with the following scalar product which makes it a Hilbert space:
We define
To apply the previous inverse mapping theorem, we first show the following lemma.
Lemma 5
The map \({\mathcal F}\) has its image included in \(F\) and is of class \(C^1\) on \(E\).
Proof
We see that \({\mathcal F}= {\mathcal F}_1+ {\mathcal F}_2\) with
and
Thanks to the construction of \(E\) and \(F\) we have \({\mathcal F}_1: E\rightarrow F\) and \({\mathcal F}_1\) is continuous, so, since \({\mathcal F}_1\) is linear, \({\mathcal F}_1\) is of class \(C^1\). The map \({\mathcal F}_2\) is a quadratic form, hence to prove that it maps \(E\) into \(F\) and is of class \(C^1\), it is sufficient to prove that it is continuous, i.e. to prove that
We choose \(r\) and \(\alpha \) so that
(One can take for example \(\alpha =3/4\) and \(r\in (0,1)\) close enough to 1) Let us call
This definition of \(\tilde{y}\) and inequality (4.2) imply that
which gives that
We also have
A classical Sobolev embedding in dimension 3 together with (4.5) imply that
Direct computations imply that
From (4.4), (4.6) and (4.7), we obtain
which, together with (4.3), gives (4.1). \(\square \)
We now consider the element \(e_0=(0,0,0)\) and we compute
Proposition 5 implies that this application is onto. Hence, taking \(g=(0,y^0)\) and applying Proposition 6, Theorem 1 easily follows (in particular because the trajectory \(\overline{y}\) can be chosen as small as we want since \(\varepsilon \) can be arbitrarily small).
References
Alabau-Boussouira, F., Léautaud, M.: Indirect controllability of locally coupled wave-type systems and applications. J. Math. Pures Appl. (9) 99(5), 544–576 (2013)
Alekseev, V.M., Tikhomirov, V.M., Fomin, S.V.: Optimal control. Contemporary Soviet Mathematics. Consultants Bureau, New York (1987) (translated from the Russian by V. M. Volosov)
Ammar-Khodja, F., Benabdallah, A., Dupaix, C., González-Burgos, M.: A generalization of the Kalman rank condition for time-dependent coupled linear parabolic systems. Differ. Equ. Appl. 1(3), 427–457 (2009)
Ammar-Khodja, F., Benabdallah, A., González-Burgos, M., de Teresa, L.: The Kalman condition for the boundary controllability of coupled parabolic systems. Bounds on biorthogonal families to complex matrix exponentials. Journal de Mathématiques Pures et Appliquées 96(6), 555–590 (2011).
Ammar-Khodja, F., Benabdallah, A., González-Burgos, M., de Teresa, L.: Recent results on the controllability of linear coupled parabolic problems: a survey. Math. Control Relat. Fields 1(3), 267–306 (2011)
Carreño, N., Guerrero, S.: Local null controllability of the \(N\)-dimensional Navier–Stokes system with \(N-1\) scalar controls in an arbitrary control domain. J. Math. Fluid Mech. 15(1), 139–153 (2013)
Coron, J.-M.: Global asymptotic stabilization for controllable systems without drift. Math. Control Signals Syst. 5(3), 295–312 (1992)
Coron, J.-M.: Contrôlabilité exacte frontière de l’équation d’Euler des fluides parfaits incompressibles bidimensionnels. C. R. Acad. Sci. Paris Sér. I Math. 317(3), 271–276 (1993)
Coron, J.-M.: On the controllability of the \(2\)-D incompressible Navier–Stokes equations with the Navier slip boundary conditions. ESAIM Control Optim. Calc. Var. 1, 35–75 (1995/1996) (electronic)
Coron, J.-M.: Control and nonlinearity, volume 136 of Mathematical Surveys and Monographs. American Mathematical Society, Providence (2007)
Coron, J.-M., Fursikov, A.V.: Global exact controllability of the \(2\)D Navier–Stokes equations on a manifold without boundary. Russ. J. Math. Phys. 4(4), 429–448 (1996)
Coron, J.-M., Guerrero, S.: Local null controllability of the two-dimensional Navier–Stokes system in the torus with a control force having a vanishing component. J. Math. Pures Appl. (9) 92(5), 528–545 (2009)
Coron, J.-M., Guerrero, S.: Null controllability of the \(N\)-dimensional Stokes system with \(N-1\) scalar controls. J. Differ. Equ. 246(7), 2908–2921 (2009)
Dulmage, A.L., Mendelsohn, N.S.: Coverings of bipartite graphs. Can. J. Math. 10, 517–534 (1958)
Fernández-Cara, E., Guerrero, S., Imanuvilov, O.Y., Puel, J.-P.: Local exact controllability of the Navier–Stokes system. J. Math. Pures Appl. (9) 83(12), 1501–1542 (2004)
Fernández-Cara, E., Guerrero, S., Imanuvilov, O.Y., Puel, J.-P.: Some controllability results for the \(N\)-dimensional Navier–Stokes and Boussinesq systems with \(N-1\) scalar controls. SIAM J. Control Optim. 45(1), 146–173 (2006) (electronic)
Fursikov, A.V., Imanuvilov, O.Y.: Controllability of evolution equations. Lecture Notes Series, vol. 34. Seoul National University Research Institute of Mathematics Global Analysis Research Center, Seoul (1996)
Fursikov, A.V., Imanuvilov, O.Y.: Exact controllability of the Navier–Stokes and Boussinesq equations. Russ. Math. Surveys 54, 565–618 (1999)
Gromov, M.: Partial differential relations, volume 9 of Ergebnisse der Mathematik und ihrer Grenzgebiete (3). Results in Mathematics and Related Areas (3). Springer, Berlin (1986)
Guerrero, S.: Controllability of systems of Stokes equations with one control force: existence of insensitizing controls. Ann. Inst. H. Poincaré Anal. Non Linéaire 24(6), 1029–1054 (2007)
Gueye, M.: Insensitizing controls for the Navier–Stokes equations. Ann. Inst. H. Poincaré Anal. Non Linéaire 30(5), 825–844 (2013)
Imanuvilov, O.Y.: On exact controllability for the Navier–Stokes equations. ESAIM Control Optim. Calc. Var. 3, 97–131 (1998) (electronic)
Imanuvilov, O.Y.: Remarks on exact controllability for the Navier–Stokes equations. ESAIM Control Optim. Calc. Var. 6, 39–72 (2001) (electronic)
Lions, J.-L., Magenes, E.: Problèmes aux limites non homogènes et applications, vol. 1. Travaux et Recherches Mathématiques, No. 17. Dunod, Paris (1968)
Lions, J.-L., Zuazua, E.: A generic uniqueness result for the Stokes system and its control theoretical consequences. In Partial differential equations and applications, volume 177 of Lecture Notes in Pure and Applied Mathematics, pp. 221–235. Dekker, New York (1996)
Mauffrey, K.: On the null controllability of a \(3\times 3\) parabolic system with non-constant coefficients by one or two control forces. J. Math. Pures Appl. (9) 99(2), 187–210 (2013)
Pothen, A., Fan, C.-J.: Computing the block triangular form of a sparse matrix. ACM Trans. Math. Software 16(4), 303–324 (1990)
Temam, R.: Navier–Stokes equations, volume 2 of Studies in Mathematics and its Applications. Theory and Numerical Analysis. revised edition, with an appendix by F. Thomasset. North-Holland Publishing Co., Amsterdam (1979)
Acknowledgments
The authors would like to thank Sergio Guerrero for fruitful discussions concerning Proposition 4.
Author information
Authors and Affiliations
Corresponding author
Additional information
Work supported by the ERC advanced grant 266907 (CPDENL) of the 7th Research Framework Programme (FP7).
Appendix A: Creation of the matrix \({L_0}\)
Appendix A: Creation of the matrix \({L_0}\)
In this appendix, we explain how the matrix \({L_0}\) at point \(\xi ^0\) (which represents all the differentiated equations of System (3.27) up to the order \(19\)) was created. The program is written in \(C^{++}\), using the library uBLAS which is well-adapted to the manipulation of sparse matrices. It is a parallel openMP algorithm, using 8 cores. We are not going to give all the technical details but just explain rapidly the spirit of the algorithm. To simplify, we will assume that the following “black boxes” (that had to be created) are at our disposal:
-
1.
An evaluation function ep which evaluates a polynomial (represented by a vector) at \(\xi ^0\). This evaluation function can be created so that it can verify that \(\xi ^0\) is not a root of the polynomial \(P(0,.,.)\). (one just has to see if the evaluation is equal to 0 whereas the polynomial has nonzero coefficients).
-
2.
A derivation function deqex which differentiates an equation of level \(m\) with respect to \(x_1,x_2 ,x_3\) or \(t\).
A partial differential equation which is a derivative of order \(m\) of some of the equations of (3.27) will be represented in a matricial form in the following way: We know that there are at most \(F(m+3)\) derivatives appearing, and we observe that the coefficients are polynomials in \((x_1,x_2,x_3)\) of an order less than 4 (it is a vector space of dimension \(35\)). Hence an equation of order \(m\) is represented by a matrix with \(F(m+3)\) lines and \(35\) columns, where on each line one can find the coefficient of the partial derivatives of \(z^1\) (or \(z^2\) appearing) corresponding to the number of this line, thanks to the natural bijection between \({\mathbb N}^4\) and \({\mathbb N}\). Since we have three equations in (3.27) and two unknowns (\(z^1\) and \(z^2\)), one can write the matrix \(M\) in the following way:
For \(i=1,2,3, A_i\) represents the derivatives of \(z^1\) appearing in the derivatives of the \(i\)-th equation of (3.27) and \(B_i\) those of \(z^2\). Hence, we can compute these \(A_i\) and \(B_i\) separately and then gather them to obtain \({L_0^0}\).
The algorithm is the following. We explain it for the first equation of (3.27) and for the unknown \(z^1\) (i.e. for \(A_1\), but it is the same for the other matrices).
-
1.
We create a matrix \(e\) that represents the equation. We use ep to fill the line of \({L_0^0}\) corresponding to the equation in a .txt file under the form \(i\) \(j\) \(A_1(i,j)\). We create a matrix \(h\) which is empty for the moment. In fact in \(e\) we will keep the equations of level \(m-1\) and in \(h\) we will fill the equations of level \(m\).
-
2.
We create a “for” loop on \(m\) which will represent the level of equations we are creating. The integer \(m\) goes from 1 to \(19\) since we differentiate \(19\) times at most.
-
3.
We create a second “for” loop in the interior of the first loop on a number \(n\) which represents one of the equations of level \(m\). Thanks to the definition of the function \(F\) given in Sect. 3.2.2, we have \(F(m-1)+1\leqslant n\leqslant F(m)\). If \(m=1\), then \(n\) goes from \(F(0)+1=2\) to \(F(1)=5\) (\(n\) represents \(\partial _1\), \(\partial _2\), \(\partial _3\) or \(\partial _t\)). If \(m=2\), then \(n\) goes from \(F(1)+1=6\) to \(F(2)=15\) (\(n\) represents \(\partial ^2_{11},\partial ^2_{12},\partial ^2_{13},\partial ^2_{1t},\partial ^2_{22},\partial ^2_{23},\partial ^2_{2t},\partial ^2_{33},\partial ^2_{3t}\) or \(\partial ^2_{tt}\)), etc. This loop is parallelized on our 8 cores. In this loop, we want to create the \(n\)-th equation denoted \(E_n\), which is of level \(m\). Hence we take a suitable equation of level \(m-1\) denoted \(E_r\) which is so that if we differentiate \(E_r\) with respect to \(1,2,3\) or \(t\), we obtain \(E_n\). For example, if we consider \(m=2\) and if we want to obtain the first equation of (3.27) differentiated two times with respect to 1, then we consider the equation \(E_r\) to be the first equation of (3.27) differentiated one time with respect to 1 and differentiated with respect to 1 to obtain \(E_n\).
-
4.
Once the loop on \(n\) is ended, we have in our matrix \(e\) all the equations of level \(m-1\) and in \(h\) we have just created all the equations of level \(m\). Now we just have to use our evaluation function ep on \(h\) to obtain the coefficients of the lines of \(A_1\) corresponding to the equations that are of level \(m\), i.e. the equations numbered from \(F(m-1)+1\) to \(F(m)\). We write these coefficients in our .txt file under the form \(i\) \(j\) \(A_1(i,j)\).
-
5.
We update now \(e\), take \(e=h\), we empty \(h\) and we can go to the following loop \(m+1\).
At the end we have created a file containing the coefficients of a sparse matrix \(A_1\) of size \((8855,14950)\). Using the same program with \(z^2\) and the two others equation we obtain five other files representing five matrices that we gather as in (4.8) to obtain the matrix \({L_0}(\xi ^0)={L_0^0}\). Our matrix \({L_0^0}\), which represents all the equations, is of size \((30360,29900)\) and has 651128 nonzero coefficients. Only 0.07 % of the coefficients are different from 0, with an average of 21.44 nonzero coefficients on each row, which is logical since we are working with coefficients that are polynomials of small degree, so we do not create many terms on each line when we differentiate the equations. In the above figure, one can observe how the nonzero coefficients of \({L_0^0}\) are distributed (Fig. 3).
Rights and permissions
About this article
Cite this article
Coron, JM., Lissy, P. Local null controllability of the three-dimensional Navier–Stokes system with a distributed control having two vanishing components. Invent. math. 198, 833–880 (2014). https://doi.org/10.1007/s00222-014-0512-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00222-014-0512-5