Abstract
We discuss and illustrate the stability issues associated with ill-posed linear operator equations and the need to have regularization methods for obtaining stable approximate solutions. As illustration, we describe two simple regularization methods, namely, the Lavrentiev regularization in the setting of a Banach spaces and Tikhonov regularization in the setting of Hilbert spaces, and discuss on the corresponding error estimates. We also indicate procedures to obtain error estimates under milder general source conditions and also to obtain better error estimates under modified methods. The discussed procedures include some of the work of the author as well.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Solvability and stability
Consider the problem of solving the operator equation
where \(T: X\rightarrow Y\) is a linear operator between normed linear spaces X and Y (over the same field \({\mathbb {K}}\in \{{\mathbb {R}}, {\mathbb {C}}\}\)) and \(y\in Y\). Equation (1.1) is said to be an ill-posed equation if for some \(y\in Y\), either
-
(1)
it does not have a solution or
-
(2)
it has more than one solution or
-
(3)
there is a solution which does not depend continuously on the data y.
In applications, one has to deal with approximate data, say \(\tilde{y}\) in place of y which may not be in R(T). In fact, if \(R(T)\not =Y\) and \(y\in R(T)\), then \(\forall \, \delta >0\), \(\exists \, y^\delta \in Y\setminus R(T)\) such that \(\Vert y-y^\delta \Vert \le \delta \).
There are many operators of practical importance for which R(T) is not even closed in Y. For instance, if T is a compact operator of infinite rank, then R(T) is not closed (cf. [11]). In this connection, let us observe the following result.
Theorem 1
Let T be a bounded linear operator between Banach spaces X and Y such that R(T) is not closed in Y. Then for every \(x\in X\) and for every sequence \((\varepsilon _n)\) of positive real numbers, there exists a sequence \((x_n)\) in X such that
Before giving the proof of the above theorem, let us recall that a linear operator \(T: X\rightarrow Y\) between normed linear spaces X and Y is said to be bounded below if there exists \(c>0\) such that \(\Vert Tu\Vert \ge c\Vert u\Vert \) for all \(u\in X\). We shall make use of part of the following lemma (cf. [11]).
Lemma 2
Let X and Y be normed linear spaces and \(T: X\rightarrow Y\) be bounded below. Then the following are true.
-
(1)
T is one-one and its inverse is continuous from its range.
-
(2)
If X is a Banach space and T is continuous, then R(T) is closed.
Proof of Theorem 1
Suppose R(T) is not closed and \(x\in X\). Let \((\varepsilon _n)\) be a sequence of positive real numbers. Since X is a Banach space and T is not bounded below, there exists a sequence \((u_n)\) in X such that
Let \(v_n = \frac{u_n}{\varepsilon _n\Vert u_n\Vert }\). Then we have
Taking \(x_n:= x + v_n\), we have \(Tx_n = Tx+Tv_n \) so that
for all \(n\in {\mathbb {N}}.\) \(\square \)
Remark 3
Note that, by Theorem 1, for each \(n\in {\mathbb {N}}\), \(\exists x_n\in X\) such that
Let us illustrate Theorem 1 in the context of general Hilbert spaces.
Example 4
Let X and Y be infinite dimensional Hilbert spaces, \((\varphi _n)\) and \((\psi _n)\) be orthonormal sequences in X and Y, respectively. Let \((\lambda _n)\) be a sequence of positive scalars such that \(\lambda _n\rightarrow 0\). Let
Let \(x\in X\) and \(x_k:=x+ \frac{1}{\sqrt{\lambda _k}}\varphi _k\) for \(k\in {\mathbb {N}}\). Then we have \(Tx_k = Tx + {\sqrt{\lambda _k}}\psi _k .\) Thus,
Remark 5
In the above example, let
for \(k\in {\mathbb {N}}\). Note that \(T_k\) is a finite rank bounded linear operator. We may assume, without loss of generality, that \( \lambda _1 \ge \lambda _2\ge \cdots \). Then we have
so that
Hence T is a compact operator (cf. [11]).
By spectral theorem, we know that every compact operator of infinite rank can be represented as in Example 4, where \(\lambda _k\) are the singular values of T (cf. [11]). Thus, the level of ill-posedness of the corresponding operator equation is closely related to the rate of decay of the singular values. \(\lozenge \)
A typical example of a compact operator is the Fredholm integral operator as in the following example (For the proof of the affirmative statements, one may refer, e.g., [11]).
Example 6
Let \(\Omega := [a, b]\) and \(k\in L^2(\Omega \times \Omega )\). For \(x\in L^2(\Omega )\), let
Then \(Tx\in L^2(\Omega )\) and the operator \(T: L^2(\Omega )\rightarrow L^2(\Omega )\) is a compact operator. The following facts are known:
-
(1)
The range of T is closed if and only if \(k(\cdot , \cdot )\) is a degenerate kernel.
-
(2)
If \(k(\cdot , \cdot )\) is degenerate, then T is not one-one.
-
(3)
If \(k(\cdot , \cdot )\) is non-degenerate, then T is of infinite rank hence it is with non-closed range.\(\lozenge \)
In view of the above example, the problem of solving a Fredholm integral equations of the first kind, namely, the equation
with non-degenerate kernel \(k(\cdot , \cdot )\) is ill-posed. Many problems in applications, such as
-
(1)
computerized tomography,
-
(2)
remote sensing,
-
(3)
geological prospecting
appear as a problem of solving a Fredholm integral equation of the first kind, which is ill-posed (cf. [1,2,3]).
Compact operators also appear in the inverse problems related to parameter identification problems in PDE through the compact imbeddings of Sobolev spaces (cf. [8]). For instance, if \(\Omega \) is a bounded open set in \({\mathbb {R}}^N\) for some \(N\in {\mathbb {N}}\), then the inclusion operator from \(H^1_0(\Omega )\) into \(L^2(\Omega )\) is compact.
Ill-posed linear operator equations also appear while dealing with the problem of solving a non-linear operator equation,
where \(F: D\subseteq X\rightarrow Y\) is a non-linear operator between Hilbert spaces. For instance, a linearization of the above equation takes the form
If F is a completely continuous, that is, if image of every bounded set under F is relatively compact, then \( F'(x_0)\) can be shown to be a compact operator (cf. [6]), and hence the linearized equation (1.2) of the non-linear equation (1.2) is also an ill-posed equation. Many inverse problems in PDE, such as parameter identification problems, are nonlinear and ill-posed (cf. [1]), and they can be formulated as a non-linear equation (1.2). For instance, consider the boundary value problem
There are problems in applications where one would like to find the function u on \(\Omega \) knowing the functions q, f and g. Such problems are the forward problems which are generally well-posed. There are also many problems of practical interest where one would like to find the functions f or q from the knowledge of u on \(\Omega \) or g. Such problems are inverse problems which are generally ill-posed. The inverse problems in which one would like to find f from the knowledge of u are called source identification problems, and those problems in which one would like to find q from the knowledge of u are called parameter identification problems. The source identification problem and parameter identification problem can be represented abstractly as
respectively. It can be shown that the source identification problem is a linear problem, whereas the parameter identification problem is non-linear; both are ill-posed, in general [1, 7].
In this paper, we shall be concerned about linear ill-posed problems only.
2 Regularization
Assume for a moment that \(y\in R(T)\), so that Eq. (1.1) has a solution, say x. Suppose \(y^\delta \in Y\) is such that
for some error level \(\delta >0\). As Eq. (1.1) is, in general, ill-posed, we would like to obtain approximations for x using solutions of a well-posed problems. Such a procedure is called a regularization method. More precisely, we have the following definition.
Definition 7
A regularization method for (1.1) involves a family \(\{R_\alpha \}_{\alpha >0}\) of bounded linear operators from Y to X with \(\alpha >0\) such that
for every \(x\in X\), and for each \(x\in X\) and \(\delta >0\), if \(y^\delta \in Y\) is such that \(y^\delta \rightarrow Tx\) as \(\delta \rightarrow 0\), then there is a parameter choice strategy, say \(\alpha :=\alpha _\delta \), depending on \(y^\delta \) and \(\delta \), such that
The family \(\{R_\alpha : \alpha >0\}\) satisfying (2.1) is called a regularization family or simply, a regularization of the operator equation (1.1).\(\lozenge \)
We describe two simple and well-studied regularization methods, namely, Lavrentiev regularization and Tikhonov regularization—one in the context of Banach spaces and the other in the context of Hilbert spaces.
2.1 Lavrentiev regularization
Let X be a Banach space and \(T: X\rightarrow X\) be a bounded linear operator with non-closed range R(T). Throughout this section, we assume the following (cf. [14]).
Assumption 8
For each \(\alpha >0\), \(T+\alpha I\) is bijective and there exists \(M>0\) such that
for all \(\alpha >0.\) \(\lozenge \)
Operators satisfying the above assumption are called weakly sectorial operators (cf. [20]). The class of weakly sectorial operators include:
-
(1)
Positive self adjoint operators on a Hilbert space.
-
(2)
Many of the Abel integral operators (cf. [20]), including
$$\begin{aligned} (Tu)(s):=\int _0^s t^{\beta -1}u(t)dt,\quad t\in [0, a], \end{aligned}$$where \(X=C[0, a]\) with sup-norm.
-
(3)
The multiplication operator
$$\begin{aligned} (T_\varphi u)(s):= \varphi (s)u(s),\quad s\in [-1, 1], \end{aligned}$$where \(\displaystyle \varphi (s):= \left\{ \begin{array}{ll} 0, &{} -1\le s\le 0,\\ s, &{} 0 <s\le 1,\end{array}\right. \) and \(X=C[-1, 1]\) with sup-norm.
Suppose \(y\in R(T)\) and \(x\in X\) is such that
By Assumption 8 on T, for each \(\alpha >0\),
is a well-posed operator equation. Note that
Hence,
where
By Assumption 8, \(\Vert S_\alpha \Vert \le M\). In other words, \(\{S_\alpha : \alpha >0\}\) is a uniformly bounded family.
Clearly, if \(x\in X\) such that \(Tx=y\), then
In view of the above, we specify a condition under which \(S_\alpha x \rightarrow 0\) as \(\alpha \rightarrow 0\). First let us prove the following general result on \(S_\alpha ,\, \alpha >0\).
Theorem 9
The following are true.
-
(1)
\(\Vert S_\alpha T\Vert \le (1+M)\alpha \) for every \(\alpha >0.\)
-
(2)
If R(T) is dense in X, then
-
(a)
\(S_\alpha u \rightarrow 0\) as \(\alpha \rightarrow 0\) for every \(u\in X\), and
-
(b)
T is injective.
-
(a)
Proof
(1) Observe that, for \(\alpha >0\),
Thus, we obtain \(\Vert S_\alpha T\Vert \le (1+M)\alpha .\)
(2) By (1), \(S_\alpha u\rightarrow 0\) for every \(u\in R(T)\). Now, suppose that R(T) is dense in X. Since \((S_\alpha )\) is uniformly bounded and since it converges pointwise to 0 on a dense subspace of a Banach space, by a result in functional analysis (cf. [11]), \(\Vert S_\alpha u\Vert \rightarrow 0\) for every \(u\in X\).
Now, to see that T is injective, let \(x\in X\) be such that \(Tx=0\). Then we have
so that \(x= \alpha (T+\alpha I)^{-1} x= S_\alpha x\rightarrow 0.\) \(\square \)
As a corollary to the above theorem we deduce the following.
Theorem 10
Let \(y\in R(T)\) and let \(x\in X\) be such that \(Tx=y\). Then we have the following.
-
(1)
If \(x\in R(T)\), then x is the unique element in R(T) such that \(Tx=y\), and
$$\begin{aligned} \Vert x-x_\alpha \Vert = O(\alpha ) \end{aligned}$$ -
(2)
If R(T) is dense in X, then x is the unique element in X such that \(Tx=y\), and
$$\begin{aligned} \Vert x-x_\alpha \Vert \rightarrow 0\quad \text{ as }\quad \alpha \rightarrow 0. \end{aligned}$$
Proof
Let \(y\in R(T)\) and let \(x\in X\) be such that \(Tx=y\).
(1) Suppose \(x\in R(T)\). Let \(u\in X\) be such that \(x=Tu\). Then, from (2.4) and from Theorem 9 (1), we have
Hence, \(\Vert x-x_\alpha \Vert = O(\alpha )\).
(2) Suppose R(T) is dense in X. Then, by Theorem 9 (2), we have \(\Vert x-x_\alpha \Vert = \Vert S_\alpha x\Vert \rightarrow 0\) as \(\alpha \rightarrow 0\). Again, by Theorem 9 (2), T is injective. Hence, x is the unique element in X such that \(Tx=y\). \(\square \)
Remark 11
In Theorems 9 and 10, the assumption that R(T) is dense in X can be weakened as follows.
-
(1)
Theorem 9(2) can be replaced by:
\(S_\alpha u \rightarrow 0\) as \(\alpha \rightarrow 0\) for every \(u\in \overline{R(T)}\).
-
(2)
Theorem 10(2) can be replaced by:
If \(x\in \overline{R(T)}\), then x is the unique element in X such that \(Tx=y\), and \(\Vert x-x_\alpha \Vert \rightarrow 0\) as \(\alpha \rightarrow 0.\) \(\lozenge \)
From Theorem 9, we also deduce the following.
Theorem 12
The family \(\{R_\alpha : \alpha >0\}\) of bounded operators \(R_\alpha : X\rightarrow X\) defined by
for \(\alpha >0\) is a regularization family.
Proof
We observe that for \(x\in X\),
By Theorem 9 (1), \(\Vert S_\alpha T\Vert \le (1+M)\alpha \). Hence,
showing that \(\{R_\alpha : \alpha >0\}\) is a regularization family. \(\square \)
Definition 13
Under the Assumption 8 on T, the family of operators \(R_\alpha ,\, \alpha >0, \) defined in Theorem 12, that is,
is called the Lavrentiev regularization of (1.1).\(\lozenge \)
If X is a Hilbert space and T is a positive self-adjoint operator, then the Assumption 8 on T is automatically satisfied. Also, we know that R(T) is dense in \(N(T)^\perp \).
Thus, in this case, Theorem 10 leads to the following result.
Theorem 14
Let X be a Hilbert space and T be a positive, self-adjoint, bounded linear operator on X . Suppose \(y\in R(T)\) . Then there exists a unique \(x\in N(T)^\perp \) such that \(Tx=y\) , and
Further, if \(x\in R(T)\), then
In the above theorem, by T being a positive and self adjoint, we mean that \(\left\langle Tx, x \right\rangle \ge 0\) for all \(x\in X\) and \(T^*=T\), where \(T^*\) is the adjoint of T.
2.2 Error estimates under noisy data
Suppose that the data y is noisy. Thus, for \(\delta >0\) we may have \(y^\delta \in X\) in place of y such that such that
Let \(x_\alpha ^\delta \in X\) be such that
Then we have
so that
Hence,
Thus, we obtain the following theorem.
Theorem 15
Let \(y\in R(T)\) and let \(x\in X\) be such that \(Tx=y\). Then, under the Assumption 8on T, we have the following.
-
(1)
\(\displaystyle \Vert x - x_\alpha ^\delta \Vert \le \Vert x-x_\alpha \Vert + M \frac{\delta }{\alpha }.\)
-
(2)
If R(T) is dense in X and \(\alpha _\delta \) is such that \(\alpha _\delta \rightarrow 0\) and \({\delta }/{\alpha _\delta } \rightarrow 0\) as \(\delta \rightarrow 0,\) then
$$\begin{aligned} \Vert x - x_{\alpha _\delta }^\delta \Vert \rightarrow 0\quad \text{ as }\quad \delta \rightarrow 0. \end{aligned}$$ -
(3)
If \(x\in R(T)\) and \(\alpha _\delta \sim \delta ^{1/2}\), then
$$\begin{aligned} \Vert x - x_{\alpha _\delta } ^\delta \Vert = O( \delta ^{1/2}). \end{aligned}$$
Proof
(1) This is a consequence of (2.7).
(2) Follows from the fact that \(\Vert x-x_\alpha \Vert \rightarrow 0\) as \(\alpha \rightarrow 0\).
(3) \(x\in R(T)\) implies \(\Vert x-x_\alpha \Vert = O(\alpha )\). Hence \(\alpha \sim \delta ^{1/2}\) implies \(\alpha +\frac{\delta }{\alpha } = O(\delta ^{1/2}).\) \(\square \)
2.3 Tikhonov regularization
Suppose X and Y are Hilbert spaces and \(T: X\rightarrow Y\) is a bounded operator. In this case, \(T^*T\) is a positive self adjoint operator. Let \(y\in Y\) be such that
It is known that the above equation has a solution if and only if \(y\in R(T)+R(T)^\perp \), equivalently,
attains its minimum at x, which satisfies (2.8) (cf. [11, 12]).
Definition 16
A solution of (2.8) is called a least-square solution or more appropriately, a least residual norm (LRN) solution of (1.1).\(\lozenge \)
It can be shown that, if there is an LRN-solution x, then there is an LRN-solution of minimal norm, say \(x^\dagger \), and it is given by (cf. [11, 12])
where \(Q: X\rightarrow X\) is the orthogonal projection onto \(N(T)^\perp \). In particular,
One would like to obtain regularized approximations for an LRN solution of minimal norm.
Let \(x_\alpha \in X\) be the unique element such that
equivalently (cf. [12]), \(x_\alpha \) minimizes
From (2.9) and (2.10), we obtain
Thus,
where
Since \(T^*T\) is a positive self-adjoint operator, we have the following theorem (cf. [12]).
Theorem 17
Suppose \(y\in R(T)\) and \(x\in N(T)^\perp \) is the unique element such that \(Tx=y\) . For \(\alpha >0\) , let \(x_\alpha \in X\) be the unique element such that (2.10) is satisfied. Then,
and the operators
define a regularization family. Further, if \(x\in R(T^*T)\), then
2.4 Error estimates under noisy data
Now, let \(y^\delta \in Y\) be such that
and let \(x_\alpha ^\delta \in X\) be such that
Then we have
It can be shown that (cf. [12])
so that
Thus, we obtain the following theorem (cf. [12]).
Theorem 18
Let \(x\in N(T)^\perp \) be such that \(Tx=y\). Then the following results are true.
-
(1)
\(\displaystyle \Vert x - x_\alpha ^\delta \Vert \le \Vert x-x_\alpha \Vert + \frac{\delta }{2\sqrt{\alpha }}.\)
-
(2)
If \(\alpha _\delta \) is such that \(\alpha _\delta \rightarrow 0\) and \(\delta /\sqrt{\alpha _\delta } \rightarrow 0\) as \(\delta \rightarrow 0\), then
$$\begin{aligned} \Vert x - x_{\alpha _\delta }^\delta \Vert \rightarrow 0\quad \text{ as }\quad \delta \rightarrow 0. \end{aligned}$$ -
(3)
If \(x\in R(T^*T)\) and \(\alpha _\delta \sim \delta ^{2/3}\),
$$\begin{aligned} \Vert x - x_{\alpha _\delta }^\delta \Vert \le O( \delta ^{2/3}). \end{aligned}$$
Remark 19
(i) In Theorem 18 (2), we may take \(\alpha _\delta :=\delta ^s\) for some \(0<s<2\) so that \(\alpha _\delta \rightarrow 0\) and \(\delta /\sqrt{\alpha _\delta }=\sqrt{\delta }\rightarrow 0\) as \(\delta \rightarrow 0\).
(ii) The rate \(O(\delta ^{2/3})\) obtained in Theorem 18 (3) is optimal for Tikhonov regularization, in the sense that, for any choice of the regularization parameter \(\alpha _\delta \), the rate \(o(\delta ^{2/3})\) is not possible (cf. [2, 12]).\(\lozenge \)
3 Illustration of source conditions
In the last section, we obtained error estimates for certain regularization methods under the assumption that the solution x belongs to R(T) or \(R(T^*T)\) if the space is a Hilbert space so that it is of the forms
respectively. Such conditions are called source conditions. Let us illustrate these conditions when the operator involved is a a compact operator.
Case 1: Let X be a Hilbert space and \(T:X\rightarrow X\) be a compact positive self-adjoint operator of infinite rank. Then
where \((\lambda _n)\) is a null sequence of positive real numbers and \(\{\varphi _n: n\in {\mathbb {N}}\}\) is an orthonormal basis of \(N(T)^\perp \). Suppose \(x\in N(T)^\perp = \overline{R(T)}\). Then
Thus,
Case 2: Let X and Y be a Hilbert spaces and \(T:X\rightarrow Y\) be a compact operator of infinite rank. Then
where \((s_n)\) is null sequence of positive real numbers, \(\{\varphi _n: n\in {\mathbb {N}}\}\) and \(\{\psi _n: n\in {\mathbb {N}}\}\) are orthonormal bases of \(N(T)^\perp \) and \(N(T^*)^\perp \), respectively. Suppose \(x\in N(T)^\perp = N(T^*T)^\perp = \overline{R(T^*T)}\). Then, as in Case 1,
Analogous to (3.2) and (3.4) we have the following situations: Let \(0<\nu \le 1\). (i) Suppose \(T:X\rightarrow X\) is a compact, positive, self adjoint operator with spectral representation as in (3.1). Then we may define
From this, it follows that,
(ii) Suppose \(T:X\rightarrow Y\) is a compact operator with singular value representation as in (3.3). Then, since \(T^*T: X\rightarrow Y\) is a compact, positive, self adjoint operator, we obtain
From this, it follows that
Clearly, for \(0<\nu <1\), the requirements in (3.5) and (3.6), are milder than the conditions in (3.2) and (3.4), respectively.
Observations: Let \((\mu _n)\) be a null sequence of positive real numbers and let \((\varphi _n)\) be an orthonormal sequence in a Hilbert space H. For \(0<\nu <1\), let
Then:
-
(1)
\(H_\nu \) is a Hilbert space (a generalized Sobolev space) with inner product:
$$\begin{aligned} \left\langle x, y \right\rangle _\nu :=\sum _{n=1}^\infty \frac{1}{\mu _n^{2\nu }}\left\langle x, \varphi _n \right\rangle \left\langle \varphi _n, y \right\rangle . \end{aligned}$$ -
(2)
If \(\nu _1< \nu _2\), then \(H_{\nu _2}\subset H_{\nu _1}\) is a proper inclusion, and it is a compact embedding.
Note that
so that \(\{\mu _n^\nu \varphi _n: n\in {\mathbb {N}}\}\) is an orthonormal basis for \(H_\nu \). Hence, for \(\nu _1< \nu _2\), if \(x\in H_{\nu _2}\), then
Thus, the inclusion operator \({{\mathcal {J}}}: H_{\nu _2}\rightarrow H_{\nu _1}\) is a compact operator with singular values \(\mu _n^{\nu _2-\nu _1}\). For \(\nu >0\),
defines an inner product on H, and its completion \(H_{-\nu }\) is linearly isometric with \((H_\nu )'\), the dual of \(H_\nu \) (cf. [17]).
The triple \((H_\nu ,\, H,\, H_{-\nu })\) is called a Gelfand triple. The family \(\{H_\nu :\, \nu \in {\mathbb {R}}\}\) is a Hilbert scale, a generalization of the Sobolev scale. More generally, a Hilbert scale is defined as follows (cf. [17]):
Definition 20
A Hilbert scale \((X_s)_{s\in {\mathbb {R}}}\) is a family of Hilbert spaces generated by some unbounded closed densely defined positive, self adjoint operator B in X with \(D(B)\subseteq X\) such that
-
(1)
\(X_0=X\),
-
(2)
For \(s>0\), \(X_s\) is the completion of \({{\mathcal {D}}}:= \bigcap _{k=1}^\infty D(B^k)\) w.r.t. the norm
$$\begin{aligned} \Vert u\Vert _s:= \Vert B^su\Vert ,\quad u\in {{\mathcal {D}}}. \end{aligned}$$ -
(3)
For \(s>0\), \(X_{-s}\) is the completion of X w.r.t. the norm
$$\begin{aligned} \Vert u\Vert _{-s} := \sup _{0\not = v\in X_s} \frac{|\left\langle u, v \right\rangle |}{\Vert v\Vert _s}.\quad \quad \quad \lozenge \end{aligned}$$
As in the case of bounded operators, in the above definition, by B being a positive and self adjoint operator, we mean that \(\left\langle Bx, x \right\rangle \ge 0\) for all \(x\in B\) and \(T^*=T\) on D(B), where \(T^*\) is the adjoint of B, which can be defined for closed densely defined operator as well (see, e.g. Nair [11]).
Remark 21
It is known that (cf. [17]), for \(s>0\), \(X_{-s}\) is linearly isometric with the dual of \(X_s\).\(\lozenge \)
4 Error estimates under general source conditions
4.1 General source conditions for Lavrentiev regularization
In this section, we shall consider a more general form of a source condition for Lavrentiev regularization, namely,
for some \(u\in X\), where \(\phi (\lambda ),\, \lambda >0\), is a continuous function such that
Assumption 22
Suppose there exists a constant \(c_0>0\) such that
\(\lozenge \)
Here is a sufficient condition on \(\phi \) to satisfy (4.3).
Lemma 23
Suppose \(\phi \) is concave, monotonically increasing and \(\phi (\lambda )\rightarrow 0\) as \(\lambda \rightarrow 0\) . Then
Proof
For \(\alpha >0\), writing a point in \([0, \phi (\lambda )]\) as
we obtain
Since \(\phi \) is monotonically increasing, we have
Thus, we obtain the required result. \(\square \)
Theorem 24
Suppose x satisfies the source condition as in (4.1) for some \(u\in X\) . Let \(\phi \) be as in Assumption22. Then
Further, if \(y^\delta \in X\) is such that \(\Vert y-y^\delta \Vert \le \delta \) and \(x_\alpha ^\delta \) is the corresponding Lavrentiev regularized solution, then
In particular, taking \(\alpha _\delta := \psi ^{-1}(\delta )\) , where \(\psi \) is the inverse of \(\alpha \mapsto \alpha \phi (\alpha )\) , then
Proof
By (2.4), we have
Hence,
To obtain the last estimate, note that
where \(\psi \) is the the function \(\alpha \mapsto \alpha \phi (\alpha )\). \(\square \)
Remark 25
If T is a positive compact self adjoint operator T with spectral representation
then
Hence, the requirement (4.1) is same as requiring x to belong to the (Hilbert) space
The above general consideration is useful in applications.\(\lozenge \)
4.2 An application
Let us consider the problem of identifying the initial temperature
from the knowledge of the final temperature
where u(s, t) satisfies the one-dimensional heat equation:
We may recall that
where
Thus, the problem is to solve the operator equation
Note that
defines a compact self-adjoint operator on \(L^2[0, \ell ]\) with spectrum
Clearly, for a given \(y\in L^2[0, \ell ]\), the above operator equation
has a solution \(x\in L^2[0, \ell ]\), i.e.,
Let us assume that y satisfies the above condition. Again
A milder assumption on x than the above would be:
Recall that \(\lambda _n:= e^{-\mu _n^2\tau }\) are the non-zero spectral values (eigenvalues) of T. Note that
Thus the (milder) condition (4.5) takes the form:
where
That is, x belongs to the space
with \(\phi \) as in (4.6). Thus, we can achieve the error estimate as in Theorem 23 under the assumption (4.5) on the initial temperature.
4.3 General source conditions for Tikhonov regularization
In the case of Tikhonov regularization, we may assume, in place of (4.1)
for some \(u\in X\). Recall that, in Tikhonov regularization, one solves the equation
in place of
Then, in place of Theorem 23, we have the following theorem (cf. [12]).
Theorem 26
Suppose the LRN solution x satisfies the source condition as in (4.8) for some \(u\in X\) . Then
Further, if \(y^\delta \in X\) is such that \(\Vert y-y^\delta \Vert \le \delta \) and \(x_\alpha ^\delta \) is such that
then
In particular, taking \(\alpha := \psi ^{-1}(\delta )\) , where \(\psi \) is the inverse of \(\alpha \mapsto \sqrt{\alpha }\phi (\alpha )\) , then
4.4 Remarks on parameter choice strategies
The parameter choice strategy we have considered so far are a priori. That is, with some a priori knowledge on the smoothness of the unknown solution.
There are many a posteriori parameter choice strategies available in the literature. Two of the simple strategies are the following:
Morozov’s type discrepancy principle (MDP): \(\alpha \) is chosen such that
Arcangeli’s type discrepancy principle (ADP): \(\alpha \) is chosen such that
The following results are known:
-
(1)
(Groetch and Guacaneme [4]): For Lavrentiev regularization, MDP (4.9) does not lead to a convergent method; whereas ADP (4.10) gives convergence.
-
(2)
(Groetch [2]): For Tikhonov regularization the order \(O(\delta ^{1/2})\) is obtained with \(x\in R(T^*)\) under MDP (4.9), and this rate cannot be improved under MDP.
-
(3)
(Groetch & Schock [5]): For Tikhonov regularization the order \(O(\delta ^{1/3})\) is obtained with \(x\in R(T^*)\) under ADP (4.10), and this order cannot be improved under the condition of \(x\in R(T^*)\).
However, it is known that for Tikhonov regularization the optimal order \(O(\delta ^{2/3})\) is obtained under apriori choice \(\alpha \sim \delta ^{2/3}\). The question of obtaining this rate under ADP was open right from its introduction in 1966.
The above problem was settled positively in 1992 by the author [10]: For Tikhonov regularization the best optimal order \(O(\delta ^{2/3})\) is attained under ADP (4.10) with \(x\in R(T^*T)\). In fact, it was shown that \(O(\delta ^{2/3})\) is attained under a general discrepancy principle:
whenever
4.5 Modification of the methods to obtain better estimates
One of the modifications for obtaining a better method is to modify the Tikhonov regularization, requiring additional smoothness for the solution as well as for the regularized approximations:
Tikhonov regularization using an unbounded operator: In this, one looks for \(x_\alpha \) which minimizes
where \(L: D(L)\subseteq X\rightarrow Z\) is a closed unbounded operator. Equivalently (cf. [18]), solve:
It is known that the operator \(T^*T+\alpha L^*L: D(L^*L)\rightarrow X\) has continuous inverse if
for some \(\gamma >0\) (see, e.g., [18]).
Note that the above condition is satisfied if L is bounded below (that is the case with many of the differential operators).
For obtaining error estimates, we use the following assumption.
Assumption 27
There exist \( a>0, \, b\ge 0,\, c>0,\, d>0\) such that
where \((X_s)_{s\in {\mathbb {R}}}\) is a Hilbert scale.\(\lozenge \)
It can be shown that, if T is injective and if the Hilbert scale is generated by \(B:=(T^*T)^{-1/2}\) and if \(L:=B^b\), then
Under the Assumption 27, we have the following result (cf. [13]):
Theorem 28
If \({\hat{x}}\in D(L)\) and \(\alpha \) chosen according to the Morozov-type discrepancy principle MDP (4.9), then
Note that, larger the b better the rate.
4.6 Further improvements
Here are a few results improving the above results (cf. [15]).
Theorem 29
If \({\hat{x}}\in D(L^*L)\) and \(\alpha \) is chosen according to the Morozov-type discrepancy principle MDP (4.9), then
Theorem 30
If \({\hat{x}}\in D(L^*L)\) , \(L^*L x = (T^*T)^\nu ,\, 0<\nu \le 1\) and \(\alpha \) is chosen according to the Morozov-type discrepancy principle MDP (4.9), then
where \(p:= {2(a\nu +b)}/[{2(a\nu +b)+a}].\)
Recently [16], the condition
for some \(0\le \theta <1\) is considered in place of the completion condition (4.12) and proved the following theorem which includes results in the Hilbert scale setting as special cases:
Theorem 31
Suppose (4.13) is satisfied in place of (4.12).
-
(1)
If \({\hat{x}}\in D(L)\) and \(\alpha \) is chosen according to the Morozov-type discrepancy principle MDP (4.9), then
$$\begin{aligned} \Vert {\hat{x}} - x_\alpha ^\delta \Vert = O(\delta ^\theta ). \end{aligned}$$ -
(2)
If \({\hat{x}}\in D(L^*L)\) and \(\alpha \) is chosen according to the Morozov-type discrepancy principle MDP (4.9), then
$$\begin{aligned} \Vert {\hat{x}} - x_\alpha ^\delta \Vert = O(\delta ^\frac{2\theta }{1+\theta }). \end{aligned}$$
Note that, taking \(\theta = \frac{b}{a+b}\), we obtain results in Theorem 28 and 29 as special cases.
References
Engl, H., M. Hanke, and A. Neubauer. 1996. Regularization of Inverse Problems. Dordrecht: Kluwer.
Groetsch, C.W. 1984. The Theory of Tikhonov Regularization for Fredholm Equations of the First Kind. London: Pitman.
Groetsch, C.W. 1993. Inverse Problems in the Mathematical Sciences. Braunschweig: Friedrich Vieweg und Sohn.
Groetsch, C.W., and J. Guacaneme. 1987. Arcangeli’s method for Fredholm interal equations of the first kind. Proceedings of the AMS 99: 2.
Groetch, C.W., and E. Schock. 1984. Asymptotic convergence rate of arcangeli’s method for iII-posed problems. Applicable Analysis 18: 3.
Joshi, M.C., and R.K. Bose. 1984. Some Topics in Nonlinear Functional Analysis. Oxford: Wiley Eastern Ltd.
Kaltenbacher, B. 2005/06. Parameter Identificationin Partial Differential Equations, Lecture Notes. University of Linz, WS
Kesavan, S. 1989. Topics in Functional Analysis and Applications. New Dehli: New Age International Ltd.
Mahale, Pallavi 2008. Regularization of Nonlinear Ill-Posed Problems under General Source Conditions, Ph.D. Thesis
Nair, M.T. 1992. A generalization of Arcangeli’s Method for ill-posed problems leading to optimal rate. Integral Equations and Operator Theory 15: 1042–1046.
Nair, M.T. 2002. Functional Analysis: A First Course. New Delhi: PHI-Learning. (Fourth Print: 2014).
Nair, M.T. 2009. Linear Operator Equations: Approximation and Regularization. Singapore: World Scientific.
Nair, M.T. 1999. On Morozov’s method for Tikhonov regularization as an optimal order yielding algorithm. Zeitschrift für Analysis und ihre Anwendungen 18 (1): 37–46.
Nair, M.T. 2000. An iterated version of Lavrentiev’s method for ill-posed equations with approximately specified data. Journal of Inverse and Ill-Posed Problems 8 (2): 193–204.
Nair, M.T. 2006. On improving error estimates for Tikhonov regularization using an unbounded operator. Journal of Analysis 14: 143–157.
Nair, M.T. 2015. A Unified Treatment for Tikhonov Regularization Using a General Stabilizing Operator. Analysis and Applications 13: 201–215.
Nair, M.T. 2016. Compact operators and Hilbert scales in ill-posed problems. Mathematics Student 85 (January—-June): 45–61.
Nair, M.T., M. Hegland, and R.S. Anderssen. 1998. A trade-off between regularity and stabilization for Tikhonov regularization. Mathematics of Computetion 66: 193–206.
Nair, M.T., and P. Ravishankar. 2008. Regularized versions of continuous Newton’s method and continuous modified Newton’s method under general source conditions. Numerical Functional Analysis and Optimization 29: 1140–1165.
Plato, R. 1995. Iterative and Other Methods for Linear Ill-Posed Equations. PhD Thesis, Technincal University of Berlin.
Zeidler, E. 1990. Nonlinear Functional Analysis and its Applications 2/B: Nonlinear Monotone Operators. New York: Springer.
Acknowledgements
The author is thankful to the referee for suggesting many improvements, and also for bringing to his notice some typographical errors in the previous drafts of the paper.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The corresponding author states that there is no conflict of interest.
Ethical approval
This article does not contain any studies with human participants or animals performed by any of the authors.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Nair, M.T. Regularization of ill-posed operator equations: an overview. J Anal 29, 519–541 (2021). https://doi.org/10.1007/s41478-020-00263-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s41478-020-00263-9
Keywords
- Ill-posed problems
- Least-square solution
- Regularization
- Hilbert scales
- Backward heat equation
- Source condition
- Lavrentiev
- Tikhonov
- Morozov
- Arcangeli
- Optimal rate