Abstract
Poisson noise removal problems have attracted much attention in recent years. The main aim of this paper is to study and propose an alternating minimization algorithm for Poisson noise removal with nonnegative constraint. The algorithm minimizes the sum of a Kullback-Leibler divergence term and a total variation term. We derive the algorithm by utilizing the quadratic penalty function technique. Moreover, the convergence of the proposed algorithm is also established under very mild conditions. Numerical comparisons between our approach and several state-of-the-art algorithms are presented to demonstrate the efficiency of our proposed algorithm.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Image deblurring under Poisson noise is an important task in various applications such as astronomical images [26, 39], positron emission tomography (PET) [4, 29], and single particle emission computed tomography (SPECT) [18, 31]. In this paper, we focus on the following image reconstruction model:
where \(g\in \mathbb {R}^n\) is the observed image, \(\mathbf {e}:=(1,\ldots ,1)^T\in \mathbb {R}^n\) is a vector of all ones, \(c\ge 0\) is a fixed background, \(A\in \mathbb {R}^{n\times n}\) is a spatial-invariant blur matrix, \(f\in \mathbb {R}^n\) is a column vector concatenated from the original image with size \(m_1\times m_2\) (\(n=m_1m_2\)), and \(\text{ Poisson }(\cdot )\) denotes the degradation by Poisson noise. We remark that A is a matrix of block Toeplitz with Toeplitz blocks when zero boundary conditions are applied, and block Toeplitz-plus-Hankel with Toeplitz-plus-Hankel blocks when Neumann boundary conditions are used [27, 28].
An approach to deal with the images corrupted by Poisson noise is the maximum a posterior (MAP) estimate [5, 7, 16, 20, 49]. Assume that the observed data \(g_i\) are independent random variables, then the probability distribution of g is the product of the probability distributions of all pixels [37], which can be described as follows:
where \((Af+c\mathbf {e})_i\) and \(g_i\) denote the ith components of \(Af+c\mathbf {e}\) and g, respectively. By taking the negative logarithm of the probability distribution, the Kullback-Leibler (KL) divergence \(D_{KL}(Af+c\mathbf {e},g)\) of \(Af+c\mathbf {e}\) from g is given by [13]
which is convex, nonnegative, and coercive on the nonnegative orthant. When the constants are omitted, we obtain the following data fidelity term
where the notation \(^T\) denotes the transpose operator and \(\log (\cdot )\) means componentwise. The maximum likelihood estimator of the original image is the minimizer of \(\Phi (f)\). It is known that the Richardson-Lucy (RL) algorithm can be applied to solve this problem [24, 32]. There are also a number of important works on RL algorithm, including [9, 14, 37, 41], to name only a few. However, the RL algorithm converges slowly and yields high noise estimate [43].
To deal with the ill-posed problem, a regularization term should be added to control the noise. Traditional regularization methods include the Tikhonov-like regularization [42] and the total variation (TV) regularization [35]. Although the Tikhonov-like regularization method has the advantage of simple calculations [17], it tends to make images unduly smooth and often fails to adequately save important image attributes such as sharp edges. By contrast, the TV regularization, first proposed by Rudin, Osher, and Fatemi for image denoising in [35] and then extended to image deconvolution in [34], has become very popular to represent the prior of images due to its ability to preserve sharp edges. In addition, Acar et al. [1] analyzed the advantages of the TV regularization over the Tikhonov-like regularization for recovering images including piecewise smooth objects.
Based on the prior information of an image, we use the TV as the regularization term. Then, a discrete version of the TV deblurring problem under Poisson noise is given by [23]
where \(\alpha \) is a regularization parameter to balance the regularization term and the KL divergence term, the nonnegative constraint \(f\ge 0\) guarantees that no negative intensities occur in the restored images, \(\Vert \cdot \Vert \) denotes the Euclidean norm of a vector, and \(D_if\in \mathbb {R}^2\) denotes the first-order finite difference of f at pixel i in both horizontal and vertical directions. More specifically, let f be extended by periodic boundary conditions, where \(f:=(f_1,f_2,\ldots , f_n)^T\). Then, for \(i=1,\ldots , n\), \(D_if\) is defined as
where
and
Here \(mod (i,m_2)\) denotes the remainder of i divided by \(m_2\).
The computational challenges of this model are that the KL divergence is non-quadratic and contains a logarithm function. The algorithms of image restoration based on the quadratic data fitting term cannot be applied to this model directly. Moreover, since the TV term is non-differentiable, it is difficult to solve the Euler-Lagrange equation associated with the minimization model (1.2). Recently, some efficient algorithms are proposed to solve the Poisson noise removal problem. Le et al. [23] proposed a gradient descent method to solve a variational image restoration model under Poisson noise. It is well known that the convergence speed of the gradient descent method is very slow. Moreover, Figueiredo et al. [15] proposed an augmented Lagrangian (PIDAL) method for deconvolving Poissonian images. But it needs to solve a TV denoising subproblem under their splitting strategy, which cannot be solved exactly. Furthermore, Setzer et al. [36] proposed an alternating split Bregman algorithm (called PIDSplit+) for the restoration of blurred images corrupted with Poissonian noise. Although the resulting subproblems have closed-form solutions, this algorithm converges slowly after reaching a low accuracy.
In this paper, we focus on developing a fast algorithm to solve the constrained minimization model (1.2). By putting the constrained set into the objective function, we obtain the following unconstrained optimization problem
where \(\delta _{\mathbb {R}_+^{n}}(\cdot )\) is the indicator function over \(\mathbb {R}_+^{n}\). Let \(u\in \mathbb {R}^n\) be an auxiliary variable that approximates \(Af+c\mathbf {e}\) in (1.3). Similarly, we introduce \(w_i\in \mathbb {R}^2\) to approximate \(D_if\) and z to approximate f. By adding the quadratic terms to penalize the differences between every pair of original and auxiliary variables, we obtain the following problem
where \(\beta _1, \beta _2, \beta _3\) are positive penalty parameters, respectively. It is clear that when \(\beta _1,\beta _2,\beta _3\) are sufficiently large, the objective function in (1.4) is close to that in (1.3).
An alternating minimization algorithm is proposed to solve the model (1.4). Alternating minimization algorithms have been applied to the cases of recovering blurred images corrupted by impulse, Gaussian, and multiplicative noises successfully [19, 21, 22, 44, 46,47,48]. For more details about multiplicative noise removal, the interested readers are referred to the papers [2, 38, 40]. Since u, z, w are decoupled, we can minimize f and (u, z, w), alternately. When \(\beta _1,\beta _2,\beta _3\) are sufficiently large, the minimizer of (1.4) is close to that of (1.3). The main advantages of the proposed algorithm are that the resulting subproblems have closed-form solutions or can be solved by fast Fourier transforms [27, 28]. Moreover, based on the fixed point iteration [8, 30], we show that, for any fixed \(\beta _1,\beta _2,\beta _3\), the sequence generated by the alternating minimization algorithm converges to a solution of (1.4). Our numerical results show that the proposed algorithm is faster than PIDAL [15], PIDSplit+ [36], and the primal dual (PD) algorithm [45].
The remaining parts of this paper are organized as follows. In Sect. 2, we give some notation that will be used throughout this paper. The alternating minimization algorithm is presented in Sect. 3. We establish the convergence of the proposed algorithm in Sect. 4. In Sect. 5, some numerical experiments are presented to demonstrate the efficiency of our proposed algorithm. Finally, we give some concluding remarks in Sect. 6.
2 Preliminaries
In this section, we briefly introduce some notation that will be used throughout this paper. \(\mathbb {R}^n\) denotes the n-dimensional Euclidean space. \(\mathbb {R}_+^{n}\) denotes the nonnegative orthant of \(\mathbb {R}^n\), i.e., \(\mathbb {R}_+^{n}:=\{x=(x_1,\ldots ,x_n)^T\in \mathbb {R}^n|x_i\ge 0, i=1, \ldots , n\}\). \(\mathbb {R}_{++}^{n}\) denotes the positive orthant of \(\mathbb {R}^{n}\), i.e., \(\mathbb {R}_{++}^{n}:=\{x=(x_1,\ldots ,x_n)^T\in \mathbb {R}^n|x_i>0, i=1,\ldots ,n\}\). The set of all \(m\times n\) matrices with real entries is denoted by \(\mathbb {R}^{m\times n}\). For any matrix \(A\in \mathbb {R}^{m\times n}\), \(\Vert A\Vert \) denotes the spectral norm of A.
Let \(D^{(1)}\) and \(D^{(2)}\) be the first-order finite difference matrices in the horizontal and vertical directions, respectively. \(D_i\in \mathbb {R}^{2\times n}\) is a matrix formed by stacking the ith rows of \(D^{(1)}\) and \(D^{(2)}\). Let \(D:=(D_1^T,\ldots ,D_n^T)^T\). \(A\succeq B\) denotes that \(A-B\) is positive semidefinite. The identity matrix is denoted by I, whose dimension should be clear from the context. The relative interior of a set G is denoted by \(\text{ int }(G)\). The indicator function of a set G is defined as
Next, we give the definition of a convex function [33, Definition 2.1].
Definition 2.1
-
(a)
A subset C of \(\mathbb {R}^n\) is convex if it includes for every pair of points the line segment that joins them, or in other words, if for every choice of \(x_0\in C\) and \(x_1\in C\) one has \([x_0,x_1]\in C\):
$$\begin{aligned} (1-\tau )x_0+\tau x_1\in C \ \ for \ \ all \ \ \tau \in (0,1). \end{aligned}$$ -
(b)
A function \(\psi \) on a convex set C is convex relative to C if for every choice of \(x_0\in C\) and \(x_1\in C\) one has
$$\begin{aligned} \psi ((1-\tau )x_0+\tau x_1)\le (1-\tau )\psi (x_0)+\tau \psi (x_1)\ \ for \ \ all \ \ \tau \in (0,1). \end{aligned}$$
The effective domain of a function \(\psi : \mathbb {R}^n\rightarrow \mathbb {R}\cup \{+\infty \}\) is defined as
We say that \(\psi \) is proper if it never equals \(-\infty \) and \(\text{ dom }(\psi )\ne \varnothing \). For any proper convex function \(\psi : \mathbb {R}^n\rightarrow \mathbb {R}\cup \{+\infty \}\) and any point \(x\in \text{ dom }(\psi )\), the subgradient of \(\psi \) at x is defined as
Let \(\psi :\mathcal {X}\rightarrow \mathbb {R}\cup \{+\infty \}\) be a closed, proper, convex function, where \(\mathcal {X}\) is a real finite dimensional Euclidean space equipped with an inner product \(\langle \cdot ,\cdot \rangle \) and its induced norm \(\Vert \cdot \Vert \). The operator
is called the proximal mapping of \(\psi \) [12].
Other notation will be defined in appropriate sections if necessary.
3 An Alternating Minimization Algorithm
In this section, we present an alternating minimization algorithm to solve (1.4). Alternating minimization algorithms have been applied to the cases of recovering blurred images corrupted by impulse, Gaussian, and multiplicative noises successfully [19, 21, 22, 44, 46,47,48]. The alternating minimization algorithm for solving (1.4) can be described in Algorithm 1.
Since the objective function in (3.1) is differentiable, \(f^{k+1}\) can be computed by solving the following equation:
where \(w^k:=((w_1^k)^T,\ldots ,(w_n^k)^T)^T\). Let \(M:=\beta _1A^TA+\beta _2D^TD +\beta _3 I\), then M is positive definite. Thus, we have
Remark 3.1
In image restoration problems, A is usually a blurring matrix and D is a difference matrix under period boundary conditions that can be diagonalized by fast Fourier transforms. The computational cost of the method is dominated by three fast discrete transforms for solving the linear system in (3.5) [27, 28]. The interested readers are refered to [27, 28] for a detailed discussion on how to handle blurring matrices with other boundary conditions.
Since A is a blurring matrix, the entries of A are nonnegative, and \(c\ge 0\). Notice that, for any \(u^0>0\), we have \(u^k>0\). Therefore, the closed-form solution for (3.2) is given by
Here, the root square of a vector means componentwise.
As for the subproblem about z in (3.3), we have
where \(\mathcal {P}\) is the projection onto \(\mathbb {R}_+^n\), i.e.,
As for the subproblem about \(w_i\) in (3.4), by the well-known two-dimensional shrinkage [3], \(w_i^{k+1}\) is given by
where \(0\cdot (0/0)=0\) is assumed.
Remark 3.2
Algorithm 1 could be applied to both the isotropic and anisotropic TV. For simplicity, we will only discuss the isotropic case in detail since it is similar to deal with the anisotropic case.
Remark 3.3
Compared with PIDAL [15], the subproblems of the alternating minimization algorithm can be solved exactly in each iteration.
4 Convergence Analysis
In this section, we establish the convergence of Algorithm 1. The following results in [8, 30] are utilized to show that the iteration sequence \(\{(f^k,u^k,z^k,w^k)\}\) generated by Algorithm 1 converges to a minimizer \((f^*,u^*,z^*,w^*)\) of \(\mathcal {Q}\).
Theorem 4.1
Let \(\mathcal {X}\) be a Hilbert space, and let \(\mathcal {T}\) be a nonexpansive self-mapping of \(\mathcal {X}\) such that T is asymptotically regular, i.e., for any x in \(\mathcal {X}\), \(\Vert \mathcal {T}^{n+1}x-\mathcal {T}^nx\Vert \) tends to 0 as n tends to \(+\infty \). Assume that \(\mathcal {T}\) has at least one fixed point. Then, for any x in \(\mathcal {X}\), a weak limit of a weakly convergent subsequence of the sequence of successive approximations \(\{\mathcal {T}^nx\}\) is a fixed point of \(\mathcal {T}\).
Although Theorem 4.1 shows the general case of the fixed point iteration in Hilbert spaces, we only use the finite dimension case for the sequence generated by Algorithm 1. Thus the weak and strong convergence coincide for Algorithm 1, i.e., the limit point of \(\{\mathcal {T}^nx\}\) is a fixed point of \(\mathcal {T}\). For the finite dimensional case about this theorem, a simplified proof can be found in [10, Theorem 5]. When \(\mathcal {X}\) is \(\mathbb {R}^n\), the sequence of successive approximations can converge to a fixed point. Therefore, we only need to prove that the operator \(\mathcal {T}\) about the iteration sequence is nonexpansive, asymptotically regular, and has a fixed point.
For the proximal operator of a function, we have the following proposition [12, Lemma 2.4].
Proposition 4.1
Let \(\phi :\mathcal {X}\rightarrow \mathbb {R}\cup \{+\infty \}\) be a closed, proper and convex function. Then, for any \(x,y\in \mathcal {X}\),
In the following, we define the nonexpansive operator [6], which is important for the analysis of the convergence.
Definition 4.1
An operator \(\mathcal {T}\) is called nonexpansive in \(\mathcal {X}\) if for any \(x_1,x_2\in \mathcal {X}\), we have
By Proposition 4.1, we obtain that \(\text{ Prox }_\phi \) is nonexpansive, i.e.,
Actually, the proximal operator is a special kind of an averaged operator, which is nonexpansive [10, Lemma 3]. Next, we construct the iteration sequence of Algorithm 1. Let
then \(M=H^TH\) by the definition of M. Thus, (3.5) is equivalent to
where
Since M is positive definite, we obtain
Let \(\varphi (u):=\mathbf {e}^Tu-g^T\log (u)\). Since \(u>0\) and \(Af+c\mathbf {e}>0\), by (3.2), we get
Moreover, by (3.3), we have
Note that, for \(w_i\), the minimizer is the two-dimensional shrinkage operator [3]. Let \(g(x)=\Vert x\Vert \), by (3.8), we obtain that
Thus, combining (4.2), (4.3), and (4.4), we get
Let
we have \(v^{k+1}=\mathcal {T}(v^k)\), which is the iteration of Algorithm 1. In the following, we show that \(\mathcal {T}\) is nonexpansive.
Lemma 4.1
The operator \(\mathcal {T}\) in (4.5) is nonexpansive.
Proof
For any \(v_1,v_2\), by the definition of \(\mathcal {T}\), we have
Let \(f_1:=\mathcal {T}_f(v_1)\) and \(f_2:=\mathcal {T}_f(v_2)\), then, by the definition of \(\mathcal {T}_f\), we obtain that
Thus, we have
where the inequality follows from Proposition 4.1 and the last equality follows from (4.7). For the operator \(\mathcal {T}_2(\mathcal {T}_f)\), we get
where the inequality holds since \(\mathcal {T}_2\) is nonexpansive. For the operator \(\mathcal {T}_3(\mathcal {T}_f)\), we have
where the inequality holds by Proposition 4.1. Therefore, combining (4.8), (4.9), (4.10) and (4.6), we obtain that
where the last equality follows from the definition of M. This implies that \(\Vert \mathcal {T}(v_1)-\mathcal {T}(v_2)\Vert \le \Vert v_1-v_2\Vert \). Thus, \(\mathcal {T}\) is nonexpansive. \(\square \)
The following lemma shows that \(\mathcal {T}\) is asymptotically regular.
Lemma 4.2
Let \(v^k:=((u^k)^T,(z^k)^T,(w^k)^T)^T\) be generated by Algorithm 1, then \(\sum _{k=1}^{+\infty }\Vert v^k-v^{k-1}\Vert ^2\) converges and \(\mathcal {T}\) is asymptotically regular.
Proof
The proof follows the lines of the proof of Lemma 3.6 in [19]. For the sake of completeness, we give it here. We consider the Taylor expansion of \(\mathcal {Q}(f,u^k,z^k,w^k)\) at \(f^{k+1}\) and let \(f=f^k\). Since \(\mathcal {Q}(f,u^k,z^k,w^k)\) about f is a quadratic function, we have
where \(\nabla _f\mathcal {Q}(f^{k+1},u^k,z^k,w^k)\) and \(\nabla _f^2\mathcal {Q}(f^{k+1},u^k,z^k,w^k)\) denote the gradient and Hessian matrix of \(\mathcal {Q}(f,u^k,z^k,w^k)\) at \(f^{k+1}\), respectively. Since \(f^{k+1}\) is a minimizer of \(\mathcal {Q}(f,u^k,z^k,w^k)\), we have
Moreover,
Thus, by substituting (4.12) and (4.13) into (4.11), we get
Since \(\mathcal {Q}(f^{k+1},u^{k+1},z^{k+1},w^{k+1})\le \mathcal {Q}(f^{k+1},u^k,z^k,w^k)\), we obtain
Combining (4.2), (4.3) and (4.4), we have
where the first inequality holds by Proposition 4.1. From [11], we know that \(\Vert D\Vert ^2\le 8\) and from [25], we have \(\Vert A\Vert \le 1\). Thus,
Substituting (4.15) into (4.14), we have
Since \(\mathcal {Q}\) is bounded below, it follows that \(\sum _{k=0}^{+\infty }\Vert v^{k}-v^{k+1}\Vert ^2\) converges. Moreover,
Note that \(v^{k}=\mathcal {T}(v^{k-1})=\mathcal {T}^2(v^{k-2})=\cdots =\mathcal {T}^{k-1}(v^{1})=\mathcal {T}^k(v^0),\) we obtain that
This implies that \(\mathcal {T}\) is asymptotically regular. Thus, we complete the proof. \(\square \)
By [45], we know that \(\mathbf {e}^Tu-g^T\log (u)\) is proper, which implies that \(\mathcal {Q}(f,u,z,w)\) is proper. Notice that \(\mathcal {Q}(f,u,z,w): \mathbb {R}^n\times \mathbb {R}^n_{++}\times \mathbb {R}^n\times \mathbb {R}^{2n}\rightarrow \mathbb {R}\cup \{+\infty \}\) is continuous, we obtain that \(\mathcal {Q}(f,u,z,w)\) is closed [33, Thoerem 1.6]. In order to show the existence of minimizers of \(\mathcal {Q}(f,u,z,w)\), by [33, Theorem 1.9], we only need to prove that \(\mathcal {Q}(f,u,z,w)\) is coercive, i.e., \(\mathcal {Q}(f,u,z,w)\) tends to infinity as \(\Vert (f^T,u^T,z^T,w^T)^T\Vert \) tends to infinity.
Lemma 4.3
If \(Ker (A)\cap Ker (D)=\{0\}\), where \(Ker (\cdot )\) denotes the null space of a matrix, then \(\mathcal {Q}(u,z,w,f)\) is coercive.
Proof
When
we have \(\Vert u\Vert \rightarrow +\infty \), or \(\Vert w\Vert \rightarrow +\infty \), or
Thus, we proceed the discussions by three cases.
Case 1. Suppose that \(\Vert u\Vert \rightarrow +\infty \). We note that \(u-\log (u)\) tends to infinity when u tends to infinity. Thus, we have \(\mathcal {Q}(f,u,z,w)\) tends to infinity.
Case 2. Suppose that \(\Vert w\Vert \rightarrow +\infty \). Then, \(\sum _{i=1}^n\Vert w_i\Vert \) tends to infinity. Since \(u-\log (u)\) is bounded below, we obtain that \(\mathcal {Q}(f,u,z,w)\) tends to infinity.
Case 3. Suppose that
We assume that \(\Vert u\Vert , \Vert w\Vert \) are bounded, otherwise the result holds by Cases 1 and 2. If \(\Vert f\Vert \) tends to infinity and \(\Vert z\Vert \) is bounded, we have \(\Vert z-f\Vert ^2\) tends to infinity. Thus, \(\mathcal {Q}(f,u,z,w)\) tends to infinity. If \(\Vert z\Vert \) tends to infinity and \(\Vert f\Vert \) is bounded, we have \(\Vert z-f\Vert ^2\) tends to infinity. Then, \(\mathcal {Q}(f,u,z,w)\) tends to infinity. If \(\Vert f\Vert \) tends to infinity, \(\Vert z\Vert \) tends to infinity, since \(\text{ Ker }(A)\cap \text{ Ker }(D)=\{0\}\), we have \(\Vert Af\Vert \) tends to infinity or \(\Vert Df\Vert \) tends to infinity as \(\Vert f\Vert \) tends to infinity. Since \(\Vert u\Vert , \Vert w\Vert \) are bounded, we obtain that \(\mathcal {Q}(f,u,z,w)\) tends to infinity. Together with Cases 1 and 2, we complete the proof. \(\square \)
Remark 4.1
The condition \(\text {Ker}(A)\cap \text {Ker}(D)=\{0\}\) in Lemma 4.3 is easy to satisfy. If A is a blurring matrix, then their rows sum up to one, while the differences matrix D (under appropriate boundary conditions) has kernel consisting of constant vectors so that their joint kernel must be the zero vector.
We now show that the set of fixed points of \(\mathcal {T}\) is nonempty.
Lemma 4.4
Suppose that \(Ker (A)\cap Ker (D)=\{0\}\). Then the set of fixed points of \(\mathcal {T}\) is nonempty.
Proof
Since \(\mathcal {Q}\) is coercive, by [33, Theorem 1.9], we obtain that the minimizer \((f^*,u^*,z^*,w^*)\) of \(\mathcal {Q}\) exists. Then, by [33, Thoerem 6.12], we get
where \( N_{\mathbb {R}_{++}^n}(u^*)\) denotes the normal cone of \(\mathbb {R}_{++}^n\) at \(u^*\). Since \(u^*>0\), i.e., \(u^*\in \text{ int }(\mathbb {R}_{++}^n)\), by [33, Proposition 6.5], we have \(N_{\mathbb {R}_{++}^n}(u^*)=\{0\}\). This implies that
where \(v^*:=((u^*)^T,(z^*)^T,(w^*)^T)^T\). Thus, by the definition of \(\mathcal {T}\), we obtain that \(v^*=\mathcal {T}(v^*)\), and \(v^*\) is a fixed point of \(\mathcal {T}\). \(\square \)
Combining Lemmas 4.1, 4.2, 4.4 and Theorem 4.1, we can get the following convergence result.
Theorem 4.2
Suppose that \(Ker (A)\cap Ker (D)=\{0\}\). For any initial point \(u^0>0,z^0,w^0\), the sequence \(\{(f^k,u^k,z^k,w^k)\}\) generated by Algorithm 1 converges to a minimizer \((f^*,u^*,z^*,w^*)\) of \(\mathcal {Q}\).
Remark 4.2
Compared with the alternating minimization algorithms in [19, 21, 22], we consider some constraints in Algorithm 1.
5 Experimental Results
In this section, we present numerical results to show the efficiency of the proposed algorithm for the images corrupted by blurs and Poisson noise. The testing images include Cameraman (\(256\times 256\)), House (\(256\times 256\)), Barbara (\(512\times 512\)), and Livingroom (\(512\times 512\)), which are shown in Fig. 1. In our numerical examples, we consider two blurring functions from MATLAB: Average blur and Gaussian blur. For Average blur, the testing kernel is \(9\times 9\), and for Gaussian blur, the testing kernel is \(9\times 9\) (standard deviation \(= 3\)). All the experiments are performed under Windows 7 and MATLAB R2015b running on a laptop (AMD A10-5750M, @ 2.50GHz, 8G RAM).
Since Poisson noise is data-dependent, the noise levels of the observed images depend on the pixel intensity M. To test different noise levels, we consider different peak intensities of the images. The noisy and blurry images in our test are simulated as follows. The original image f is first scaled with the peak intensities. Then the scaled image is convolved with the blur kernel A and the background is added. Finally, the Poisson noise is added in Matlab using the function poissrnd.
To evaluate the performance of different algorithms, the signal-to-noise ratio (SNR) is used to measure the quality of the restored image:
where f is the original image and \(\hat{f}\) is the restored image. We terminate Algorithm 1 by checking whether the relative error of the successive iterates satisfies
where \(\text{ tol }\) is set to be \(10^{-3}\) in our experiments. The maximum iteration number of Algorithm 1 is set to be 500. For the penalty parameters \(\beta _1, \beta _2, \beta _3\) in Algorithm 1, the initial values of \(\beta _1, \beta _2, \beta _3\) are set to 0.4, 0.4 and 0.001, respectively, and then \(\beta _1, \beta _2, \beta _3\) are increased by \(\tau \beta _1, \tau \beta _2, \tau \beta _3\) every twenty iterations, where \(\tau =1.01\) in our experiments. The regularization parameter \(\alpha \) will be selected from \(\{1,10, 20, 50, 100, 150, 200, 250, 300\}\) for the fast convergence and satisfactory accuracy. We remark that \(\beta _1,\beta _2,\beta _3\) should make the minimizer of (1.4) close to the minimizer of (1.3) in the quadratic penalty (1.4). Unfortunately, large \(\beta _i,i=1,\ldots ,3\), usually cause numerical difficulties, since (1.4) becomes very ill-conditioned. Hence, it is more practical to minimize (1.4) with moderate values of \(\beta _i,i=1,\ldots ,3\).
5.1 Comparisons to the Existing Methods
In this subsection, we compare our algorithm with PIDAL [15], PIDSplit+ [36], and PD [45]. The stopping criterion of PIDAL, PIDSplit+, and PD is the same as (5.1) and the maximum iteration number of the three algorithms is set to be 1000. For the PIDAL algorithm [15], the Chambolle’s projection algorithm [11] is used to solve the TV denoising subproblem. The number of iterations of Chambolle’s algorithm is set to 20 and the corresponding step size is set to 0.248, which were suggested in [15]. We use the final values of an inner iteration loop as the initial values for the next loop. Moreover, as suggested in [15, 36], we set \(\gamma =50/\alpha \) for the PIDAL and PIDSplit+ algorithms, where \(\alpha \) is the regularization parameter and \(\gamma \) is the penalty parameter. In the PIDAL, PIDSplit+, and PD algorithms, the parameter \(\alpha \) is also tested from \(\{1, 10, 20, 50, 100, 150, 200, 250, 300\}\) to get the best recovery performance in terms of SNR values and CPU time.
In Table 1, we show the SNR values and CPU time (in seconds) for the Cameraman, House, Barbara, and Livingroom images corrupted by Average blur and Poisson noise with different peak intensities M and c, respectively. We can see that our proposed algorithm outperforms PIDAL, PIDSplit+, and PD for these images in terms of CPU time, and the SNR values of the proposed algorithm are almost the same as those of the other algorithms. Therefore, under almost the same accuracies, our proposed algorithm is much faster than the other three algorithms. Moreover, the PIDAL and PIDSplit+ algorithms need more CPU time than PD for these images.
The visual comparison of the Cameraman, House, Barbara, and Livingroom images corrupted by Average blur and Poisson noise with peak intensity \(M=80\) and \(c=10\) is shown in Fig. 2. We can observe that these images restored by Algorithm 1 are almost the same as those restored by the PIDAL, PIDSplit+, and PD algorithms. However, the CPU time of the proposed algorithm is much less than those of the other three algorithms. Moreover, one can hardly see the differences among these images restored by the four algorithms. In Fig. 3, we show the difference images of the House image of Fig. 2 for different algorithms. It is shown that there are still some differences in these images restored by PIDAL, PIDSplit+, PD, and Algorithm 1.
Table 2 shows the SNR values and CPU time for the Cameraman, House, Barbara, and Livingroom images corrupted by Gaussian blur and Poisson noise with different peak intensities M and c, respectively. It can be seen that the CPU time of the proposed algorithm is less than the other three algorithms. The accuracies of the proposed algorithm are almost the same as those of PIDAL, PIDSplit+, and PD in terms of SNR values. Moreover, the PD algorithm performs better than PIDAL and PIDSplit+ in terms of CPU time.
Figure 4 shows the comparison on the Cameraman, House, Barbara, and Livingroom images corrupted by Gaussian blur and Poisson noise with peak intensity \(M=40\) and \(c=1\). We can observe that the SNR values of the restored images by the four algorithms are almost the same, and the computational time required by the proposed algorithm is much less than those required by PIDAL, PIDSplit+, and PD.
In Fig. 5, we plot the SNR values versus CPU time for the Cameraman, House, Barbara, and Livingroom images corrupted by Gaussian blur and Poisson noise with \(M=20, c=1\). We can observe that the PIDAL and PIDSplit+ algorithms are fast at the initial iterations. But the convergence speed of the two algorithms is slow after reaching a low accuracy. Algorithm 1 needs less CPU time than PIDAL, PIDSplit+, and PD for reaching high accurate solutions. Moreover, the PD algorithm is slower than the proposed algorithm for all the stages. It needs more iterations to reach the same accuracy than the proposed algorithm. The proposed algorithm is fast in all the iterations for low and high accurate solutions. Thus, Algorithm 1 outperforms PIDAL, PIDSplit+ in terms of CPU time.
6 Concluding Remarks
In this paper, we have proposed an alternating minimization algorithm for the restoration of blurred images corrupted by Poisson noise, which minimizes the KL divergence plus TV regularization regularization model with nonnegative constraint. By introducing new variables, the objective function is separable. The approximation model (1.4) is derived from the classical penalty method. Then, we can minimize each variables, alternatingly. Furthermore, based on the fixed point theory, we establish the convergence of the proposed algorithm under very mild assumptions. Our preliminary numerical experiments show that the proposed algorithm is much faster than PIDAL [15], PIDSplit+ [36] and PD [45].
References
Acar, R., Vogel, C.R.: Analysis of bounded variation penalty methods for ill-posed problems. Inverse Probl. 10(6), 1217–1229 (1994)
Aubert, G., Aujol, J.-F.: A variational approach to removing multiplicative noise. SIAM J. Appl. Math. 68(4), 925–946 (2008)
Bai, M., Zhang, X., Shao, Q.: Adaptive correction procedure for TVL1 image deblurring under impulse noise. Inverse Probl. 32(8), 085004 (2016)
Bailey, D.L., Townsend, D.W., Valk, P.E., Maisey, M.N.: Positron Emission Tomography-Basic Sciences. Springer, New York (2005)
Bardsley, J.M., Goldes, J.: An iterative method for edge-preserving MAP estimation when data-noise is Poisson. SIAM J. Sci. Comput. 32(1), 171–185 (2010)
Bauschke, H.H., Combettes, P.L.: Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Springer, New York (2011)
Bertero, M., Boccacci, P., Desiderà, G., Vicidomini, G.: Image deblurring with Poisson data: from cells to galaxies. Inverse Probl. 25(12), 123006 (2009)
Browder, F.E., Petryshyn, W.V.: The solution by iteration of nonlinear functional equations in Banach spaces. Bull. Am. Math. Soc. 72(3), 571–575 (1966)
Brune, C., Sawatzky, A., Burger, M.: Primal and dual Bregman methods with application to optical nanoscopy. Int. J. Comput. Vis. 92(2), 211–229 (2011)
Burger, M., Sawatzky, A., Steidl, G.: First order algorithms in variational image processing. In: Glowinski, R., Osher, S., Yin, W. (eds.) Splitting Methods in Communication, Imaging, Science, and Engineering, pp. 345–407. Springer, Switzerland (2016)
Chambolle, A.: An algorithm for total variation minimization and applications. J. Math. Imaging Vis. 20(1), 89–97 (2004)
Combettes, P., Wajs, V.: Signal recovery by proximal forward-backward splitting. Multiscale Model. Simul. 4(4), 1168–1200 (2005)
Csiszar, I.: Why least squares and maximum entropy? An axiomatic approach to inference for linear inverse problems. Ann. Stat. 19(4), 2032–2066 (1991)
Dey, N., Blanc-Feraud, L., Zimmer, C., Roux, P., Kam, Z., Olivo-Marin, J.-C., Zerubia, J.: Richardson-Lucy algorithm with total variation regularization for 3D confocal microscope deconvolution. Micros. Res. Tech. 69(4), 260–266 (2006)
Figueiredo, M.A.T., Bioucas-Dias, J.M.: Restoration of Poissonian images using alternating direction optimization. IEEE Trans. Image Process. 19(12), 3133–3145 (2010)
Geman, S., Geman, D.: Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans. Pattern Anal. Mach. Intell. 6(6), 721–741 (1984)
Golub, G.H., Hansen, P.C., O’Leary, D.P.: Tikhonov regularization and total least squares. SIAM J. Matrix Anal. Appl. 21(1), 185–194 (1999)
Green, P.J.: Bayesian reconstructions from emission tomography data using a modified EM algorithm. IEEE Trans. Med. Imaging 9(1), 84–93 (1990)
Guo, X., Li, F., Ng, M.K.: A fast \(\ell \)1-TV algorithm for image restoration. SIAM J. Sci. Comput. 31(3), 2322–2341 (2009)
Hohage, T., Werner, F.: Inverse problems with Poisson data: statistical regularization theory, applications and algorithms. Inverse Probl. 32(9), 093001 (2016)
Huang, Y.-M., Lu, D.-Y., Zeng, T.: Two-step approach for the restoration of images corrupted by multiplicative noise. SIAM J. Sci. Comput. 35(6), A2856–A2873 (2013)
Huang, Y.-M., Ng, M.K., Wen, Y.-W.: A fast total variation minimization method for image restoration. Multiscale Model. Simul. 7(2), 774–795 (2008)
Le, T., Chartrand, R., Asaki, T.J.: A variational approach to reconstructing images corrupted by Poisson noise. J. Math. Imaging Vis. 27(3), 257–263 (2007)
Lucy, L.B.: An iterative technique for the rectification of observed distributions. Astron. J. 79(6), 745–754 (1974)
Ma, L., Ng, M.K., Yu, J., Zeng, T.: Efficient box-constrained TV-type-\(\ell ^1\) algorithms for restoring images with impulse noise. J. Comput. Math. 31(3), 249–270 (2013)
Molina, R.: On the hierarchical Bayesian approach to image restoration: applications to astronomical images. IEEE Trans. Pattern Anal. Mach. Intell. 16(11), 1122–1128 (1994)
Ng, M.K.: Iterative Methods for Toeplitz Systems. Oxford University Press, London (2004)
Ng, M.K., Chan, R.H., Tang, W.-C.: A fast algorithm for deblurring models with Neumann boundary conditions. SIAM J. Sci. Comput. 21(3), 851–866 (1999)
Ollinger, J.M., Fessler, J.A.: Positron-emission tomography. IEEE Signal Process. Mag. 14(1), 43–55 (1997)
Opial, Z.: Weak convergence of the sequence of successive approximations for nonexpansive mappings. Bull. Am. Math. Soc. 73(4), 591–597 (1967)
Panin, V.Y., Zeng, G.L., Gullberg, G.T.: Total variation regulated EM algorithm. IEEE Trans. Nuclear Sci. 46(6), 2202–2210 (1999)
Richardson, W.H.: Bayesian-based iterative method of image restoration. J. Opt. Soc. Am. A 62(1), 55–59 (1972)
Rockafellar, R.T., Wets, R.J.-B.: Variational Analysis. Springer, Berlin (2009)
Rudin, L.I., Osher, S.: Total variation based image restoration with free local constraints. In: Proceedings of IEEE International Conference Image Process, vol 1, pp. 31–35. Austin, TX, (1994)
Rudin, L.I., Osher, S., Fatemi, E.: Nonlinear total variation based noise removal algorithms. Physica D 60(1–4), 259–268 (1992)
Setzer, S., Steild, G., Teuber, T.: Deblurring Poissonian images by split Bregman techniques. J. Vis. Commun. Image Rep. 21(3), 193–199 (2010)
Shepp, L.A., Vardi, Y.: Maximum likelihood reconstruction for emission tomography. IEEE Trans. Med. Imaging 1(2), 113–122 (1982)
Shi, J., Osher, S.: A nonlinear inverse scale space method for a convex multiplicative noise model. SIAM J. Imaging Sci. 1(3), 294–321 (2008)
Starck, J.L., Pantin, E., Murtagh, F.: Deconvolution in astronomy: A review. Publ. Astron. Soc. Pac. 114(800), 1051–1069 (2002)
Steidl, G., Teuber, T.: Removing multiplicative noise by Douglas-Rachford splitting methods. J. Math. Imaging Vis. 36(2), 168–184 (2010)
Teuber, T., Steidl, G., Chan, R.H.: Minimization and parameter estimation for seminorm regularization models with I-divergence constraints. Inverse Probl. 29(3), 035007 (2013)
Tikhonov, A.N., Arsenin, V.Y.: Solutions of Ill-Posed Problems. Winston, Washington, DC (1977)
Vio, R., Bardsley, J., Wamsteker, W.: Least-squares methods with Poissonian noise: Analysis and comparison with the Richardson-Lucy algorithm. Astron. Astrophys. 436(2), 741–755 (2005)
Wang, Y., Yang, J., Yin, W., Zhang, Y.: A new alternating minimization algorithm for total variation image reconstruction. SIAM J. Imaging Sci. 1(3), 248–272 (2008)
Wen, Y.-W., Chan, R.H., Zeng, T.: Primal-dual algorithms for total variation based image restoration under Poisson noise. Sci. China Math. 59(1), 141–160 (2016)
Wen, Y.-W., Ng, M.K., Huang, Y.-M.: Efficient total variation minimization methods for color image restoration. IEEE Trans. Image Process. 17(11), 2081–2088 (2008)
Yang, J., Zhang, Y., Yin, W.: An efficient TVL1 algorithm for deblurring multichannel images corrupted by impulsive noise. SIAM J. Sci. Comput. 31(4), 2842–2865 (2009)
Zeng, T., Li, X., Ng, M.K.: Alternating minimization method for total variation based wavelet shrinkage model. Commun. Comput. Phys. 8(5), 976–994 (2010)
Zhang, X., Javidi, B., Ng, M.K.: Automatic regularization parameter selection by generalized cross-validation for total variational Poisson noise removal. Applied Opt. 56(9), D47–D51 (2017)
Acknowledgements
The authors are grateful to the anonymous referees for their careful reading of this paper and their valuable suggestions, which have helped improve the quality of this paper substantially. The authors would also like to thank Professor You-Wei Wen for sending us the code of the primal dual algorithm in [45].
Author information
Authors and Affiliations
Corresponding author
Additional information
Xiongjun Zhang: The research of this author was supported in part by the Fundamental Research Funds for the Central Universities under Grant 230-20205170463-610.
Michael K. Ng: The research of this author was supported in part by the HKRGC GRF 1202715, 12306616, 12200317 and HKBU RC-ICRS/16-17/03.
Minru Bai: The research of this author was supported in part by the National Science Foundation of China under Grant 11571098.
Rights and permissions
About this article
Cite this article
Zhang, X., Ng, M.K. & Bai, M. A Fast Algorithm for Deconvolution and Poisson Noise Removal. J Sci Comput 75, 1535–1554 (2018). https://doi.org/10.1007/s10915-017-0597-2
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10915-017-0597-2