1 Introduction

Image restoration aims at recovering an underlying image u from its observed degradation f. Mathematically, the basic image restoration model is usually formulated as \(f=Ku+n\), with K being a bounded linear blurring (or convolution) operator and n a white additive Gaussian noise with variance \(\sigma ^2\). A classical way to solve this ill-posed inverse problem is to add a regularization, resulting in the following energy functional

$$\begin{aligned} \min \limits _{u}J(u)+\frac{\lambda }{2}\Vert Ku-f\Vert _2^2. \end{aligned}$$
(1)

Here, J(u) denotes the regularization term and \(\lambda \) is a positive regularization parameter.

As is well known, several popular regularization techniques have been developed in image processing. Among them, partial differential equation and wavelet frame regularized methods are two popular approaches, which have been studied extensively and made great successes. One successful example of variational methods is the total variation (TV) [1] regularized model as

$$\begin{aligned} \min \limits _{u}\mathrm{TV}(u)+\frac{\lambda }{2}\Vert Ku-f\Vert _2^2. \end{aligned}$$
(2)

Numerically, solving for the above model, there exist several highly efficient numerical methods. Thereinto, simultaneously recovering the degraded image and estimating the regularization parameter \(\lambda \) adaptively is a challenging subject. Up to now, various approaches have been sprung up for the parameter selection automatically, such as Morozov’s discrepancy principle [2,3,4,5,6], the generalized cross-validation method [7, 8], the L-curve approach [9] and the variational Bayesian method [10, 11]. One thing to be noted is that, when the noise variance is available, Morozov’s discrepancy principle is preferred to achieve the optimal parameter \(\lambda \) adaptively.

This model performs well in preserving important detail features for image denoising. Unfortunately, the numerous staircase effect inevitably emerges due to the TV regularized framework. To overcome this drawback, researchers recently introduced the total generalized variation (TGV) as penalty functional in image processing. More specifically, the second-order TGV with weight \(\alpha \) (\({\mathrm{TGV}}_{\alpha }^2\)) regularized models [12,13,14,15,16,17,18,19] have been achieved extensive research and attention. Thereinto, applied for image restoration, the resulting model is given by

$$\begin{aligned} \min \limits _{u}{\mathrm{TGV}}_{\alpha }^2(u)+\frac{\lambda }{2}\Vert Ku-f\Vert _2^2. \end{aligned}$$
(3)

Calculating for the minimization problem (3), Bredies et al. [14] proposed a spatially dependent regularization parameter selection algorithm based on statistical methods. Later, He et al. [18] introduced an adaptive parameter estimation approach using the discrepancy principle.

Another well-adopted regularizer technique is the wavelet frame-based methods [20,21,22,23,24] with the \(\ell \)1-norm of frame coefficients. This shrinkage method makes the relevant theoretical analyses and calculations easier, and performs well in image processing too. However, the only fly in the ointment is that the Gibbs-like oscillations emerge frequently around the image discontinuities.

Therefore, to better reconstruct the degraded image and simultaneously preserve image features, a new edge-preserving regularization scheme is reported in this work. Namely, by integrating the merits of \({\mathrm{TGV}}_{\alpha }^2\) and wavelet frame transform, and avoiding their main shortcomings, we concentrate on a novel hybrid regularizers model for image restoration. The optimization problem is established in the following form

$$\begin{aligned} \min \limits _{u}{\mathrm{TGV}}_{\alpha }^2(u)+\beta \Vert Wu\Vert _1+\frac{\lambda }{2}\Vert Ku-f\Vert _2^2, \end{aligned}$$
(4)

where W is the wavelet frame transform. It is noteworthy that, applying Morozov’s discrepancy principle, we will present a fast numerical algorithm that can be used to achieve the regularization parameter \(\lambda \) in (4) automatically. Therefore, the concerned image reconstruction problem is formulated as

$$\begin{aligned} \min \limits _{u}{\mathrm{TGV}}_{\alpha }^2(u)+\beta \Vert Wu\Vert _1 \ \ \mathrm{s.t.}\ \ \Vert Ku-f\Vert _2^2\le m^2, \end{aligned}$$
(5)

where \(m^2=\tau n_1n_2\sigma ^2\), with \(\tau \) being a noise-dependent predetermined parameter and \(n_1\times n_2\) the image pixels. Generally, as discussed in [8], one can simply set \(\tau =1\).

Subsequently, let us define the characteristic function \(I_D(u)\) as

$$\begin{aligned} I_D(u)=\left\{ \begin{array}{l@{\quad }l} 0, &{}u\in D\triangleq \{u: \Vert Ku-f\Vert _2^2\le m^2\},\\ +\infty , &{}\mathrm{otherwise}. \end{array} \right. \end{aligned}$$
(6)

Then, the constrained optimization problem (5) can be transferred into an unconstrained one as follows:

$$\begin{aligned} \min \limits _{u}{\mathrm{TGV}}_{\alpha }^2(u)+\beta \Vert Wu\Vert _1+I_D(u). \end{aligned}$$
(7)

Our significant contributions of this article can be summarized as follows. First off, on the basis of the TGV and wavelet frame-based methods, we put forward a new hybrid regularizers model for image restoration. The inclusion of multiple regularizers helps to obtain more accurate and stable numerical solutions. The second important advantage is to develop an extremely efficient alternating minimization algorithm for solving the resulting model without any inner iteration, which simultaneously recovers the degenerated image and adaptively selects the optimal parameter \(\lambda \) using the discrepancy principle.

The outline of this article is generalized as follows. Section 2 gives a summary of the necessary definitions and the basic properties on the proposed model. In Sect. 3, we describe in more detail the alternating minimization method that adaptively updates the regularization parameter in each iteration step. And the convergence proof is also analyzed in Sect. 4 in brief. Numerical results aiming at demonstrating the effectivity of the new algorithm are provided in Sect. 5. Finally, concluding remarks are drawn in Sect. 6.

2 Preliminaries

Our objective in this section is to give a brief introduction and summarize the properties on the model (7). Referring to [12, 13, 17], we begin with the concept of second-order TGV.

Definition 1

Let \({\varOmega }\subset {\mathbb {R}}^{d}\) be a bound domain, and \(\alpha =(\alpha _0,\alpha _1)>0\). Then, the second-order total generalized variation for \(u\in L^1({\varOmega })\) is defined as the value of the functional

$$\begin{aligned} \begin{aligned} {\mathrm{TGV}}_{\alpha }^2(u)&=\sup \Big \{\int _{{\varOmega }} u\mathrm{div}^2 \vartheta {d}x \big | \vartheta \in C_{c}^{2}({\varOmega }, S^{{d}\times {d}}),\\&\quad \Vert \vartheta \Vert _\infty \le \alpha _0, \Vert \mathrm{div} \vartheta \Vert _\infty \le \alpha _1 \Big \}, \end{aligned} \end{aligned}$$
(8)

where \(S^{{d}\times {d}}\) represents the set of all symmetric \({d}\times {d}\) matrices and \(C_{c}^{2}({\varOmega }, S^{{d}\times {d}})\) is the space of compactly supported symmetric \({d}\times {d}\) matrix fields. Moreover, the space of bound generalized variation (BGV) of order 2 endowed with

$$\begin{aligned} \begin{aligned}&\mathrm{BGV}_\alpha ^2({\varOmega })=\Big \{u\in L^1({\varOmega })\big |{\mathrm{TGV}}_{\alpha }^2(u) <\infty \Big \},\\&\quad \Vert u\Vert _{\mathrm{BGV}_\alpha ^2}=\Vert u\Vert _1+{\mathrm{TGV}}_{\alpha }^2(u), \end{aligned} \end{aligned}$$
(9)

is a Banach space. Furthermore, thanks to [25], if \({\varOmega }\subset {\mathbb {R}}^{{d}}\) is a bounded Lipschitz domain, then \(\mathrm{BGV}_\alpha ^2({\varOmega })=\mathrm{BV}({\varOmega })\) for all \((\alpha _0,\alpha _1)>0\) in the sense of topologically equivalent Banach spaces.

In the following, we focus on the dimension \({d}=2\) and denote the spaces: \(U=C_{c}^{2}({\varOmega }, {\mathbb {R}}), V=C_{c}^{2}({\varOmega }, {\mathbb {R}}^2)\), and \(W=C_{c}^{2}({\varOmega }, S^{2\times 2})\). Based on Refs. [12, 17], the discretized \({\mathrm{TGV}}_{\alpha }^2(u)\) of \(u\in U\) is then rewritten as

$$\begin{aligned} {\mathrm{TGV}}_{\alpha }^2(u)=\min \limits _{p}\alpha _1\Vert \nabla u-p\Vert _1 +\alpha _0\Vert \varepsilon (p)\Vert _1, \end{aligned}$$
(10)

where \(p=(p_1,p_2)^\mathrm{T}\in V\), and \(\varepsilon (p)=\frac{1}{2}(\nabla p+\nabla p^\mathrm{T})\) stands for the symmetrized derivative. Here, the operators \(\nabla u\) and \(\varepsilon (p)\) are characterized by

$$\begin{aligned} \nabla u=\begin{bmatrix} u_x \\ u_y \end{bmatrix}, \quad \varepsilon (p)=\begin{bmatrix} {p_1}_x&\quad \frac{1}{2}({p_1}_y+{p_2}_x)\\ \frac{1}{2}({p_1}_y+{p_2}_x)&\quad {p_2}_y \end{bmatrix}.\nonumber \\ \end{aligned}$$
(11)

Next, we briefly review the concepts of tight frame and tight wavelet frame for \(L^2({\mathbb {R}}^2)\). More details regarding this issue can be found in [21].

Definition 2

A countable set \({\mathcal {X}}\subset L^2({\mathbb {R}}^2)\) is called a tight frame if

$$\begin{aligned} f=\sum \limits _{g\in {\mathcal {X}}}\langle f,g\rangle g, \quad \forall f\in L^2({\mathbb {R}}^2), \end{aligned}$$
(12)

where \(\langle \cdot ,\cdot \rangle \) is the inner product of \(L^2({\mathbb {R}}^2)\).

Furthermore, for given \({\varPsi }=\{\psi _1,\ldots ,\psi _r\}\subset L^2({\mathbb {R}}^2)\), the wavelet system \({\mathcal {X}}({\varPsi })\) is defined by the collection of the dilations and the shifts of \({\varPsi }\) as

$$\begin{aligned} {\mathcal {X}}({\varPsi })=\{\psi _{l,s,t}: 1\le l\le r; s,t\in {\mathbb {Z}}\}, \end{aligned}$$

where \(\psi _{l,s,t}\) is characterized by

$$\begin{aligned} \psi _{l,s,t}=\left\{ \begin{array}{l@{\quad }l} 2^{s}\psi _{l}(2^s\cdot -t), &{}s\ge 0,\\ 2^{2s}\psi _{l}(2^s\cdot -2^st), &{}s< 0. \end{array} \right. \end{aligned}$$
(13)

If \({\mathcal {X}}({\varPsi })\) forms a tight frame of \(L^2({\mathbb {R}}^2)\), then the system \({\mathcal {X}}({\varPsi })\) is called a tight wavelet frame and each function \(\psi _{i}\in {\varPsi }(i=1,\ldots ,r)\) is called a (tight) framelet.

At last, let us return to the existence of the solution to (7).

Theorem 1

The optimization problem (7) admits a solution.

Proof

Let \(\{u^{k}\}\) be a bounded minimizing sequence. By the compactness property in the space \(\mathrm{BV}({\varOmega })\), there exists a subsequence of \(\{u^{k}\}\) denoted by the same symbol and \(u^*\in \mathrm{BV}({\varOmega })\), such that \(\{u^{k}\}\) converges to \(u^*\) in \(L^{1}({\varOmega })\). Subsequently, according to the standard arguments in [12, 21, 26, 27], the functions \({\mathrm{TGV}}_{\alpha }^2(u)\), \(\Vert Wu\Vert _1\) and \(I_D(u)\) are all lower semi-continuous, proper and convex, and so is their weighted sum. Therefore, this leads to that

$$\begin{aligned} \begin{aligned}&\inf \left\{ {\mathrm{TGV}}_{\alpha }^2(u)+\beta \Vert Wu\Vert _1+I_D(u)\right\} \\&\quad \ge \liminf \limits _{k\rightarrow +\infty }\left\{ {\mathrm{TGV}}_{\alpha }^2(u^k)+\beta \Vert Wu^k\Vert _1+I_D(u^k)\right\} \\&\quad \ge {\mathrm{TGV}}_{\alpha }^2(u^*)+\beta \Vert Wu^*\Vert _1+I_D(u^*), \end{aligned} \end{aligned}$$

which implies that \(u^*\) is a minimizer of the problem (7). \(\square \)

3 Numerical algorithm

With the formulation of \({\mathrm{TGV}}_{\alpha }^2\) in (10), this results in the minimization problem as follows:

$$\begin{aligned} \min \limits _{u,p}\alpha _1\Vert \nabla u-p\Vert _1 +\alpha _0\Vert \varepsilon (p)\Vert _1+\beta \Vert Wu\Vert _1+I_D(u). \end{aligned}$$
(14)

By the variable splitting technique [20, 28,29,30,31], we introduce four auxiliary variables dvwz and consider the following constrained optimization problem:

$$\begin{aligned} \begin{aligned}&\min \limits _{u,p,d,v,w,z}\alpha _1\Vert d\Vert _1 +\alpha _0\Vert v\Vert _1+\beta \Vert w\Vert _1+I_{D'}(z),\\&\mathrm{s.t.}\ \ d=\nabla u-p,\ v=\varepsilon (p),\ w=Wu,\\&\quad z=Ku,\ z\in {D'}\triangleq \left\{ z: \Vert z-f\Vert _2^2\le m^2\right\} . \end{aligned} \end{aligned}$$
(15)

To deal with the above constrained problem, we convert it into an unconstrained one by adding the quadratic penalty functions. This yields

$$\begin{aligned} \begin{aligned}&\min \limits _{u,p,d,v,w,z}\alpha _1\Vert d\Vert _1 +\alpha _0\Vert v\Vert _1+\beta \Vert w\Vert _1+I_{D'}(z)\\&\quad +\frac{\gamma _1}{2}\Vert d-(\nabla u-p)\Vert _2^2+\frac{\gamma _2}{2}\Vert v-\varepsilon (p)\Vert _2^2\\&\quad +\frac{\gamma _3}{2}\Vert w-Wu\Vert _2^2+\frac{\gamma }{2}\Vert z-Ku\Vert _2^2, \end{aligned} \end{aligned}$$
(16)

with \(\gamma _1,\gamma _2,\gamma _3,\gamma >0\) being four penalty parameters. This formulation of the problem is very advantageous because the optimization problem (16) can be solved by employing an efficient alternating minimization method. This results in the following iterative framework:

$$\begin{aligned} \begin{aligned}&\left( u^{k+1},p^{k+1},d^{k+1},v^{k+1},w^{k+1},z^{k+1}\right) \\&\quad = \mathrm{arg} \min \limits _{u,p,d,v,w,z}\alpha _1\Vert d\Vert _1 +\alpha _0\Vert v\Vert _1+\beta \Vert w\Vert _1+I_{D'}(z)\\&\qquad +\frac{\gamma _1}{2}\Vert d-(\nabla u-p)-{\tilde{d}}^k\Vert _2^2 \\&\qquad +\frac{\gamma _2}{2}\Vert v-\varepsilon (p)-{\tilde{v}}^k\Vert _2^2\\&\qquad +\frac{\gamma _3}{2}\Vert w-Wu-{\tilde{w}}^k\Vert _2^2 +\frac{\gamma }{2}\Vert z-Ku-{\tilde{z}}^k\Vert _2^2, \end{aligned} \end{aligned}$$
(17)

with the updates for \({\tilde{d}}^{k+1},{\tilde{v}}^{k+1},{\tilde{w}}^{k+1}\) and \({\tilde{z}}^{k+1}\):

$$\begin{aligned} {\tilde{d}}^{k+1}= & {} {\tilde{d}}^{k}+\gamma _1\left( \left( \nabla u^{k+1}-p^{k+1}\right) -d^{k+1}\right) , \end{aligned}$$
(18)
$$\begin{aligned} {\tilde{v}}^{k+1}= & {} {\tilde{v}}^{k}+\gamma _2\left( \varepsilon \left( p^{k+1}\right) -v^{k+1}\right) , \end{aligned}$$
(19)
$$\begin{aligned} {\tilde{w}}^{k+1}= & {} {\tilde{w}}^{k}+\gamma _3\left( Wu^{k+1}-w^{k+1}\right) , \end{aligned}$$
(20)
$$\begin{aligned} {\tilde{z}}^{k+1}= & {} {\tilde{z}}^{k}+\gamma \left( Ku^{k+1}-z^{k+1}\right) . \end{aligned}$$
(21)

More precisely, to implement the algorithm (17), we can perform this minimization efficiently by iteratively minimizing with respect to updvw and z, respectively, i.e.,

$$\begin{aligned} {\left\{ \begin{array}{ll} u^{k+1} =\mathrm{arg}\min \limits _{u}\frac{\gamma _1}{2}\Vert (\nabla u-p^k)-d^k+{\tilde{d}}^k\Vert _2^2\\ \qquad +\,\frac{\gamma _3}{2}\Vert Wu-w^k+{\tilde{w}}^k\Vert _2^2 +\frac{\gamma }{2}\Vert Ku-z^k+{\tilde{z}}^k\Vert _2^2,\\ p^{k+1}=\mathrm{arg}\min \limits _{p}\frac{\gamma _1}{2}\Vert (p-\nabla u^{k+1})+d^k-{\tilde{d}}^k\Vert _2^2\\ \qquad +\,\frac{\gamma _2}{2}\Vert \varepsilon (p)-v^k +{\tilde{v}}^k\Vert _2^2,\\ d^{k+1}=\mathrm{arg}\min \limits _{d}\alpha _1\Vert d\Vert _1 +\frac{\gamma _1}{2}\Vert d-\nabla u^{k+1}+p^{k+1}-{\tilde{d}}^k\Vert _2^2, \\ v^{k+1}=\mathrm{arg}\min \limits _{v}\alpha _0\Vert v\Vert _1 +\frac{\gamma _2}{2}\Vert v-\varepsilon (p^{k+1})-{\tilde{v}}^k\Vert _2^2, \\ w^{k+1}=\mathrm{arg}\min \limits _{w}\beta \Vert w\Vert _1 +\frac{\gamma _3}{2}\Vert w-Wu^{k+1}-{\tilde{w}}^k\Vert _2^2, \\ z^{k+1}=\mathrm{arg}\min \limits _{z}I_{D'}(z) +\frac{\gamma }{2}\Vert z-Ku^{k+1}-{\tilde{z}}^k\Vert _2^2. \end{array}\right. } \end{aligned}$$
(22)

In the first step, for solving the subproblem with respect to u, the Karush–Kuhn–Tucker (KKT) necessary conditions assert that

$$\begin{aligned} \begin{aligned}&(\gamma K^{T}K+\gamma _1\nabla ^{T}\nabla +\gamma _3W^{T}W)u^{k+1}= \gamma K^{T}(z^k-{\tilde{z}}^{k}) \\&\quad +\gamma _1\nabla ^{T}(d^k+p^k -{\tilde{d}}^{k})+\gamma _3W^{T}(w^k-{\tilde{w}}^{k}), \end{aligned} \end{aligned}$$

which means that

$$\begin{aligned} u^{k+1}= & {} \Big (\gamma K^{T}K-\gamma _1{\varDelta }+\gamma _3I\Big )^{-1} \Big [\gamma K^{T}(z^k-{\tilde{z}}^k) \nonumber \\&+\gamma _1\nabla ^{T}(d^k+p^k -{\tilde{d}}^{k}) +\gamma _3W^{T}(w^k-{\tilde{w}}^k)\Big ], \end{aligned}$$
(23)

where \({\mathcal {A}}^{T}\) denotes the adjoint of \({\mathcal {A}}, \nabla ^{T}\nabla =-{\varDelta }\) and \(W^{T}W=I\). Notice that, under the periodic boundary condition, \(K^{T}K\) and \(\nabla ^{T}\nabla \) are all block circulant, so that they can be diagonalized by fast Fourier transform (FFT). Hence, \(u^{k+1}\) is calculated by FFT and inverse FFT efficiently. Alternatively, (23) can be solved by discrete cosine transform (DCT) under the Neumann boundary condition with mirror extension and assuming that K is symmetric (see [32]). In our simulation results, we apply the periodic boundary condition and FFTs.

Next, for the p-subproblem, by differentiating the second equation of (22) of both hand sides with respect to p, we have

(24)

which we can rewrite as in the following formulation:

$$\begin{aligned} {\left\{ \begin{array}{ll} \left( \gamma _2 \nabla _1^{T}\nabla _1+\frac{\gamma _2}{2}\nabla _2^{T}\nabla _2 +\gamma _1I\right) p_1^{k+1}+\frac{\gamma _2}{2}\nabla _2^{T}\nabla _1 p_2^{k+1} \\ \quad =\gamma _1 \nabla _1 u^{k+1}+\gamma _1\left( {\tilde{d}}_1^{k}-d_1^k\right) +\gamma _2\nabla _1^{T} (v_1^k-{\tilde{v}}_1^{k}) \\ \qquad +\,\gamma _2 \nabla _2^{T}\left( v_3^k-{\tilde{v}}_3^{k}\right) , \\ \frac{\gamma _2}{2}\nabla _1^{T}\nabla _2 p_1^{k+1}+\left( \frac{\gamma _2}{2}\nabla _1^{T}\nabla _1+\gamma _2 \nabla _2^{T}\nabla _2+\gamma _1I\right) p_2^{k+1} \\ \quad =\gamma _1\nabla _2 u^{k+1}+\gamma _1\left( {\tilde{d}}_2^{k}-d_2^k\right) +\gamma _2\nabla _2^{T} \left( v_2^k-{\tilde{v}}_2^{k}\right) \\ \qquad +\,\gamma _2\nabla _1^{T}\left( v_3^k-{\tilde{v}}_3^{k}\right) . \end{array}\right. } \end{aligned}$$
(25)

As can be seen from (25), in essence, it is a system of linear equations in two unknowns \(p_1^{k+1}\) and \(p_2^{k+1}\). This observation leads to that the coefficient matrix associated with \((p_1^{k+1},p_2^{k+1})\) can be diagonalized blockwise under the Fourier transform.

Before going further, let us introduce some necessary notations used in what follows. In the sequel, for notational convenience, we denote the expressions by

$$\begin{aligned} a_1= & {} \gamma _2\nabla _1^{T}\nabla _1+\frac{\gamma _2}{2} \nabla _2^{T} \nabla _2+\gamma _1I,\ \ a_2=\frac{\gamma _2}{2}\nabla _1^{T}\nabla _2, \\ a_3= & {} \frac{\gamma _2}{2}\nabla _1^{T}\nabla _1+\gamma _2 \nabla _2^{T}\nabla _2+\gamma _1I, \\ b_1= & {} \gamma _1\nabla _1 u^{k+1}+\gamma _1({\tilde{d}}_1^{k}-d_1^k)+\gamma _2\nabla _1^{T} (v_1^k-{\tilde{v}}_1^{k})\\&+\gamma _2 \nabla _2^{T}(v_3^k-{\tilde{v}}_3^{k}), \\ b_2= & {} \gamma _1\nabla _2 u^{k+1}+\gamma _1({\tilde{d}}_2^{k}-d_2^k)+\gamma _2\nabla _2^{T} (v_2^k-{\tilde{v}}_2^{k})\\&+\gamma _2\nabla _1^{T}(v_3^k-{\tilde{v}}_3^{k}). \end{aligned}$$

This convention, together with (25), yields that

$$\begin{aligned} {\left\{ \begin{array}{ll} a_1{p_1^{k+1}}+a_2^{T}{p_2^{k+1}}=b_1, \\ a_2{p_1^{k+1}}+a_3{p_2^{k+1}}=b_2. \end{array}\right. } \end{aligned}$$
(26)

It is noteworthy that linear operators \(a_1,a_2\) and \(a_3\) are all block-circulant matrices under the periodic boundary condition, and which further can be diagonalized by FFT. As a consequence, this together with the Cramer’s rule, two variables \(p_1^{k+1}\) and \(p_2^{k+1}\) in the system (26) can be efficiently computed in the Fourier domain.

As for the dv and w subproblems shown in (22), the generalized shrinkage formula, similarly as in [33], can be adopted preferentially. Namely,

$$\begin{aligned} d^{k+1}= & {} \mathrm{shrink}\Big (\big (\nabla u^{k+1}-p^{k+1}\big ) +{\tilde{d}}^k,\frac{\alpha _1}{\gamma _1}\Big ), \end{aligned}$$
(27)
$$\begin{aligned} v^{k+1}= & {} \mathrm{shrink}\Big (\varepsilon (p^{k+1})+{\tilde{v}}^k, \frac{\alpha _0}{\gamma _2}\Big ), \end{aligned}$$
(28)
$$\begin{aligned} w^{k+1}= & {} \mathrm{shrink}\Big (Wu^{k+1}+{\tilde{w}}^k,\frac{\beta }{\gamma _3}\Big ), \end{aligned}$$
(29)

with \(\mathrm{shrink}(t,\delta )=\mathrm{sign}(t)\cdot \max (|t|-\delta ,0)\).

At last, we are now in the position to solve the z-subproblem. It can be written as

$$\begin{aligned} z^{k+1}=\mathrm{arg}\min \limits _{z}\frac{\lambda ^{k+1}}{2}\Vert z-f\Vert _2^2 +\frac{\gamma }{2}\Vert z-(Ku^{k+1}+{\tilde{z}}^k)\Vert _2^2, \end{aligned}$$
(30)

which indicates that

$$\begin{aligned} z^{k+1}=(\lambda ^{k+1}f+\gamma (Ku^{k+1}+{\tilde{z}}^k))/ (\lambda ^{k+1}+\gamma ), \end{aligned}$$
(31)

where the regularization parameter \(\lambda ^{k+1}\) is updated by the discrepancy principle in the \((k+1)\)th iteration. Obviously, the solutions of \(\lambda ^{k+1}\) and \(z^{k+1}\) are related together. More precisely, if \(\Vert (Ku^{k+1}+{\tilde{z}}^k)-f\Vert _2^2\le m^2 \ (\star )\) holds, we set \(\lambda ^{k+1}=0\) and \(z^{k+1}=Ku^{k+1}+{\tilde{z}}^k\). On the contrary, by the discrepancy principle, we should solve the following equation:

$$\begin{aligned} \Vert z^{k+1}-f\Vert _2^2=m^2. \end{aligned}$$
(32)

To this end, the relationship (31), together with (32), leads to that

$$\begin{aligned} \lambda ^{k+1}=(\gamma \Vert f-(Ku^{k+1}+{\tilde{z}}^k)\Vert _2/m)-\gamma . \end{aligned}$$
(33)

In conclusion, putting all of these elements together, this results in the following algorithmic framework: alternating minimization method, devoted for solving (7).

figure b

4 Convergence analysis

In this section, as far as the convergence property of the above iterative algorithm is concerned, we will present the main theorem in what follows. Here, we only give the basic proof frameworks and do not repeat the lengthy proving process. Similar to the classical alternating direction method of multiplier (ADMM) developed in [34,35,36], we have the following theorem.

Theorem 2

For given \(\gamma _1, \gamma _2, \gamma _3, \gamma >0\), the sequence \(\{u^{k},p^{k},d^{k},v^{k}, w^{k},z^{k},\lambda ^{k}\}\) generated by the proposed algorithm from any initial point converges to a solution of (16).

Fig. 1
figure 1

Recovered results by using three different models on Lena image. a Original image, b noisy image, c wavelet frame, d TGV model, e our hybrid scheme

Fig. 2
figure 2

Locally enlarged images recovered by using three different models on Lena image. a Original image, b noisy image, c wavelet frame, d TGV model, e our hybrid scheme

Proof

It follows from (17) that the (up)-subproblem and dvwz subproblems are decoupled each other. Thus, six variables can be grouped into two blocks (up) and (dvwz). Therefore, our method can be regarded as an application of ADMM. Concerning the convergence proof of the proposed approach, let us first construct the Lagrangian functional as follows:

$$\begin{aligned}&{\mathscr {L}}(u,p,d,v,w,z;{\tilde{d}},{\tilde{v}},{\tilde{w}},{\tilde{z}})= \alpha _1\Vert d\Vert _1+\alpha _0\Vert v\Vert _1 \nonumber \\&\quad +\beta \Vert w\Vert _1+I_{D'}(z)+\frac{\gamma _1}{2}\Vert d-(\nabla u-p)-{\tilde{d}}\Vert _2^2 \nonumber \\&\quad +\frac{\gamma _2}{2}\Vert v-\varepsilon (p)-{\tilde{v}}\Vert _2^2 +\frac{\gamma _3}{2}\Vert w-Wu-{\tilde{w}}\Vert _2^2 \nonumber \\&\quad +\frac{\gamma }{2}\Vert z-Ku-{\tilde{z}}\Vert _2^2, \end{aligned}$$
(34)

and denote three variables by \(X=(u,p),Y=(d,v,w,z)\) and \(Z=({\tilde{d}}, {\tilde{v}}, {\tilde{w}}, {\tilde{z}})\). Recurring to the standard arguments on ADMM stated above, for given \(\gamma _1,\gamma _2,\gamma _3,\gamma >0\), then the sequence generated by the resulting iterative framework (22) converges to the saddle point of (34), and the proof is completed. \(\square \)

5 Experimental results

In this section, we illustrate the effectiveness of the proposed hybrid model with different wavelet frames on five \(256 \times 256\) test images: Lena, Cameraman, Boat, Peppers and Butterfly. We also compare the recovered results with four closely related TGV, wavelet frame, TV+wavelet (TVW for short) and deep learning-based methods, by measuring the reconstruction quality, staircase effect suppression and edge-preserving ability. For fair comparisons, under the discrepancy principle, these regularized models are all solved by adopting the regularization parameter update algorithms.

Table 1 Comparison of the recovered results using three different models on Lena image
Fig. 3
figure 3

Recovered results by using three different models on Cameraman image. a Original image, b degraded image, c wavelet frame, d TV+wavelet frame, e our hybrid scheme

An important remark is that all images are processed by our scheme with the parameters \(\gamma _1=0.5, \gamma _2=1, \gamma _3=1\) and \(\gamma =10^{(\mathrm{BSNR}/10-1)}\) for reconstructing the reasonable results, with \(\mathrm{BSNR}=10\log _{10}(\Vert f\Vert _2^2/\Vert n\Vert _2^2)\). The other parameters are firmly fixed to \(\alpha _0=3, \alpha _1=1.5\) and \(\beta =0.5\). Additionally, as is suggested in [5, 6], the parameter \(\tau \) can be selected as \(\tau =-\tau _0\times \mathrm{BSNR}+1.09\), with \(\tau _0=0.03\) for image denoising and \(\tau _0=0.006\) for image deblurring, respectively. All experiments are implemented using MATLAB R2011b on a PC with Intel(R) Core(TM) i5 CPU and 4 GB of RAM under Windows 7.

The stopping criterion for all the tested algorithms is set to \(\Vert u^{k+1}-u^k\Vert _2/\Vert u^k\Vert _2<10^{-4}\) or the number of iterations is larger than 1000. The quality of the recovered image is quantitatively measured by peak signal-to-noise ratio (PSNR), which is defined as

$$\begin{aligned} \mathrm{PSNR}=10\log _{10}\Big (\frac{255^2\cdot n_1n_2}{\Vert u-{\tilde{u}}\Vert _2^2}\Big ), \end{aligned}$$
(35)

where u and \({\tilde{u}}\) denote the original image and the restored data, respectively. Also, the optimal \(\lambda \) is chosen in achieving the best restoration with respect to the PSNR value. Furthermore, the Pratt’s figure of merit (FOM) criterion [37] is employed to evaluate the edge-preserving ability of different models. Meanwhile, we also compare their recovered results by adapting the feature similarity (FSIM) index [38] for image quality assessment. Generally speaking, the larger PSNR, FOM and FSIM values normally indicate that the restoration is of higher quality.

Table 2 Comparison of the recovered results using three different models on Cameraman image
Fig. 4
figure 4

Recovered results by using three different models on Boat image. a Original image, b degraded image, c wavelet frame, d TV+wavelet frame, e our hybrid scheme

Example 1

First, we validate the ability of the proposed hybrid regularizers strategy for image denoising and compare it with two successful methods: the wavelet frame-based model and the TGV model. Here, the wavelet frame is a 4-scale redundant Haar frame. The original image Lena is shown in Fig. 1a. Figure 1b (\({\hbox {PSNR}}=28.1281\) dB) stands for its noisy version corrupted by white random Gaussian noise with standard variance 10. Figure 1c, d denotes the denoised versions by the wavelet frame model and the TGV model, respectively. And we display in Fig. 1e the performance of our novel scheme. The local enlarged images and recovered results by three different models are separately listed in Fig. 2 and Table 1. As might be expected, it follows from Fig. 1 and Table 1 that our proposed hybrid model provides the best restoration, visually and quantitatively, in terms of suppressing noise and preserving details over some existing sophisticated numerical methods.

Example 2

We take another standard test image Cameraman (Fig. 3a) as an example for image deconvolution. Its degraded version shown in Fig. 3b (\({\hbox {PSNR}}=23.0625\) dB) is blurred by Gaussian convolution with a \(5\times 5\) window and a standard deviation of 3, and noisy by white Gaussian noise with variance \(\sigma ^2=15\). In this case, the wavelet frame is selected as the 4-scale Daubechies-4 wavelet. The deconvolution results by the wavelet frame model, the TV+wavelet model and our method are displayed in Fig. 3c–e and Table 2, respectively.

Example 3

In Fig. 4, we compare the denoised and deblurred results for image Boat, by using the wavelet frame model, the TV+wavelet model and our proposed strategy. We use 2-scale Coiflet filter for our wavelet frame to deal with the contaminated image (Fig. 4b, \({\hbox {PSNR}}=24.1873\) dB), blurred by motion blur with parameters “len = 6” and “theta = 30,” and noisy by white Gaussian noise with variance \(\sigma ^2=15\). Subsequently, provided Fig. 4 and Table 3 indicate the deconvolution results by using three different models in more detail.

Table 3 Comparison of the recovered results using three different models on Boat image

Example 4

Subsequently, to further evaluate the performance of the addressed hybrid regularizers approach to image deblurring, we use 4-scale symmetric framelet [39] for our model and deal with the contaminated Peppers image. Figure 5b (\({\hbox {PSNR}}=22.3220\) dB) denotes the degenerated version blurred by a \(6\times 6\) averaging filer and noisy by white Gaussian noise with variance \(\sigma ^2=20\). Furthermore, the restored results by the TGV model, the TV+wavelet model and our proposed algorithm are presented in Table 4 and Fig. 5 at great length.

Fig. 5
figure 5

Recovered results by using three different models on Peppers image. a Original image, b degraded image, c TGV model, d TV+wavelet frame, e our hybrid scheme

Table 4 Comparison of the recovered results using three different models on Peppers image
Table 5 Comparison of the recovered results using three different models on five test images

Example 5

At last, with the aim of further illustrating the performance, we compare our reconstructions with the TGV and another popular and powerful deep learning-based methods [40, 41]. It is worth pointing out that the Daubechies wavelet D6 is chosen for wavelet frame in the implementation of our algorithm. As suggested in [40], the MLP approach is carried out with the Gaussian window width of 2 and a stride size of 3. Five degraded images are all corrupted by additive Gaussian noise with standard variance 15. Recovered results and measurable comparisons obtained using three different strategies are intuitively depicted and listed in Table 5 and Fig. 6, respectively.

Fig. 6
figure 6

Recovered results by using three different models on five test images: Lena, Cameraman, Boat, Peppers and Butterfly. a1a5 Degraded image, b1b5 MLP, c1c5 TGV model, d1d5 our hybrid scheme

Observing the restorations in Figs. 3, 4, 5 and 6 gives that the oscillation and staircasing artifacts are frequently produced by the canonical wavelet frame and TV-based methods. However, the images recovered by our novel model are more visually natural and veritable. Other comparisons outlined in Tables 234 and 5, especially in achieving higher PSNR, FOM and FSIM values than those of the wavelet frame, TGV, TV+wavelet and deep learning-based efficient methods, concertedly illustrate the outstanding performance of the proposed approach to image deblurring, with respect to both restoration accuracy and edge-preserving ability.

6 Conclusions

In this article, by incorporating the advantages of two recently developed wavelet frame-based and TGV methods, we introduce a novel hybrid regularizers model for image denoising and deblurring. Associated with the alternating minimization method, the proposed framework is calculated by an efficient adaptive parameter estimation algorithm, where the parameter \(\lambda \) is changing automatically during the iterations. Convergence of the algorithm is also briefly described. Finally, in comparison with some state-of-the-art techniques, experimental results distinctly illustrate the unexampled superiority of our developed strategy in solving the image restoration problem, in terms of reconstruction quality, staircasing effect reduction and details preservation.