1 Introduction

Multiplicative noise removal is one of the most primary tasks in the research of coherent imaging systems [1,2,3,4]. It is crucial to find an effective method that can protect the details of digital images as much as possible while smoothing the oscillations. Denote an observed image by \(f=f(x)\), \(x\in { \Omega }\), where \({ \Omega }\) is a bounded open subset of \(\mathbb {R}^{2}\) and has a Lipschitz boundary. Usually the degradation process of coherent imaging system is given mathematically as

$$\begin{aligned} f=u\circ \eta \end{aligned}$$
(1)

where u is unknown ideal image, \(\circ \) refers to the componentwise multiplication, \(\eta \) is the noise which obeys a certain probability distribution. In a real-world application scenario, for example, the Gamma distribution is studied in synthetic aperture radar (SAR) images [3], and the Rayleigh distribution is researched in ultrasound imaging [4]. In the scope of this paper, we focus on dealing with multiplicative noise in L-look SAR image [5]. It is well known that the noise obeys a Gamma distribution

$$\begin{aligned} p(\eta )=\frac{L^{L}}{\Gamma (L)}\eta ^{L-1}\exp \{-L\eta \}\chi _{\{\eta \ge 0\}} \end{aligned}$$
(2)

where \(\chi _{\{\eta \ge 0\}}\) is the indicator function of the set \(\{\eta \ge 0\}\), the mean value of \(\eta \) is 1, and the variance of noise \(\eta \) is 1/L.

Various techniques such as TV regularization, non-local filtering, sparse representation and tight-frame approach have been applied to deal with multiplicative Gamma noise. We refer the reader to see [6,7,8,9,10,11,12,13,14,15]. Focusing on the variational based method, one of the landmarks is the proposal of the AA model [14]. Authors derived a new fidelity term from the maximum a posteriori estimation under the assumption that the multiplicative noise obeys the Gamma distribution. The goal is to minimize the following energy functional

$$\begin{aligned} \min _{u}\int _{ \Omega }|Du|+\lambda \int _{ \Omega }\left( \log u+\frac{f}{u}\right) \mathrm{d}x \end{aligned}$$
(3)

where \(\int _{ \Omega }|Du|\) stands for the total variation (TV) of u. However, the model lacks of global convexity since it is strictly convex only for \(u\in (0,2f)\). Undeniably, it promotes the development of a series of global convex models based on TV regularization, which have good performance for piecewise constant images. For instance, Huang, Ng and Wen [7] used the transformation \(u=\exp {z}\) in (3), and considered the TV regularization for z. Zhao, Wang and Ng [10] proposed rewritting the degradation equation (1) as \(f\circ w=u\), where \(w(x)=\frac{1}{\eta (x)}\), and derived the following model for multiplicative noise removal:

$$\begin{aligned} \min _{w,u}\int _{ \Omega }\frac{1}{2}(w-\mu )^{2}+\alpha _{1}|fw-u|\mathrm{d}x+\alpha _{2}\int _{ \Omega }|Du| \end{aligned}$$

where \(\mu \) can be set to be the mean value of w. The model is jointly convex for (uw). On the flip side, TV regularization is known to yield some undesired effects, such as the staircase and blocky effect [16]. To mitigate this inherent effect, researchers begin to focus on the study of replacing TV with higher order differential operators [17, 18]. Besides that, Chan and Strong [19] introduced the adaptive total variation and proposed the following model:

$$\begin{aligned}\min \int _{ \Omega }g(x)|\nabla u|\mathrm{d}x\end{aligned}$$

where the weight function g(x) controls the regularization level of different regions. Yet, most of the existing weight functions are designed according to \(|\nabla u|\), namely, the structural information of the image. After the signal dependence of multiplicative noise is captured in [20], several authors begin to explore the use of gray level indicator in the weight function.

Another approach for multiplicative noise removal is the diffusion based approach. In essence, anisotropic diffusion is to iteratively modeficate the observed image through a partial differential equation (PDE) to suppress noise. This kind of method is inspired by the work of [21], where the nonlinear diffusion equation (PM model) is proposed for the images corrupted by additive Gaussian noise. Lately, Zhou et al. [11] proposed a doubly degenerated (DD) diffusion model to address the multiplicative noise removal problem

$$\begin{aligned}&\frac{\partial u}{\partial t} = \mathrm{div}(g(u,|\nabla u|)\nabla u)\quad in\,~ { \Omega }_{T}\\&\frac{\partial u}{\partial \mathbf {n}}=0\quad on \,~ \partial { \Omega }\times (0,T)\\&u=f\quad on \,~{ \Omega }\times \{t=0\} \end{aligned}$$

Here, the diffusion coefficient is given as follows

$$\begin{aligned}g(u,|\nabla u|)=\frac{2|u|^{\alpha }}{M^{\alpha }+|u|^{\alpha }}\cdot \frac{1}{\left( 1+|\nabla u|^{2}\right) ^{\frac{1-\beta }{2}}}\end{aligned}$$

where \(\alpha >0,\) \(0<\beta <1\), \(M=\sup _{x\in { \Omega }}u(x,t)\), \(\frac{2|u|^{\alpha }}{M^{\alpha }+|u|^{\alpha }}\) is a gray level indicator, and \(\frac{1}{\left( 1+|\nabla u|^{2}\right) ^{\frac{1-\beta }{2}}}\) is devoted to the edge detection. In the work of [22], a fractional-order nonlinear diffusion model is proposed, which can handle the multiplicative noise well and protect the texture. Because these equations are highly non-linear, the CFL condition forces the time step to be small, resulting in many iterations of the algorithm. Although the DD model uses the fast explicit diffusion (FED) to accelerate the calculation, the nature of explicit diffusion still has a great limitation on the time step.

Recently, Euler’s elastica regularization has received extensive attention because of its good geometric interpretation. Many functional models based on Euler’s elastica have been proposed for image inpainting [23, 24], segmentation [25, 26], restoration [27], deconvolution [28] and other tasks. Since they integrate curvature information involving second derivatives, the corresponding Euler–Lagrange equations are usually fourth-order. It is very time consuming to calculate their evolution equations directly, which raises a nontrivial issue of developing effective and efficient algorithms to solve them. The work of [29] approximates the Euler’s elastica based image inpainting model by an elliptic system consisting of a weighted TV equation and a linear equation. Furthermore, several fast numerical algorithms based on operator splitting technique have been widely used to solve diverse models arising from image processing [30, 31]. They convert the original problem into a few simple and easily solved subproblems, also allow for a broader class of regularizers. However, parameter tuning will greatly affect its convergence rate [32], and the convergence of non-convex problems is limited. These still do not affect the pursuit of such an algorithm for nonlinear and non-convex problems [26, 33, 34], because the numerical implementation is indeed effective.

In this article, our main contributions are twofold. Firstly, since Euler’s elastica functional is not suitable for multiplicative noise, we propose a noval multiplicative denoising model based on adaptive Euler’s elastica regularization by introducing the gray level indicator [20, 35]. Secondly, two effective and efficient numerical algorithms are proposed. On one hand, using the gradient descent algorithm, an efficient semi-implicit discrete scheme for the gradient descent flow is designed, and then an additive operator splitting algorithm is employed to accelerate the calculation. On the other hand, a restricted proximal augmented Lagrangian algorithm is proposed, which transforms the original complicated functional minimization problem into an iterative update of a series of simple sub-problems. Numerical experiments illustrate that the new model can restore some geometric features of the image well.

The paper is organized as follows. In Sect. 2, the working mechanism of Euler’s elastica energy in image processing is reviewed, and then a model based on adaptive Euler’s elastica is developed for images contaminated by multiplicative Gamma noise. In Sect. 3.1, the model is implemented by the gradient descent method, a semi-implicit discrete scheme with AOS algorithm is designed to solve the evolution equation. Additionally, a restricted proximal augmented Lagrangian algorithm is presented in Sect. 4. Experimental results are discussed in Sect. 5. Section 6 concludes the paper by highlighting the significance of the proposed model.

2 Adaptive Euler’s Elastica Regularization Combined with Convex Fidelity Terms for Multiplicative Noise Removal

In this section, we review the related concepts and geometric meanings of Euler’s elastica energy. Then, benefiting from the geometric characteristics of adaptive Euler’s elastica energy, a new multiplicative Gamma noise removal model is described, which can maintain some features of the image.

2.1 The Brief Review on Variational Geometry

In image analysis, perhaps we can consider that a collection of curves have to be dealt with simultaneously. Let u(x) denote an image on a domain \({ \Omega }\), so the image is a stack of curves

$$\begin{aligned}\gamma _{\lambda }: u(x)=\lambda \quad x=(x_{1},x_{2})\in { \Omega }\end{aligned}$$

To weigh the quality of each individual curve \(\gamma _{\lambda }\), the first step is to establish some suitable measure \(E[\gamma _{\lambda }]\).

The first natural geometric measure is the length L. In addition, curvature k plays a more explicit role, since human can easily observe and process convex or concave features of shapes and boundaries. Generally, the second order measure for smooth curve \(\gamma _{\lambda }\) is given

$$\begin{aligned}E_{k}[\gamma _{\lambda }]=\int _{\gamma _{\lambda }}f(k)\mathrm{d}s\end{aligned}$$

where f(k) is some suitable function of curvature. In order to effectively penalize local singularities, such as kinks or corners, a special but very useful case is Euler’s elastica measure

$$\begin{aligned}E_{el}[\gamma _{\lambda }]=\int _{\gamma _{\lambda }}(a+bk^{2})\mathrm{d}s\end{aligned}$$

with some positive constant coefficients a and b. The ratio a/b manifests the relative importance of the total length versus total squared curvature.

Thus, for image denoising, the minimization criterion becomes

$$\begin{aligned} E_{el}[u]=\int ^{+\infty }_{-\infty }\int _{\gamma _{\lambda }}(a+bk^{2})\mathrm{d}s\mathrm{d}\lambda \end{aligned}$$

Using the celebrated co-area formula, this can be lifted to be

$$\begin{aligned} E_{el}[u]=\int _{ \Omega } \left( a+b\left( \nabla \cdot \frac{\nabla u}{|\nabla u|}\right) ^{2}\right) |\nabla u|\mathrm{d}x \end{aligned}$$
(4)

When \(b=0\), the functional degenerates into the total variation of u. This reveals that the geometric essence of TV measure is the sum of the perimeters of image level set curves.

It is worth noting that the functional \(E_{el}[u]\) consists of two parts, TV and the curvature term \(I:=\int _{ \Omega }\left( \nabla \cdot \frac{\nabla u}{|\nabla u|}\right) ^{2}|\nabla u|\mathrm{d}x\), where I is a higher-order term with stronger regularity. Generally, higher-order derivatives play a more important role. For example, Poincar\(\acute{\mathrm{e}}\)’s inequality and embedding theorem tell us that lower derivatives can be controlled by higher derivatives. Therefore, if the adaptive strategy [19, 20] is adopted to make the model suitable for the multiplicative denoising problem, a weight function can be designed to replace the constant a in (4) to correct the basic regularization process caused by TV. In Sect. 2.3, the effects of this modification will be experimentally demonstrated.

2.2 Proposed Model

In this subsection, we aim to establish a multiplicative noise removal model based on adaptive Euler’s elastica. In particular, by analyzing the characteristics of multiplicative noise, the gray level indicator \(\alpha (u)\) is introduced, namely

$$\begin{aligned} \alpha (u)=\left( \frac{G_{\sigma }u}{M}\right) ^p \end{aligned}$$
(5)

where p is a positive constant, \(G_{\sigma }\) represents a Gaussian kernel whose width parameter is \(\sigma \), and \(M=\max G_{\sigma }u\).

In fact, according to the degradation model (1), the expectation and variance of the noisy image can be obtained through simple calculations as

$$\begin{aligned}&\mathrm{E}(f(x))=\mathrm{E}(u(x)\eta (x))=u(x)\\&\mathrm{Var}(f(x))=u^{2}(x)\mathrm{Var}(\eta (x))=\frac{u^{2}(x)}{L}\end{aligned}$$

It is not difficult to find that the variance of noisy image depends on expectation, which reveals the signal-dependent nature of multiplicative noise pollution levels. As illustrated in Fig. 1a, the greater the original signal strength, the greater the influence of noise. Therefore, we expect the regularization level to be higher where the original image pixel value is higher.

Fig. 1
figure 1

The signal-dependent nature of the fluctuations and gray level indicator

However, as image denoising is a typical inverse problem, the original pixel value can only be approximated by the pre-processed noisy image. Thus, in the gray level indicator (5), the image is first preprocessed through a small-scale Gaussian filter. And then the gray level is compressed to [0,1]. The idea of the exponential parameter p is derived from gamma correction [36], which is used to avoid the compression of the pixel information at the intermediate gray level. Since the noise pollution level is positively correlated with the gray level, the gray level indicator can be used as a weight function in the regularization term to adjust the degree of regularization adaptively. Specifically, the gray level indicator \(\alpha (u)\) becomes very small at low gray levels, which leads to preserve image information at low gray levels. In high gray level regions, \(\alpha (u)\) is closing to 1, so that outliers are eliminated. As shown in Fig. 1b, \(\alpha \) coincides with the pollution level at each pixel of the noisy image.

Therefore, combined with the fidelity term in [20], we propose the following multiplicative denoising functional model based the adaptive Euler’s elastica

$$\begin{aligned} \min _{u}\int _{ \Omega }\left( \alpha +b\left( \nabla \cdot \frac{\nabla u}{|\nabla u|}\right) ^{2}\right) |\nabla u|\mathrm{d}x+\eta \int _{ \Omega }\left( u-f\log u\right) \mathrm{d}x \end{aligned}$$
(6)

where positive constants b, \(\eta \) leverage the importance of regularity and fidelity. In particular, here \(\alpha =\alpha (f)\) is considered to be a known coefficient function.

2.3 The Role of Regularization Terms

In fact, the change of coefficient pairs \((\alpha , b)\) in model (6) will cause the degradation of the model. Specifically, the model (6) covers four types of regularization terms: (1) adaptive Euler’s elastica regularization: \(\alpha =\alpha (f), b\ne 0\); (2) standard Euler’s elastica regularization: \(\alpha \) is a positive constant, \(b \ne 0\); (3) adaptive total variation: \(\alpha =\alpha (f),b=0\); 4) TV (AA model): \(\alpha \) is a positive constant, \(b=0\). Next we will test these four cases to present the role of regularization term.

Fig. 2
figure 2

The role of ingredients in the regularization. af are Color image, noisy image and four restored images in sequence

Table 1 Parameters used in the comparison study and PSNR

In Fig. 2, four kinds of denoising results are shown by taking "Color.png" as the test image. The parameters and the peak signal to noise ratio (PSNR) are recorded in Table 1. By comparison, the classical AA model, which adopts total variation regularization, has a lower PSNR value and an obvious staircase effect in the recovery result (Fig. 2c). PSNR values have been significantly improved under Euler’s elastica framework or when the adaptive total variation technique is applied, but there are still some dissonant factors in the restored image. For example, the "fish scale" phenomenon is shown in Fig. 2d, and some halos still appear in Fig. 2e. Naturally, this means that their level set curves are not very smooth and appear jagged in many places. On the whole, when adaptive total variation is combined with curvature constraint, a more ideal result (Fig. 2f) can be achieved. Thus, the proposed adaptive Euler’s elastica regularization is meaningful.

In conclusion, the adaptive Euler’s elastica regularization consists of two parts: the adaptive total variation and the curvature-driven term. They control the length and contour of the level set curves respectively. Therefore, it has a stronger ability to deal with details, especially the gradient information.

3 The Additive Operator Splitting (AOS) Algorithm

3.1 The Gradient Descent Flow

The basis of the calculation in this section is the variational method and gradient descent flow. That is, the solution of the original minimization functional model is transformed into the calculation of the gradient descent flow corresponding to the first-order variation of the functional.

The first variation of Euler’s elastica energy (4) has been derived in [23], and please refer to lemma 1 below for details.

Lemma 1

Let \(\phi \in C^{1}(\mathbb {R},[0,\infty ))\) be a given function and

$$\begin{aligned}R[u]=\int _{ \Omega }\phi (k)|\nabla u| \mathrm{d}x\end{aligned}$$

Then the first variation is given by

$$\begin{aligned}\frac{\partial R}{\partial u}=-\nabla \cdot \mathbf {V}\end{aligned}$$

where the flux field \(\mathbf {V}\) is

$$\begin{aligned}\mathbf {V}=\phi (k)\mathbf {n}-\frac{\mathbf {t}}{|\nabla u|}\frac{\partial (\phi ' (k)|\nabla u|)}{\partial \mathbf {t}}\end{aligned}$$

Here \(\mathbf {n}\) is the ascending normal field, and \(\mathbf {t}\) is the targent field. Namely,

$$\begin{aligned}k=\nabla \cdot \frac{\nabla u}{|\nabla u|}, \quad \mathbf {n}=\frac{\nabla u}{|\nabla u|}, \quad \mathbf {t}=\mathbf {n}^{\bot }, \quad \frac{\partial }{\partial \mathbf {t}}=\mathbf {t}\cdot \nabla \end{aligned}$$

Therefore, the Euler Lagrange equation corresponding to the minimization problem (6) is

$$\begin{aligned}-\nabla \cdot \left[ (\alpha +bk^{2})\mathbf {n}-\frac{2b}{|\nabla u|}\frac{\partial (k|\nabla u|) }{\partial \mathbf {t}}\mathbf {t} \right] +\eta \left( 1-\frac{f}{u}\right) =0 \end{aligned}$$

Then, the gradient descent method is adopted to obtain the following steepest descent marching

$$\begin{aligned} \frac{\partial u}{\partial t}=\nabla \cdot \left[ (\alpha +bk^{2})\mathbf {n}-\frac{2b}{|\nabla u|}\frac{\partial (k|\nabla u|) }{\partial \mathbf {t}}\mathbf {t} \right] -\eta \left( 1-\frac{f}{u}\right) \end{aligned}$$
(7)

The natural boundary conditions along \(\partial { \Omega }\) are

$$\begin{aligned} \frac{\partial u}{\partial \mathbf {v}}=0\quad \text {and}\quad \frac{\partial (2bk|\nabla u|)}{\partial \mathbf {v}}=0 \end{aligned}$$
(8)

3.2 Explicit Scheme

From now on, we present the numerical scheme for the evolution equation (7). Let (ij) denote the digital pixel locations, and choose a small time step \(\tau \), then time be digitized to \(n\tau \), \(n=0,1,\ldots \). Thus \(u^{n}_{i,j}\) denotes the value of u at pixel (ij) at time \(n\tau \). For convenience of description, the right end of equation (7) is denoted as H(u).

If the explicit scheme is used directly to discretize the equation, we have

$$\begin{aligned}\frac{u^{n+1}_{i,j}-u^{n}_{i,j}}{\tau }=H(u^{n})_{i,j}\end{aligned}$$

Then the update criterion is

$$\begin{aligned} u^{n+1}_{i,j}=u^{n}_{i,j}+\tau H(u^{n})_{i,j} \end{aligned}$$
(9)

the time step \(\tau \) must be very small to ensure the strict stability of the algorithm. So it needs more iterations to achieve its convergence. In Table 2, we use Cameraman and Hourglass as test images . We can see that the explicit scheme is very inefficient because it is strictly limited by a very small time step. At the same time, we can see that the computational cost depends on the image size and the noise level.

Table 2 The calculation efficiency of explicit scheme
Fig. 3
figure 3

Restorations obtained by the explicit scheme

Therefore, we propose a new and efficient numerical PDE scheme in the next subsection.

3.3 Semi-implicit Scheme with AOS

In order to improve the efficiency of computation, a semi-implicit scheme is designed to reduce the limitation of time step on the efficiency of the algorithm. Note that the right term of the equation in (7) actually consists of three parts. For simplicity, the combination of the cross partial derivative and the nonlinear source term is denoted as F(u). The coefficient of the diffusion term and F(u) are fixed to the value at \(n\tau \). So the following semi-implicit scheme is obtained

$$\begin{aligned}&\frac{u^{n+1}_{i,j}-u^{n}_{i,j}}{\tau }=\nabla \cdot \left( \frac{\alpha +b(k^{n})^{2}}{|\nabla u^{n}|}\nabla u^{n+1}\right) _{i,j}+F(u^{n})_{i,j} \end{aligned}$$
(10)
$$\begin{aligned}&F(u^{n})=-\nabla \cdot \left( \frac{2b}{|\nabla u^{n}|^{3}}(\nabla u^{n})^{\bot }\cdot \nabla (k^{n}|\nabla u^{n}|)(\nabla u^{n})^{\bot } \right) -\eta \left( 1-\frac{f}{u^{n}} \right) \end{aligned}$$
(11)

For brevity, set \(G(u^{n})=\frac{\alpha +b(k^{n})^{2}}{|\nabla u^{n}|}\nabla u^{n+1}\). So far, the constructed scheme (10) is just suitable for the additive operator decomposition (AOS) algorithm proposed in [37], which is a fast algorithm for solving nonlinear differential equations. The core of the algorithm is to divide the original iteration (10) into two iterations, that is, to split the two-dimensional diffusion process into the one-dimensional diffusion of the rows and columns of \(u^{n}\), and two intermediate results \(U_{1}^{n+1}\) and \(U_{2}^{n+1}\) are obtained. Then the average of them is the result of the nth iteration, \(u^{n+1}=\frac{U^{n+1}_{1}+U^{n+1}_{2}}{2}\). Therefore, the problem is transformed into solving the following algebraic equations

$$\begin{aligned}&\frac{U_{1}^{n+1}-U^{n}_{1}}{\tau }=2G^{n}_{x}U^{n+1}_{1}+F^{n} \end{aligned}$$
(12)
$$\begin{aligned}&\frac{U_{2}^{n+1}-U^{n}_{2}}{\tau }=2G^{n}_{y}U^{n+1}_{2}+F^{n} \end{aligned}$$
(13)

where \(G_{x}^{n}\) and \(G_{y}^{n}\) are two one-dimensional diffusion operators. Take the solution of equation (13) as an example for a simple explanation. Let the image size be \(M \times N\), update the jth column of u sequentially. Namely, traverse j from 1 to N, using the Thomas algorithm to calculate the equations

$$\begin{aligned}(I-2\tau G^{n}_{y,j}) U^{n+1}_{2,j}=u^{n}_{j}+F^{n}_{j}\end{aligned}$$

for the jth column of \(u^{n}\), due to the boundary condition (8), let \(u_{0,j}^{n}=u_{1,j}^{n}\) and \(u_{N+1,j}^{n}=u_{N,j}^{n}\) to expand the image formally. Thus, the corresponding column diffusion operator \(G^{n}_{y,j}\) is obtained

$$\begin{aligned}\begin{bmatrix} -G^{n}_{\frac{3}{2},j}&{} G^{n}_{\frac{3}{2},j}&{}0 &{} \cdot \cdot \cdot &{} 0 &{} 0 &{}0 \\ G^{n}_{\frac{3}{2},j}&{} -G^{n}_{\frac{3}{2},j}-G^{n}_{\frac{5}{2},j} &{}G^{n}_{\frac{5}{2},j}&{} \cdot \cdot \cdot &{}0 &{} 0 &{}0\\ \vdots &{} \vdots &{}\vdots &{} \cdot \cdot \cdot &{} \vdots &{} \vdots &{}\vdots \\ 0&{} 0 &{}0 &{} \cdot \cdot \cdot &{}G^{n}_{N-\frac{3}{2},j}&{} -G^{n}_{N-\frac{3}{2},j}-G^{n}_{N-\frac{1}{2},j} &{}G^{n}_{N-\frac{1}{2},j}\\ 0&{} 0 &{}0 &{} \cdot \cdot \cdot &{} 0&{}G^{n}_{N-\frac{1}{2},j}&{} -G^{n}_{N-\frac{1}{2},j} \end{bmatrix} \end{aligned}$$

where \(G^{n}_{y,j}\) is an \(N\times N\) tridiagonal matrix. Similarly, (12) can also be solved. Specifically, the process of solving (10) by AOS algorithm is written in Algorithm 1.

figure a

It is worth noting that the AOS algorithm is absolutely stable for any time step \(\tau \). The following is a simple theorem, please refer to reference [37] for details.

Theorem 1

For any time step \(\tau \), semi-implicit scheme (12) and (13) is sufficient to guarantee absolute stability.

Proof

\(G_{x,i}^{n}\) and \(G_{y,j}^{n}\) are diagonally dominant matrices, and the diagonal elements are all negative. For any \(\lambda _{1}>0\), \(G_{y,j}^{n}\) is a strictly diagonally dominant matrix, \((\lambda _{1} I-G_{y,j}^{n})x=0\) has no non-zero solutions. Hence, \(\lambda _{1}\) is not an eigenvalue of \(G_{y,j}^{n}\), and the eigenvalues of \(G_{y,j}^{n}\) are nonpositive. Suppose \(\lambda _{1}\le 0\) is the eigenvalue of \(G_{y,j}^{n}\), then the eigenvalue \((1-2\tau \lambda _{1})^{-1}\) of \((I-2\tau G_{y,j}^{n})^{-1}\) satisfies \((1-2\tau \lambda _{1})^{-1}\in (0,1]\). Similarly, assuming that \(\lambda _{2}\le 0\) is the eigenvalue of \(G_{x,i}^{n}\), we can obtain that the eigenvalue \((1-2\tau \lambda _{2})^{-1}\) of \((I-2\tau G_{x,i}^{n})^{-1}\) satisfies \((1-2\tau \lambda _{2})^{-1}\in (0,1]\). Thus, the semi-implicit scheme is stable for any \(\tau >0\). \(\square \)

4 The Restricted Proximal Augmented Lagrangian Method

By introducing auxiliary variables, a proximal augmented Lagrangian method is implemented to relax model (6), and the alternate direction multiplier method (ADMM) is used to solve the minimal element iteratively. It is worth noting that a simple truncation strategy in [34] was adopted to ensure the consistency between the model and the algorithm based on augmented Lagrangian. That is, when the coefficient b of the curvature term is equal to 0, the original problem is reduced to an adaptive total variation model, so the corresponding numerical algorithm should directly solve the adaptive total variation model from the very beginning.

4.1 Conventional Augmented Lagrangian Functional

Now, we introduce four variables \(\mathbf{p} , \mathbf{n} , h, v\), and cast the functional (6) into a constrained minimization problem

$$\begin{aligned}&\min _{u,\mathbf{p} ,\mathbf{n} ,\mathbf{m} ,v}\int _{ \Omega }\left( \alpha +bh^{2}\right) |\mathbf{p} |\mathrm{d}x+\eta \int _{ \Omega }\left( v-f\log v\right) \mathrm{d}x \nonumber \\&\mathbf{p} =\nabla u, \mathbf{n} =\frac{p}{|p|}, h=\nabla \cdot \mathbf{n} , v=u \end{aligned}$$
(14)

where the virtue of v is to split the nonlinear term and the higher-order term. To avoid the degenerate case of \(\mathbf{p} =\mathbf{0} \), the constraint \(\mathbf{n} =\frac{p}{|p|}\) is amended to its equivalent form \(|\mathbf{p} |=\mathbf{n} \cdot \mathbf{p} , |\mathbf{n} |\le 1\), as suggested in [33, 38]. The crucial constraint \(|\mathbf{n} |\le 1\) prevents \(|\mathbf{n} |\) from being unbounded when \(\mathbf{p} =\mathbf{0} \). To impose the constraint \(|\mathbf{n} |\le 1\) a.e. in \({ \Omega }\), we define a characteristic function

$$\begin{aligned} \delta _{ \Omega }(\mathbf{n} )=\left\{ \begin{matrix} 0 &{}\quad |\mathbf{n} |\le 1\\ +\infty &{}\quad otherwise \end{matrix}\right. \end{aligned}$$

So far, the Lagrangian functional can be shown as

$$\begin{aligned} \begin{aligned} L (u,\varvec{p},\varvec{n},h,v;\lambda _{1},\varvec{\lambda _{2}},\lambda _{3},\lambda _{4})&=\int _{ \Omega }\left( \alpha +bh^{2}\right) |\varvec{p}|+\eta (v-f\log v)\mathrm{d}x+\delta _{ \Omega }(\varvec{n})\\&\quad +\int _{ \Omega }\lambda _{1}(|\varvec{p}|-\varvec{n}\cdot \varvec{p})\mathrm{d}x+\frac{r_{1}}{2}\int _{ \Omega }(|\varvec{p}|-\varvec{n}\cdot \varvec{p})^{2}\mathrm{d}x\\&\quad +\int _{ \Omega }\varvec{\lambda _{2}}\cdot (\varvec{p}-\nabla u)\mathrm{d}x+\frac{r_{2}}{2}\int _{ \Omega }|\varvec{p}-\nabla u|^{2}\mathrm{d}x\\&\quad +\int _{ \Omega }\lambda _{3}(v-u)\mathrm{d}x+\frac{r_{3}}{2}\int _{ \Omega }(v-u)^{2}\mathrm{d}x\\&\quad +\int _{ \Omega }\lambda _{4}(h-\nabla \cdot \varvec{n})\mathrm{d}x+\frac{r_{4}}{2}\int _{ \Omega }(h-\nabla \cdot \varvec{n})^{2}\mathrm{d}x\\ \end{aligned} \end{aligned}$$
(15)

Here, \(\lambda _{1},\lambda _{3},\lambda _{4}\in \mathbb {R}\) and \(\varvec{\lambda _{2}}\in \mathbb {R}^{2}\) are Lagrangian multipliers, and \(r_{i}\) \((i=1,2,3,4)\) are positive scalars as weighting parameters. The aim of the augmented Lagrangian method is to acquire a saddle point of (15), which approximates the solution of the original problem (6). To this end, the alternating direction multiplier method is used to calculate, i.e. given the previous iterate \((u^{k},\varvec{p}^{k},\varvec{n}^{k},h^{k},v^{k};\lambda _{1}^{k},\varvec{\lambda _{2}}^{k},\lambda _{3}^{k},\lambda _{4}^{k})\), then solve the subproblems corresponding to variables \(u,\varvec{p},\varvec{n},h,v\) in turn and update the Lagrangian multipliers.

However, in order to make the iteration results more stable and obtain the consistency between the algorithm and the model, the subproblems of u and \(\mathbf{p} \) are modified.

4.2 Proximal Operators and Truncation Techniques

On one hand, for the u-subproblem, we add a quadratic constraint, so that the updated \(u^{k+1}\) does not deviate too far from the result \(u^{k}\) of the last iteration. So, the u-subproblem is

$$\begin{aligned}u^{k+1}=\mathrm{arg} \min _{u}{} L (v^{k+1},u,\varvec{p}^{k},\varvec{n}^{k},h^{k};\lambda _{1}^{k},\varvec{\lambda _{2}}^{k},\lambda _{3}^{k},\lambda _{4}^{k})+\frac{\tau }{2}\Vert u-u^{k}\Vert _{S}^{2}\end{aligned}$$

S is a self-adjoint positive semidefinite operator, for simplicity, where we consider the proximal term in \(L^{2}\) setting. In the experiment, adding the quadratic term is helpful for us to select appropriate parameters to get the convergent solution. In addition, the existence and uniqueness of solutions to u-subproblem is supported by the theory of proximal operators. Below we provide a brief overview, one can see [32, 39, 40] for details.

For a convex function g, the proximal operator of g respect to a semi-definite matrix S at y is defined by

$$\begin{aligned} {\mathbf {prox}}_{g,S}(y):=\arg \min _{z}\left\{ g(z)+\frac{1}{2\mu }\Vert z-y\Vert ^{2}_{S}:z\in \mathbb {R}^{d}\right\} \end{aligned}$$

where \(\mu >0\). If g is strictly convex and coercive then its proximity operator exists uniquely [41]. In particular, when the matrix S is exactly the identity matrix I, then the above proximity operator is reduced to

$$\begin{aligned} {\mathbf {prox}}_{g,I}(y):=\arg \min _{z}\left\{ g(z)+\frac{1}{2\mu }\Vert z-y\Vert ^{2}_{2}:z\in \mathbb {R}^{d}\right\} \end{aligned}$$

and the associated optimal value specifies its Moreau envelope

$$\begin{aligned} M_{\mu g}(y):=g({ \mathbf {prox}}_{g,I}(y))+\frac{1}{2\mu }\Vert { \mathbf {prox}}_{g,I}(y)-y\Vert ^{2}_{2} \end{aligned}$$

The Moreau envelope is a continuously differentiable function, even when g is not, and its gradient [39] is given by

$$\begin{aligned} \nabla M_{\mu g}(y)=\frac{1}{\mu }\left( y-{ \mathbf {prox}}_{g,I}(y)\right) \end{aligned}$$

For instance, when g is the \(l_{1}\) norm, \(g(z)=\Vert z\Vert _{1}=\sum |z_{i}|\), then the proximal operator is determined by soft thresholding

$$\begin{aligned}{\mathbf { prox}}_{g,I}(y_{i})=\mathrm{sign}(y_{i})\cdot \max (|y_{i}|-\mu ,0)\end{aligned}$$

the associated Moreau envelope is the Huber function

$$\begin{aligned}M_{\mu g}(y_{i})=\left\{ \frac{1}{2\mu }y_{i}^{2},|y_{i}|\le \mu ;|y_{i}|-\frac{\mu }{2},|y_{i}|\ge \mu \right\} \end{aligned}$$

and its gradient is the saturation function \(\mathrm{sign}(y_{i})\cdot \min (|y_{i}|/\mu ,1)\).

On the other hand, for \(\varvec{p}\)-subproblem, the cutting-off strategy proposed in [34] is used to eliminate the connection with the variable \(\varvec{n}\), which is according to the dependent order of variables. Therefore, the \(\varvec{p}\)-subproblem is reduced to

$$\begin{aligned}\varvec{p}^{k+1}=\mathrm{arg} \min _{\varvec{p}}\int _{ \Omega }\left( \alpha +b(h^{k})^{2}\right) |\varvec{p}|+\varvec{\lambda _{2}^{k}}\cdot (\varvec{p}-\nabla u^{k+1})+\frac{r_{2}}{2}|\varvec{p}-\nabla u^{k+1}|^{2}\mathrm{d}x\end{aligned}$$

Because \(\varvec{n}\) is defined as \(\varvec{p}/|\varvec{p}|\) to express some normal fields, it is reasonable to assume that \(\varvec{p}\) does not depend on \(\varvec{n}\). However, this strategy is very critical. When \(b=0\), the closed iterative loop is formed with subproblems of \(\varvec{p}^{k+1}\) and \(u^{k+1}\), which is exactly in line with the augmented Lagrangian method for solving the adaptive total variation model.

At this point, we write the pseudo-code of the alternating iterative algorithm in Algorithm 2. From now on, we will name our method as RPALM (restricted proximal augmented Lagrangian method) for brevity.

figure b

4.3 The Solutions of Minimization Subproblems

In this section, let’s show how to solve the subproblems in each iteration of Algorithm 2. For ease of description, we state some mathematical notations in the discrete setting.

Let \({ \Omega }=[1,N]\times [1,M]\) be a set of \(N\times M\) points in \(\mathbb {R}^{2}\), denote the Euclidean space \(\mathbb {R}^{N\times M}\) as V and \(Q=V\times V\). For \(u\in V\) and \(\mathbf{p} \in Q\), given \((i,j)\in [1,N]\times [1,M]\), then \(u(i,j)\in \mathbb {R}\) and \(\mathbf{p} (i,j)=(p_{1}(i,j),p_{2}(i,j))\in \mathbb {R}^{2}\). The standard Euclidean inner products can be defined as

$$\begin{aligned}&(u,v)_{V}=\sum _{i,j}u(i,j)v(i,j) \\&(\mathbf{p} ,\mathbf{q} )_{Q}=(p_{1},q_{1})_{V}+(p_{2},q_{2})_{V} \end{aligned}$$

Using periodic boundary condition, the discrete forward and backward difference operators for \(u\in V\) are defined as follows

$$\begin{aligned}&\partial ^{+}_{x}u(i,j)=\left\{ \begin{matrix} u(i+1,j)-u(i,j) &{}\quad 1\le i<M\\ u(1,j)-u(M,j)\,~\,~ &{}\quad i=M \end{matrix}\right. \\&\partial ^{+}_{y}u(i,j)=\left\{ \begin{matrix} u(i,j+1)-u(i,j), &{}\quad 1\le j<N\\ u(i,1)-u(i,N),\,~\,~ &{}\quad j=N \end{matrix}\right. \\&\partial ^{-}_{x}u(i,j)=\left\{ \begin{matrix} u(i,j)-u(i-1,j), &{}\quad 1< i\le M\\ u(1,j)-u(M,j),\,~\,~ &{}\quad i=1 \end{matrix}\right. \\&\partial ^{-}_{y}u(i,j)=\left\{ \begin{matrix} u(i,j)-u(i,j-1), &{}\quad 1< j\le N\\ u(i,1)-u(i,N),\,~\,~ &{}\quad j=1 \end{matrix}\right. \end{aligned}$$

Furthermore, the discrete gradient operator \(\nabla :V\rightarrow Q\) can be defined as follows

$$\begin{aligned}\nabla u(i,j)=(\partial ^{+}_{x}u(i,j),\partial ^{+}_{y}u(i,j))\end{aligned}$$

and the negative adjoint or divergence operator div:\(Q\rightarrow H\) of \(\nabla \) is given by

$$\begin{aligned}\text {div}{} \mathbf{p} (i,j)=\partial ^{-}_{x}p_{1}(i,j)+\partial ^{-}_{y}p_{2}(i,j)dx\end{aligned}$$

4.3.1 Solution for v-Subproblem

Because the objective problem (16)

$$\begin{aligned} \min _{v}\int _{ \Omega }\eta (v-f\log v)+\lambda _{3}^{k}v+\frac{r_{3}}{2}(v-u^{k})^{2}\mathrm{d}x \end{aligned}$$

is a convex nonlinear functional with respect to v. Newton iterative algorithm with second order convergence can be used, the first and second order variation of v are

$$\begin{aligned}&\delta ^{1}_{v}=\eta \left( 1-\frac{f}{v}\right) +\lambda _{3}^{k}+\frac{r_{3}}{2}(v-u^{k})\\&\delta ^{2}_{v}=\eta \frac{f}{v^{2}}+\frac{r_{3}}{2} \end{aligned}$$

thus

$$\begin{aligned} v^{k+1}=v^{k}-\frac{\delta ^{1}_{v^{k}}}{\delta ^{1}_{v^{k}}} \end{aligned}$$
(21)

4.3.2 Solution for u-Subproblem

For \(u-\)subproblem (17)

$$\begin{aligned}\min _{u}\int _{ \Omega }-\varvec{\lambda _{2}}^{k}\cdot \nabla u+\frac{r_{2}}{2}|\varvec{p}^{k}-\nabla u|^{2}-\lambda _{3}u+\frac{r_{3}}{2}(v^{k+1}-u)^{2}+\frac{\tau }{2}(u-u^{k})^{2}\mathrm{d}x\end{aligned}$$

whose optimality condition yields a linear equation

$$\begin{aligned} -r_{2}\text {div}(\nabla u) +r_{3}u+\tau u=-\text {div}\varvec{\lambda _{2}}^{k}-r_{2}\text {div}\varvec{p}^{k}+\lambda _{3}^{k}+r_{3}v^{k+1}+\tau u^{k} \end{aligned}$$

which can be efficiently solved by fast Fourier techniques. The updated \(u^{k+1}\) can be expressed explicitly as

$$\begin{aligned} u^{k+1}=\mathcal {F}^{-1}\left( \frac{\mathcal {F}(-\text {div}\varvec{\lambda _{2}}^{k}-r_{2}\text {div}\varvec{p}^{k}+\lambda _{3}^{k}+r_{3}v^{k+1}+\tau u^{k})}{-r_{2}\mathcal {F}(\Delta )+r_{3}+\tau }\right) \end{aligned}$$
(22)

where \(\mathcal {F}\) and \(\mathcal {F}^{-1}\) mean fast Fourier transform (FFT) and its inverse transform respectively.

4.3.3 Solution for \(\mathbf{p} \)-Subproblem

It is easy to see that the \(\varvec{p}\)-subproblem (18) can be recast as

$$\begin{aligned}\min _{\varvec{p}}\int _{ \Omega }(\alpha +b(h^{k})^{2})|\varvec{p}|+\frac{r_{2}}{2}\left| \varvec{p}-\left( \nabla u^{k+1}-\frac{\varvec{\lambda }_{2}^{k}}{r_{2}}\right) \right| ^{2}\mathrm{d}x\end{aligned}$$

we use the soft thresholding operator defined in [42] to find the closed-form formula for the minimizer

$$\begin{aligned} \varvec{p}^{k+1}=shrinkage\left( \nabla u^{k+1}-\frac{\varvec{\lambda }_{2}^{k}}{r_{2}},\frac{\alpha +b(h^{k})^{2}}{r_{2}}\right) \end{aligned}$$
(23)

where

$$\begin{aligned}shrinkage(x,y)=\mathrm{sign}(x)\cdot \max (|x|-y,0)\end{aligned}$$

4.3.4 Solution for \(\mathbf{n} \)-Subproblem

For \(\mathbf{n} \)-subproblem, we actually solve for

$$\begin{aligned} \min _\mathbf{n }\int _{ \Omega }-\lambda _{1}^{k}{} \mathbf{n} \cdot \mathbf{p} ^{k+1}+\frac{r_{1}}{2}(|\mathbf{p} ^{k+1}|-\mathbf{n} \cdot \mathbf{p} ^{k+1})^{2}-\lambda _{4}^{k}\text {div}{} \mathbf{n} +\frac{r_{4}}{2}(h^{k}-\text {div}{} \mathbf{n} )^{2}\mathrm{d}x \end{aligned}$$
(24)

and then projecting the minimizer into the unit ball. Here, let’s approximate \(\mathbf{n} \) in the second term with \(\mathbf{n} ^{k}\), and quadratic term \(\int _{ \Omega }\frac{r_{4}}{2}(h^{k}-\text {div}{} \mathbf{n} )^{2}\mathrm{d}x\) is linearized follows [43, 44], i.e.

$$\begin{aligned} \int _{ \Omega }\frac{r_{4}}{2}(h^{k}-\text {div}{} \mathbf{n} )^{2}\mathrm{d}x\approx \int _{ \Omega }&\frac{r_{4}}{2}(h^{k}-\text {div}{} \mathbf{n} ^{k})+r_{4}\langle \nabla (h^{k}-\text {div}{} \mathbf{n} ^{k}),\mathbf{n} -\mathbf{n} ^{k}\rangle \\&+\frac{1}{2\delta }|\mathbf{n} -\mathbf{n} ^{k}|^{2}\mathrm{d}x \end{aligned}$$

where \(\delta >0\) is a constant. Hence, the corresponding Euler–Lagrange equation can be simplified as

which generates a closed-form solution \(\mathbf{n} ^{k+1}\). In addition, the Euler–Lagrange equation of (24) can also be directly solved, which is just a linear coupled PDEs. Nevertheless, due to the singularity of the operator \(\nabla (\text {div})\), it is impossible to directly apply FFT technique to solve the problem. To overcome this difficulty, a quadratic penalty term \(\frac{\mu }{2}\int _{ \Omega }|\mathbf{n} -\mathbf{n} ^{k}|^{2}\mathrm{d}x\) should be added to (24). And then, let’s project \(\mathbf{n} ^{k+1}\) onto the unit circle, which is

$$\begin{aligned} {\varvec{n}}^{k+1}=\left\{ \begin{matrix} {\varvec{n}}^{k+1}, &{} if\, |{\varvec{n}}^{k+1}|\le 1\\ \frac{{\varvec{n}}^{k+1}}{|{\varvec{n}}^{k+1}|}, &{}otherwise \end{matrix}\right. \end{aligned}$$
(25)

4.3.5 Solution for h-Subproblem

As for the h-subproblem

$$\begin{aligned}\min _{h}\int _{ \Omega }b|\mathbf{p} ^{k+1}|h^{2}+\lambda _{4}^{k}h+\frac{r_{4}}{2}(h-\text {div}{} \mathbf{n} ^{k+1})\mathrm{d}x\end{aligned}$$

We can easily derive the following update formula from its optimality condition

$$\begin{aligned} h^{k+1}=\frac{r_{4}\text {div}{} \mathbf{n} ^{k+1}-\lambda _{4}^{k}}{2b|\mathbf{p} ^{k+1}|+r_{4}} \end{aligned}$$
(26)

5 Results and Discussion

In this paper, we tested several classical images, such as, the Pepper, Cone, etc. (see Fig. 4). Quantitatively, the performance of algorithm is evaluated by using two quality metrics named as the Peak Signal to Noise Ratio (PSNR) [45] and the Structural SIMilarity (SSIM) [46]. PSNR is defined as

$$\begin{aligned} \mathrm{PSNR}=10\log _{10}\frac{MN|\max u-\min u|^{2}}{||\hat{u}-u||_{L^{2}}^{2}}\mathrm{dB} \end{aligned}$$

where u and \(\hat{u}\) denote the original image and the recovered image respectively, M and N are the size of image. While SSIM is defined as

$$\begin{aligned} \mathrm{SSIM}=\frac{(2\mu _{u}\mu _{\hat{u}}+c_{1})(2\sigma _{u\hat{u}}+c_{2})}{(\mu _{u}^{2}\mu _{\hat{u}}^{2}+c_{1})(\sigma _{u}^{2}+\sigma _{\hat{u}}^{2}+c_{2})} \end{aligned}$$

Here \(\mu _{u},\mu _{\hat{u}}\) are the mean values of images u and \(\hat{u}\), \(\sigma _{u},\sigma _{\hat{u}}\) represent the variances of \(u,\hat{u}\), and \(\sigma _{u\hat{u}}\) is the covariance of u and \(\hat{u}\). Note that the larger PSNR and larger SSIM means the better restoration effect. For each experiment, the parameters are selected to obtain the maximum PSNR and SSIM. For convenience, the two algorithms of our developed model are denoted as AOS and RPALM. All experiments of both AOS and RPALM algorithms proposed in this paper are performed under Ubuntu 16.04 and Python 3.6 running on a Desktop Computer with an Intel Core (TM) i7-9700K CPU at 3.60GHz with 8GB of RAM.

Fig. 4
figure 4

Ground truth images used in the paper. a Pepper, size: \(256\times 256\). b Pillsetc, size: \(512\times 384\). c Cameraman, size: \(256\times 256\). d Hourglass, \(128 \times 128\). e Brain, size: \(210\times 210\). f Cone, size: \(301 \times 301\). g Letter, size: \(315\times 169\). h Coloredchips, size: \(518\times 391\). i Moon, size: \(358\times 537\)

Fig. 5
figure 5

Denoising results of image Pepper for different \(\tau \) with fixed the remaining parameters. \(\sigma =0.8,p=1.4,b=0.001,\eta =0.001.\) The number of iterations, PSNR and SSIM were recorded respectively for the recovery results. a Noisy (\(L=10\)), b 5,26.77, 0.7870, c 7,27.09, 0.7869, d 22,27.77, 0.8200, e 166,27.93, 0.8236

In our model, \(\alpha \) and b are the coefficients of the adaptive Euler’s elastica energy. The ratio \(\alpha /b\) represents the relative importance of the total length versus total squared curvature. For \(b=0\), the minimization problem (6) degenerates into an adaptive total variation multiplicative denoising model. The balance parameter \(\eta \) is the coefficient of the fidelity term. In the case of denoising, this parameter generally does not need to be too large to obtain a cleaner restored image.

5.1 Discussion of Parameters in AOS Algorithm

In AOS algorithm, in addition to model parameters \(\sigma ,p,b,\eta \), time step \(\tau \) is also introduced. In Theorem 1 we emphasize the absolute stability of semi-implicit schemes. Although the larger \(\tau \) is, the faster the convergence rate is, we still need to choose the appropriate \(\tau \) to find a balance between accuracy and efficiency. Figure 5 shows examples of different time step \(\tau \) with fixed remaining parameters by the above techniques. From this, we find that Fig. 5b and c look blurry visually, and d and e have good visual effects. But the number of iterations at \(\tau =0.001\) was about eight times that at \(\tau =0.01\). Therefore, \(\tau =0.01\) is an appropriate choice in this group of experiments.

5.2 Discussion of Parameters in RPALM Algorithm

In the RPALM algorithm, in addition to the inherent parameters of the model, due to constraints and linearization techniques, six parameters \(\tau , r_{1}, r_{2}, r_{3}, r_{4}\) and \(\delta \) are also introduced. These optimal parameters usually need to be tuned manually, which is quite a troublesome task. However, it is worth noting that in the process of using the operator splitting technique to transform the original model into a constrained optimization problem, the most critical constraint is \(\mathbf{p} =\nabla u\). Moreover, the penalty term \(\frac{\tau }{2}\Vert u-u^{k}\Vert ^{2}_{S}\) in the u-subproblem also plays an important role in enhancing the robustness of the algorithm. Therefore, according to the above analysis and some experimental tests, we divide all the parameters into two categories. The first category contains high-sensitive parameters which are \(\tau \) and \(r_{2}\), their values may directly affect the convergence of the algorithm. The second category contains low-sensitive parameters which are \(r_{1}, r_{3}, r_{4}, \delta \), the change of PSNR caused by their perturbation is small.

Moreover, the experiment on Pillsetc in Fig. 6 shows that when appropriate parameters are taken, after multiple iterations of the RPALM algorithm, the auxiliary variable \(\mathbf{p} \) is sufficiently close to \(\nabla u\) and v is sufficiently close to u, which verifies that the constraint (14) is satisfied, and the RPALM algorithm is meaningful.

Fig. 6
figure 6

Numerical convergence of RPALM. Parameter selection: \(p=1.1, b=1.2, \eta =1, \delta =120,\tau =85,r_{1}=70,r_{2}=32,r_{3}=15,r_{4}=40.\) (c), (d) are the two components of the auxiliary variable \(\mathbf{p} \) in the RPALM algorithm and (e) is the auxiliary variable v

5.3 Efficient of Solving Higher Order Model

In Sect. 3.2, we have witnessed that the traditional gradient descent method for solving the second-order functional minimization model is very time-consuming, so it is necessary to develop more efficient numerical algorithms. Adressing this challenge, this paper proposed two algorithms, AOS and RPALM, in Sects. 3.3 and 4 respectively. Hereon, we use the same test images as Fig. 3, the noisy Cameraman (\(L=4\)) and the noisy Hourglass (\(L=10\)), to show the significant improvement in the computational efficiency of the AOS and RPALM. The parameters selection and performance indicators (including PSNR, number of iterations and running CPU-time) are listed in Table 3. Note that in Table 2, the traditional gradient descent method is strictly limited by the time step, and it takes tens of thousands of iterations to process an image, while the algorithm in this paper can achieve the optimal restoration by hundreds of iterations at most, the computational efficiency has been improved by nearly 300 times.

Table 3 Comparison of AOS and RPALM algorithms

It’s noteworthy that the AOS algorithm converges for any time step \(\tau \), if a larger \(\tau \) is set, the noise will be removed faster, but this will sacrifice calculation accuracy. In addition, in Fig. 7e, f, k, l, we plot the curve of numerical energy (i.e. the value of the functional in (6)) vary with number of iterations. One can easily see that the two algorithms proposed for the multiplicative noise removal model based on adaptive Euler’s elastica are both convergent. Furthermore, the RPALM algorithm can make the functional energy drop to a certain threshold quickly, after which the energy drop speed becomes very slow.

5.4 Denoising Performance

In this subsection, we provide amount numerical results to illustrate the image restoration capability of our method. The results are compared with those obtained by some state of the art algorithms,

  • The AA model based on TV regularization proposed by Aubert and Aujol [14];

  • The HNZ model proposed by Huang et al. [7] is derived based on logarithmic transformation and TV regularization;

  • The ZWN model proposed by Zhao et al. [10] involves the TV regularization term, the terms of variance of the inverse of noise, and the fidelity term;

  • The DD model based on adaptive diffusion technology proposed by Zhou et al. [11];

  • The FD model based on tight frame proposed by Wang et al. [12];

  • The MuLoG+BM3D algorithm proposed by Deleddalle et al. [13], which integrates multichannel logarithm, block matching and 3-D filtering.

Among them, the AA model is solved by gradient descent method, the HNZ, ZWN and FD models are solved by alternating minimization algorithm, and DD model is solved by fast explicit diffusion algorithm. In the experiment, parameters are selected for them to obtain the maximum PSNR. For the MuLoG+BM3D algorithm, the free parametres are set as suggested in [13]. For the AOS and RPALM algorithms developed in this paper, the selection of parameters is tabulated in Tables 4 and 5 respectively.

Fig. 7
figure 7

Convergence analysis of the algorithms

Table 4 The parameter values of AOS for numerical experiments
Table 5 The parameter values of RPALM for numerical experiments

Next, we report the numerical experiments of multiplicative noise removal for the original test images Fig. 4e–i. All test images are corrupted by multiplicative noise with \(L\in \left\{ 1,4,10\right\} \), and the corresponding results are depicted in Figs. 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21 and 22. In particular, for ZWN and FD models, only the case with noise level \(L=4\) and \(L=10\) is recorded, because the expectation of the inverse of the noise is bounded only when \(L>1\).

For a comparison of the performance quantitatively, we list the PSNR and SSIM values of the restored results in Table  6, where the best results are shown in bold face. These tables show that both AOS and RPALM algorithms are competitive to other algorithms.

As illustrated in Figs. 8, 9 and 10, when the image is multiplied by a noise that obeys a Gamma distrbution with a heavy right tail, there will be many unusually bright outliers in a noisy image. The denoised results of the AA algorithm, the HNZ algorithm, the ZWN algorithm and the FD algorithm, much more white spots remains compared with those other algorithms. Moreover, the contrast of the details for these four reference algorithms are noticeably reduced due to the oversmoothing during noise removal, please observe the wrinkled areas in the brain computed tomography images. Essentially, the DD model and ours model adopt the adaptive smoothing strategy to significantly improve these problems. They can better maintain the image contrast than those three methods based on the standard TV regularizqation. As a non-local method,

MuLoG+BM3D algorithm performs well in contrast protection and noise removal.

As shown in Figs. 11, 12 and 13, significant staircase effect can be seen in the results of the TV-based methods. Besides, in the original Cone image, there do not

seem to be two similar image blocks, which affects the performance of the MULOG+BM3D method. Intuitively, Fig. 11i–m demonstrate that our model can better fit the “inclined plane”.

Fig. 8
figure 8

The Brain image: a original. b Noisy: \(L=1\). ch: denoised results

Fig. 9
figure 9

The Brain image: a original. b Noisy: \(L=4\). cj: denoised results

Fig. 10
figure 10

The Brain image: a original. b Noisy: \(L=10\). cj: denoised results

Fig. 11
figure 11

The Cone image: a Original. b Noisy: \(L=1\). ch: Denoised results. im 3D drawing

In addition, we ran more tests and showed them in Figs. 14, 15, 16, 17, 18, 19, 20, 21 and 22. All the algorithms obtained reliable results. Although the MuLoG+BM3D algorithm can recover fine structures very well, artificial effect is inevitably caused by

Fig. 12
figure 12

The Cone image: a Original. b Noisy: \(L=4\). cj: Denoised results

Fig. 13
figure 13

The Cone image: a original. b Noisy: \(L=10\). cj: Denoised results

Fig. 14
figure 14

The Letter image: a original. b Noisy: \(L=1\). ch: Denoised results

stacking the image blocks in the BM3D method [47], please see Figs. 14f and 17f. Furthermore, the restored images obtained by our model have smoother background than the results obtained by the DD model, please see Fig. 17e, g and h. This is due to the curvature-driven regular term maintaining the rounded curve and the stronger regularity in the uniform region than the lower-order gegular term. In Figs. 17, 18, 19, 20, 21 and 22, whether it is the gray contrast between the chess pieces in the image ColoredChips or the pothole informa-

tion in the lower right of the image Moon, our algorithms can make it better protected.

Fig. 15
figure 15

The Letter image: a original. b Noisy: \(L=4\). cj: Denoised results

Fig. 16
figure 16

The Letter image: a Original. b Noisy: \(L=10\). cj: Denoised results

Fig. 17
figure 17

The ColoredChips image: a original. b Noisy: \(L=1\). ch: Denoised results

Fig. 18
figure 18

The ColoredChips image: a original. b Noisy: \(L=4\). cj: Denoised results

Fig. 19
figure 19

The ColoredChips image: a original. b Noisy: \(L=10\). cj: Denoised results

Fig. 20
figure 20

The Moon image: a original. b Noisy: \(L=1\). ch: Denoised results

Comparing the denoised results and the test image, we can clearly find that our method can reduce artificial effect, preserve the tiny geometrical structures and supresse the noise successfully. Also, as presented in Table 6, the proposed model delivers competitive results in terms of PSNR value and structure preservation.

6 Conclusion

In this paper, we study the restoration of the images polluted by multiplicative Gamma noise. Based on the gray level indicator, an adaptive Euler’s elastica regularization model is proposed to deal with the problem. The model is solved in the spatial domain, both two algorithms are effective and efficient. One is a fast algorithm based on PDE numerical solution, and the other is an optimization method.

Fig. 21
figure 21

The Moon image: a original. b Noisy: \(L=4\). cj: Denoised results

Fig. 22
figure 22

The Moon image: a Original. b Noisy: \(L=10\). cj: Denoised results

Some experiments conducted on both the synthetic and real images indicated the effectiveness and efficiency of the proposed method. The visual impression shows that the proposed model performs well in protecting tiny structures, maintaining contrast, and restoring a smooth image background. In addition, the quantitative measurement on PSNR and SSIM, indicate that the proposed method possesses more remarkable capacity of noise reduction than some other classical methods.

Table 6 Parameters and performance indicators used in the comparison