Abstract
A spatially varying Gamma mixture model prior is employed for tomographic image reconstruction, ensuring effective noise elimination and the preservation of region boundaries. We define a line process, modeling edges between image segments, through appropriate Markov random field smoothness terms which are based on the Student’s t-distribution. The proposed algorithm consists of two alternating steps. In the first step, the mixture model parameters are automatically estimated from the image. In the second step, the reconstructed image is estimated by optimizing the maximum-a-posteriori criterion using the one-step-late expectation–maximization and preconditioned conjugate gradient algorithms. Numerical experiments on various photon-limited image scenarios show that the proposed model outperforms the compared state-of-the-art reconstruction models.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Emission tomography (ET) and tomographic reconstruction have gained tremendous attention in the last decades. This is due to the crucial role that medical imaging has come to play in the diagnosis of human disease, as well as in monitoring the progress of therapeutic methods. From a computational-model point of view, tomographic reconstruction is a typical inverse problem. After noisy emission data are collected, the goal is to reconstruct the original object, thereby inverting in this sense the image generation process [10, 16, 19].
Several approaches have been proposed for tomographic reconstruction, a considerable part of which focuses on methods based on iterative image reconstruction. This family of methods involves an iterative procedure of repeated projections and backprojections that progressively refines the image estimate [2, 38]. Iterative tomographic reconstruction techniques are separated into two basic categories: algebraic and statistical [1, 40], with the latter having proved to be the most popular. Shepp and Vardi have defined the problem in terms of maximization of a likelihood function [35], proposing the widely used maximum likelihood expectation–maximization (MLEM) algorithm. In EM and its variants, a probabilistic model is solved with an iterative scheme of alternating steps: the expectation (E-step) and the maximization step (M-step). In the E-step, moments of latent model variables are computed with respect to the current non-latent parameter estimate. In the M-step, non-latent parameters are optimized, with respect to current latent variable moments. The two interrelated steps are iterated until convergence. In practice, slight variants of the basic EM algorithm are used, such as MAP-EM, where part of the model random variables can be treated as parameters to be optimized, and their prior plays the standard role of a penalty term.
The ordered subsets EM (OSEM) algorithm has been proposed [14, 17] as an efficient variation of the MLEM algorithm, reducing the reconstruction time and the computational cost, as well as facilitating clinical use. The kernel-based EM algorithm that has been presented in [37] for PET image reconstruction is also of note. The row action maximum likelihood algorithm (RAMLA) [5] has been proposed as another MLEM variation, further speeding up convergence rate. Furthermore, with respect to speeding up reconstruction, also notable is the parallel tomographic reconstruction technique that has recently been proposed in [26], suitable for use with GPU-ready machines.
Bayesian maximum a posteriori methods, tantamount to penalized maximum likelihood [4], have been employed for tomographic reconstruction. Methods of this family impose a prior probability density function (pdf) on the reconstruction, encouraging image smoothness or other desirable traits (e.g., sharp edges). When the reconstruction problem is posed as an inverse problem, prior pdfs are in fact not only useful but necessary as well, as reconstruction from projections is in general an ill-posed problem. As such, in order to proceed a form of regularization is required [8,9,10]. In this sense of posing tomographic reconstruction as an inverse problem, a prior pdf acts as an objective function regularizer. With respect to traits that are deemed desirable for the reconstruction, image smoothness is perhaps one of the most usual constraints. The image smoothness constraint is based on the preassumption that high-frequency components in the reconstruction are more likely to be present due to noise. A common smoothness prior is the Markov random field (MRF) prior, formulated as a Gibbs distribution. Numerous Gibbs/MRF-based methods have been proposed, using various potential functions [12, 13, 21]. Under Gibbs/MRF priors, local differences between neighboring pixels are penalized in order to encourage small values, hence enforcing output image smoothness.
Another popular prior is the total variation (TV) prior [6, 11, 25, 28, 36, 39], with application to many imaging tasks, including tomographic reconstruction. The motivation behind using TV priors is that they provide a principled, coherent way to favor a smooth solution with sharper edges, in the premise that such solutions should have lower total intensity variation [6]. The penalty associated with total variation is in practice typically formulated as the discrete sum of gradient norms over the reconstructed image. In tomographic reconstruction with TV priors, the objective function is the weighted sum of the TV prior term and a data term. The data term is formulated as the distance between the reconstructed object and the original projections [11, 36]. Prior models and associated algorithms that can’t readily be categorized as Gibbs/MRF or TV priors have otherwise been proposed [7, 15]. A notable example is the clustered intensity histogram algorithm [15], where Gamma-distributed priors are used to enforce emission positivity constraints.
More complex prior models are in general possible, formulated as hierarchical probabilistic models [3]. In hierarchical probabilistic models, multiple stochastic and deterministic parameters may interact to model the reconstruction as a data generation process. The expectation–maximization (EM) algorithm is typically used to solve hierarchical probabilistic models [30], provided the model is suitably constructed (for example, pdf conjugacy requirements need to be met [3]). The tomographic reconstruction model and algorithm proposed in the current work is a hierarchical probabilistic model, solved with a MAP-EM algorithm. We propose a spatially varying mixture model [22, 23, 27, 32, 33] based on a Gamma mixture prior. In order to account for the modeling of edges between image segments, appropriate MRF smoothness priors on pixel label mixing proportions are defined. These priors model edges as random, latent variables, an assumption known in the literature as a \({ line \; process}\) model. The proposed model is based on a MAP tomographic reconstruction formula that uses a Gamma mixture prior and encourages edge preservation by modeling edge presence as a Student’s t-distribution on the pixel contextual mixing proportions.
In the proposed model, parameters are automatically estimated from the observed data. In effect, this yields location-dependent smoothing which cannot be modeled by the standard Gibbs distribution. The image reconstruction process integrates this prior in a standard MAP-EM iterative algorithm, where the proposed spatially varying Gamma mixture model (SVGammaMM) parameters and the unknown image are estimated in an alternating scheme. Numerical experiments using photon-limited images validate the method’s usefulness with respect to results using common state-of-the-art MRF-based priors.
The remainder of this paper is organized as follows. The image formation model and proposed reconstruction method is presented in Sect. 2. We present numerical results in Sect. 3, testing our model and comparing it against other well-known reconstruction models. In Sect. 4, we discuss our conclusions and perspectives for future work.
2 Image Formation Model and Image Reconstruction
Let \(\mathbf{f}\) be the vectorized form of the image to be reconstructed. Let also \(\mathbf{g}\) be the observed projections (sinogram), also in vectorized form, and let \(\mathbf{H}\) represent the projection matrix. Penalized likelihood models rely on a stochastic interpretation of reconstruction regularization by introducing an appropriate prior \(p(\mathbf{f})\) for the image \(\mathbf{f}\). The likelihood function \(p(\mathbf{g}|\mathbf{f})\) is related to the posterior probability \(p(\mathbf{f}|\mathbf{g})\) by the Bayes rule, \(p(\mathbf{f}|\mathbf{g})\propto p(\mathbf{g}|\mathbf{f})p(\mathbf{f})\).
In emission tomography, the likelihood \(p(\mathbf{g}|\mathbf{f})\) is a Poisson distribution if the detector counts are mutually independent and not corrupted by additional data errors. Formally, we can write:
with respect to \(\mathbf{f}\), where M is the number of projection measurements, \(g_{m}\) is the \(m^{\text {th}}\) component of \(\mathbf{g}\) and \([\mathbf{Hf}]_{m}\) is the \(m^{\text {th}}\) component of the vector \(\mathbf{Hf}\). MAP estimates for the image \(\mathbf{f}\) may be obtained by maximizing the log-posterior:
We assume that the N intensity values \(\{ { f}_n \}_{n=1}^N\) of the reconstructed image \(\mathbf{f}\) are independent and identically distributed. These can be assumed to be generated by a finite mixture model [20]:
where \(\pi =\lbrace \pi _{\mathbf{j}} \rbrace _{\mathbf{j}=1}^{\mathbf{J}}\) is the prior probability vector of a pixel membership on class j. The \(\lbrace \theta _j \rbrace _{j=1}^J\) parameters control the shape of the kernel functions \(\phi \). Thus, there is a natural correspondence between pixel class-membership and kernels, and we can classify the pixels according to posterior class memberships [32]. A standard and well-known choice of a kernel function are the Gaussian distribution [32] or the Gamma distribution [3]. The Gamma distribution is nonzero valued on positive real numbers only, the positivity constraint of the reconstruction problem is inherently satisfied. The general form of a mixture-of-Gammas prior pdf is:
In Eq. (4), the set of parameters \(\theta \) involves the vector q, with elements \(\{ q_j \}_{j=1, \ldots , J}\) and the vector r, \(\{ r_j \}_{j=1, \ldots , J}\), which parameterize the gamma density. The gamma density is defined only for \(\mathbf{f}>0\). Furthermore, \(r_j>0\) is the mean and \(r_j^2/q_j\) the variance of the j-th component. Like other positivity preserving priors, it motivates a slight bias due to the difference between mean and mode. The mixing proportions (weights) \(\pi _j\) are positive and satisfy the constraint:
In this paper, we extend the above assumptions (Eqs. 3, 4, 5) which assumes global mixing weights \(\pi _j\) and assume that mixing proportions vary pixel-wise. Formally, we assume that the conditional distribution of \(\mathbf f\) given a latent variable \(\mathbf{z}\) is:
where the prior distribution for the latent variable \(\mathbf{z}\) is multinomial [32]:
where \({z}^n\) is a binary vector \([{z}^n_1 {z}^n_2 \ldots {z}^n_J]^T\) with a single component equal to 1, \({z}^n_j=1\) and all others equal to 0, i.e., a “one-hot” vector representation. Pixel n is estimated to be part of segment j if and only if \({z}^n_j=1\). Note that, while the vector \({z}^n\) is binary and one-hot, the corresponding \({\pi }^n\) vector \([{\pi }^n_1 {\pi }^n_2 \ldots {\pi }^n_J]^T\) is a probability vector with \(\pi ^n_j \in [0,1] \forall j \in [1,J]\) and \(\sum _{j=1}^J \pi ^n_j = 1\). Also, as a direct consequence of Eq. (7), the two vectors are related by \(<z^n_j> = \pi ^n_j\), where \(<\cdot>\) denotes random variable expectation.
Our model differs from the standard Gamma mixture model in the definition of the mixing proportions. More precisely, each pixel \(f_n\), \(n=1,\ldots ,N\) has a distinct vector of mixing proportions denoted by \(\pi _{j}^{n}\), \(j=1,\ldots ,J\), with J being the number of Gamma kernels. We call these parameters contextual mixing proportions to distinguish them from the mixing proportions of a standard Gamma mixture model.
Apart from inherently assuming an image segmentation by means of the Gamma mixture prior, the edges in the image are preserved because the local differences of the contextual mixing proportions are considered to follow a univariate Student’s t-distribution (or referred to as simply t-distribution for brevity). Following the definition of the Student’s t-distribution [18], we can write the Student’s t as a two-step generative process:
where \(\gamma _d(n)\) is the set of neighbors of the pixel indexed by n, with respect to the \(d{\mathrm{th}}\) adjacency type. This model first draws \(u_j^{nk}\) from a Gamma distribution parameterized by \(\nu _{jd}\) and then considers that the local differences of the mixing proportions follow a Gaussian distribution with zero mean and standard deviation \(\beta _{jd}^2/u_j^{nk}\). We consider D different adjacency types (e.g., horizontal, vertical, diagonal). For example, if d=“horizontal adjacency” and n is a non-border pixel, we have \(|\gamma _d(n)|=2\) as pixel n has one horizontal neighbor to its left pixel and one horizontal neighbor to its right. The total number of neighbors for all adjacency types \(d \in [1,D]\) is considered equal to \(\varGamma \).
Note that Eq. (8) effectively defines a Gibbs distribution on the contextual mixing proportions \(\pi \). The Gibbs distribution has the general form \(e^{-\int \beta V(f(x,y), \gamma (x,y))dxdy}\) where \(\beta \) and \(V(\cdot )\) are Gibbs parameters and clique potentials, respectively. In the current model, these are defined by a zero-centered normal distribution on neighbor differences (8), parametrized by \(\beta _{jd}\) and line process \(\mathbf{u}\). Note that this use is quite different from the typical use of the Gibbs distribution, which is defined directly over pixel value estimates.
Formally, the Student’s t-distribution is defined as
The t-distribution is heavy-tailed, which means that it is more likely to generate values that fall far from its mean. As the number of degrees of freedom \(\nu \) increases, the t-distribution approaches the normal distribution [3]. On the contrary, the closer \(\nu \) is to zero, the more heavy-tailed the distribution is.
The proposed generative model can be examined in Fig. 1. The model imposes edge preservation through the Student’s t-distribution on the difference of the mixing proportions. The variables \(u_j^{nk}\), effectively modeling a continuous line process, provide a detailed description of the boundary structure of the image.
Estimation of model parameters is obtained through a standard MAP-EM approach, consisting of a two-step alternating optimization scheme. In the first step, the parameters of the SVGammaMM model are estimated using the EM algorithm [32] given a fixed image estimate \(\mathbf {f}\). In the second step, the image estimate \(\mathbf {f}\) is updated while the model parameters are kept fixed.
The EM algorithm consists of performing alternating E-steps and M-steps until model likelihood convergence. In the E-step, moments of the hidden variables \(\mathbf{z}\) and \(\mathbf{u}\) are computed:
where t denotes iteration index. \(\psi (\cdot )\) denotes the digamma function, while z, \(\eta \) are determined by:
In the M-step, the non-stochastic parameters r, q, \(\beta \) are updated with:
In addition, the degrees of freedom parameter \(\nu _{jd}^{(t+1)}\) are estimated as the solution of the below equation:
Thereafter, the MAP estimate for the random process \(\pi \) is computed. The elements of \(\pi \) are updated by solving the following quadratic equation:
with coefficients:
While the vector elements of \(\pi \) are probability distributions, the real nonnegative solutions will not, in general, satisfy the sum-to-unity constraint. In order to get proper mixing weight vectors, we perform a projection step onto the constraints subspace using the quadratic programming algorithm described in [32].
After having updated the SVGammaMM parameters with the updates presented above, the next step consists in estimating the reconstruction \(\mathbf {f}\) given these parameters. An update for \(\mathbf {f}\) should maximize Eq. 2. The intensity of the \(n^{\text {th}}\) pixel of the unknown image can therefore be updated as follows:
This “one-step-late” (OSL) update [12, 38] uses the previous image estimate to evaluate the derivative term in Eq. (23).
Algorithm 1 summarizes the steps of the proposed tomographic reconstruction algorithm. The algorithm stops when the estimated image does not change significantly or when a predefined number of iterations is reached.
3 Numerical Results
The performance of the proposed spatially varying mixture models for the tomographic reconstruction problem was evaluated using the well-known Shepp-Logan phantom and a phantom consisting of three regions of relative intensities, represented by a hot disk, a cold disk and a background ellipse (Fig. 2). We call this phantom Elliptical phantom. The resolution of the reconstructed objects \(\mathbf{f}\), for both cases is \(256 \times 256\) pixels. Projections are observed for tilt angles {\(0^\circ \), \(1^\circ \), \(2^\circ \), \(\ldots \), \(179^\circ \)}. The observed sinograms \(\mathbf{g}\) are of size \(367 \times 180\), for both cases again. We have set \(J = 3\) and \(J = 5\) clusters for the mixture models taking into account the segments of the two phantoms. The algorithm stopped when \(\epsilon =10^{-3}\) or when 60 iterations were reached. The presented algorithm, namely the spatially varying Gamma mixture model (SVGammaMM), was evaluated with respect to the standard MLEM [35], the established MAP-EM algorithm with a Gibbs [12] and a TV prior [25]. The MLEM method solves the problem by performing a maximum likelihood fit over the Poisson-distributed emissions. The Gibbs and TV prior models penalize the solution according to a prior distribution model. The Gibbs prior term is defined as a MRF over the reconstruction itself [12], contrary to the proposed hierarchical model, as:
where \(\beta \) is a parameter controlling prior weight with respect to the data term. \(V(\cdot )\) is a clique potential function, penalizing reconstruction pixel values that are different from their neighbors, denoted here as \(\gamma (x,y)\). The TV prior term takes the form [6, 25]:
Furthermore, we have compared our algorithm to a spatially varying Gaussian mixture model (SVGMM) [24], a standard Gaussian mixture model (GMM) and a standard Gamma mixture model (GammaMM). The latter two models define likelihood according to Eq. 3, with different choices for the kernel function \(\phi \).
For the image estimation step (see algorithm 1) we have tried employing two alternatives. We ran tests either using an OSL update (Eq. 23) or preconditioned conjugate gradients (PCG) to compute the image estimate. Concerning parameters of the competing algorithms (e.g., the weight parameter between data term and total variation term in the TV prior method), tests were run over different parameter values, and the optimal ones, result-wise, were chosen to be presented for comparison.
A number of performance indices were used. To this end, degraded images were generated from the initial images by modifying the total photon counts. More specifically, images having 75, 55, 35 and 15 photons/pixel on average were generated to degrade the signal quality, and for the Elliptical phantom, 80, 56, 36 and 24 photons/pixel were simulated. Since the various performance indices are so similar for the whole set of photon counts, only comparative statistics for one rate of photons per pixel is illustrated (75 photons/pixel for the Shepp–Logan phantom and 80 photons/pixel for the Elliptical phantom).
Furthermore, we have artificially degraded the original phantoms by smoothing with a Gaussian blur filter. We have used two different levels of smoothing, \(\sigma =2\) and \(\sigma =4\). We have run tests corresponding to the original, non-smoothed phantoms as well as the smoothed phantoms, in order to test for our algorithm’s robustness also to this type of degradation.
At first, the algorithms were put in test in terms of the improvement in signal to noise ratio (ISNR) with respect to a reconstruction obtained by a simple filtered backprojection using the Ram–Lak filter:
where \(\mathbf{f}\) is the ground truth image, \(\mathbf{f}_{FBP}\) is the reconstructed image by filtered backprojection and \(\hat{\mathbf{f}}\) is the reconstructed image using the proposed image model. Practically, ISNR measures the improvement (or deterioration) in the quality of the reconstruction of the proposed method with respect to the reconstruction obtained by filtered backprojection.
Moreover, the consistency of the method was measured by the bias (BIAS) and the variance (VAR) of the reconstructed images:
with
where \({\mathbf{f}}\) is the ground truth image and \(\hat{\mathbf{f}}^{k}\), for \(k=1,\ldots ,K\), is the \(k^{\text {th}}\) reconstructed image, obtained from \(K=40\) different realizations for each noise level. Finally, we also considered for the evaluation the structural similarity index (SSIM) [34], which represents the visual distortion between the reconstructed image and the ground truth. It is computed to be close to 1 for all the resulted images except from MLEM and MAP-EM with Gibbs prior, hence it is not included in the illustrated comparative results.
The statistical comparisons for the aforementioned algorithms are shown in Fig. 3 for the Shepp–Logan phantom. For all the quantities, their mean values over the \(K=40\) experiments are shown. All of the obtained ISNR values are very close to the mean values as their standard deviations over the whole set of experiments are very small. The ISNR, bias and variance for the Shepp–Logan phantom, for 75 photons per pixel on average, are shown. The statistical comparisons for the Elliptical phantom for 80 photons per pixel are illustrated in Fig. 4. It may be observed from all these indices for both phantoms that the proposed spatially varying mixture model performs better with respect to the other priors. Specifically, for the Shepp–Logan phantom with 75 photons per pixel and for the Elliptical phantom with 80 photons, ISNR reaches its peak for the spatially varying Gamma mixture model, carried out by a preconditioned conjugate gradient optimizer (SVGammaMM (PCG)), as shown in Figs. 3a and 4a. The spatially varying Gamma mixture methods have similar ISNR values but always larger than the other algorithms. The proposed methods similarly outperform the competition for the case where we perform reconstructions over the blur-degraded versions of the phantoms, for either blur intensity (Figs. 5, 6, 7, 8).
The same stands for the bias which yields its minimum through the SVGammaMM (PCG) for both phantoms as it can be observed in Figs. 3c and 4c. Furthermore, the variance of the estimates is relatively consistent for the spatially varying mixture model for both phantoms. This is due to the mixture nature of the priors. The results confirm the effectiveness of the proposed model.
In general, the images provided by the proposed spatially varying mixture model are visually sharper. Estimated images obtained using this model (SVGammaMM, PCG) are illustrated in Figs. 9 and 10, where we show reconstructions on input data degraded with Gaussian blur (\(\sigma =2\)). For comparison, estimates obtained using MLEM, MAP-EM with a Gibbs prior and a TV prior and a filtered backprojections (FBP) estimate are shown in the same Figures. Note that the performance of SVGammaMM (OSL) is always very close to SVGammaMM (PCG), with only small differences, and with at times slightly better performance.
In addition, to highlight the accuracy of the proposed model, the estimated image intensities along a scan line for both phantoms are shown in Figs. 11 and 12. Results using the SVGammaMM (PCG) model and a MAP-EM with a Gibbs prior are compared. It can be observed that the spatially varying Gamma mixture model provides values which are closer to the ground truth.
Finally, the required execution time of the proposed models on a standard PC using MATLAB without any optimization is on average 4 minutes, which is explained by the number of required computations for the EM or PCG for the reconstruction and the EM for the estimation of the model parameters.
4 Conclusion and Future Work
The main goal of this work was to explore the possibility of the development of a robust model for tomographic image reconstruction purposes. We have proposed an alternative hierarchical and spatially constrained mixture model in order to enforce them to preserve image edges. We have presented a spatially varying Gamma mixture model with a Student’s t MRF-based prior governing the contextual mixing proportions. Spatially varying mixture models are characterized by the dependence of their mixing proportions on location (contextual mixing proportions) and they have been previously successfully used in image segmentation.
The main contribution of this work is the effectiveness of the adaptive MRF priors which may capture spatial coherence and preserve image boundaries, retaining them unsmoothed. The proposed Gamma mixture component of the model assumes a multiple-mode histogram in the object, and enforces positivity naturally. Furthemore, an important property of the proposed model is the automatic estimation of model parameters from the data which is crucial, as many state-of-the-art reconstruction algorithms rely on empirical parameter selection. Numerical experiments on various photon-limited image scenarios showed that the proposed models are more accurate than the corresponding results of the compared methods.
With regard to directions for future improvements, an important perspective of this study would be to explore how to automatically estimate the number of components of the mixture model assumed in the image reconstruction framework. This remains an important open problem in the mixture modeling literature [3, 29].
Furthermore, the space complexity of the proposed algorithm could perhaps be improved upon. Taking into account that for each iteration we require space for the Markov random field values \(\pi \) and the line process \(\mathbf {u}\), we can easily see that our space complexity is \(O(NJ\varGamma )\). For values \(J=5\) and a standard \(8-\) neighborhood, this may mean a requirement of up to \(\sim 50\) times the space of the reconstruction. Perhaps a grid-update strategy [31] could be envisaged and improved upon to save on required space.
References
Ahmadi, S., Rajabi, H., Sardari, D., Babapour, F., Rahmatpour, M.: Attenuation correction in SPECT during image reconstruction using inverse Monte Carlo method a simulation study. Rom. Rep. Phys. 66, 200–211 (2014)
Arcadu, F., Nilchian, M., Studer, A., Stampanoni, M., Marone, F.: A forward regridding method with minimal oversampling for accurate and efficient iterative tomographic algorithms. IEEE Trans. Image Process. 25(3), 1207–1218 (2016)
Bishop, C.: Pattern Recognition and Machine Learning. Springer, Berlin (2006)
Brendel, B., Teuffenbach, M., No, P.B., Pfeiffer, F., Koehler, T.: Penalized maximum likelihood reconstruction for x-ray differential phase-contrast tomography. Med. Phys. 43(1), 188–194 (2016)
Browne, J.A., Pierro, A.R.D.: A row-action alternative to the EM algorithm for maximizing likelihood in emission tomography. IEEE Trans. Med. Imaging 15, 687–699 (1996)
Burger, M., Osher, S.: A guide to the TV zoo. In: Level Set and PDE Based Reconstruction Methods in Imaging, pp. 1–70. Springer (2013)
Chen, Y., Ma, J., Feng, Q., Luo, L., Shi, P., Chen, W.: Nonlocal prior Bayesian tomographic reconstruction. J. Math. Imaging Vis. 30, 133–146 (2008)
Estellers, V., Soatto, S.: Detecting occlusions as an inverse problem. J. Math. Imaging Vis. 54(2), 181–198 (2016)
Garrigos, G., Rosasco, L., Villa, S.: Iterative regularization via dual diagonal descent. J. Math. Imaging and Vis.(2017)
Golub, G.H., Hansen, P.C., O’Leary, D.P.: Tikhonov regularization and total least squares. SIAM J. Matrix Anal. Appl. 21(1), 185–194 (1999)
Goris, B., Van den Broek, W., Batenburg, K.J., Mezerji, H.H., Bals, S.: Electron tomography based on a total variation minimization reconstruction technique. Ultramicroscopy 113, 120–130 (2012)
Green, P.J.: Bayesian reconstructions from emission tomography data using a modified EM algorithm. IEEE Trans. Med. Imaging 9(1), 84–93 (1990)
Higdon, D.M., Bowsher, J.E., Johnson, V.E., Turkington, T.G., Gilland, D.R., Jaczszak, R.J.: Fully Bayesian estimation of Gibbs hyperparameters for emission computed tomography data. IEEE Trans. Med. Imaging 16(5), 516–526 (1997)
Hm, H.M.H., Larkin, B.S.: Accelerated image reconstruction using ordered subsets of projection data. IEEE Trans. Med. Imaging 13, 601–9 (1994)
Hsiao, I.T., Rangarajan, A., Gindi, G.: Joint MAP Bayesian tomographic reconstruction with a Gamma-mixture prior. IEEE Trans. Med. Imaging 11(12), 1466–1475 (2002)
Idier, J.: Bayesian Approach to Inverse Problems. Wiley, Hoboken (2013)
Kim, D., Ramani, S., Fessler, J.A.: Combining ordered subsets and momentum for accelerated X-ray CT image reconstruction. IEEE Trans. Med. Imaging 34(1), 167–178 (2015)
Kotz, S., Balakrishnan, N., Johnson, N.: Continuous Multivariate Distributions. Wiley, New York (2000)
Lei, J., Liu, S.: An image reconstruction algorithm based on the regularized minimax estimation for electrical capacitance tomography. J. Math. Imaging Vis. 39(3), 269–291 (2011)
McLachlan, G.: Finite Mixture Models. Wiley, Hoboken (2000)
Mumcuoglu, E.U., Leahy, R., Cherry, S.R., Zhou, Z.: Fast gradient-based methods for Bayesian reconstruction of transmission and emission PET images. IEEE Trans. Med. Imaging 13(4), 687–701 (1994)
Nikou, C., Galatsanos, N., Likas, A.: A class-adaptive spatially variant mixture model for image segmentation. IEEE Trans. Image Process. 16(4), 1121–1130 (2007)
Nikou, C., Likas, A., Galatsanos, N.: A Bayesian framework for image segmentation with spatially varying mixtures. IEEE Trans. Image Process. 19(9), 2278–2289 (2010)
Papadimitriou, K., Nikou, C.: Tomographic image reconstruction with a spatially varying Gaussian mixture prior. In: 2015 IEEE International Conference on Image Processing (ICIP), pp. 4002–4006 (2015)
Rudin, L.I., Osher, S., Fatemi, E.: Nonlinear total variation based noise removal algorithms. Physica D 60(1–4), 259–268 (1992)
Sabne, A., Wang, X., Kisner, S.J., Bouman, C.A., Raghunathan, A., Midkiff, S.P.: Model-based iterative CT image reconstruction on GPUs. In: Proceedings of the 22nd ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, pp. 207–220 (2017)
Sanjay-Gopal, S., Hebert, T.: Bayesian pixel classification using spatially variant finite mixtures and the generalized EM algorithm. IEEE Trans. Image Process. 7(7), 1014–1028 (1998)
Sawatzky, A., Brune, C., Kösters, T., Wuebbeling, F., Burger, M.: EM-TV methods for inverse problems with Poisson noise. In: Level Set and PDE Based Reconstruction Methods in Imaging, pp. 71–142. Springer (2013)
Sfikas, G., Nikou, C., Galatsanos, N., Heinrich, C.: Majorization-minimization mixture model determination in image segmentation. In: 2011 IEEE conference on Computer Vision and Pattern Recognition (CVPR), pp. 2169–2176 (2011)
Sfikas, G., Nikou, C., Galatsanos, N.: Edge preserving spatially varying mixtures for image segmentation. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR). Anchorage, Alaska, USA (2008)
Sfikas, G., Nikou, C., Heinrich, C., Galatsanos, N.: On the optimization of probability vector MRFS in image segmentation. In: IEEE International Workshop on Machine Learning for Signal Processing (MLSP), pp. 1–6. IEEE (2009)
Sfikas, G., Nikou, C., Galatsanos, N., Heinrich, C.: Spatially varying mixtures incorporating line processes for image segmentation. J. Math. Imaging Vis. 36(2), 91–110 (2009)
Sfikas, G., Heinrich, C., Zallat, J., Nikou, C., Galatsanos, N.: Recovery of polarimetric stokes images by spatial mixture models. J. Opt. Soc. Am. A 28(3), 465–474 (2011)
Sheikh, H.R., Wang, Z.: Bovik, A.C., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004)
Shepp, L., Vardi, Y.: Maximum likelihood reconstruction for emission tomography. IEEE Trans. Med. Imaging 1, 113–122 (1982)
Sidky, E.Y., Pan, X.: Image reconstruction in circular cone-beam computed tomography by constrained, total-variation minimization. Phys. Med. Biol. 53(17), 4777 (2008)
Wang, G., Qi, J.: Pet image reconstruction using kernel method. IEEE Trans. Med. Imaging 34(1), 61–71 (2015)
Wernick, M.N., Aarsvold, J.N.: Emission Tomography: the Fundamentals of PET and SPECT. Elsevier, Atlanta (2004)
Zhang, H., Wang, L., Yan, B., Li, L., Cai, A., Hu, G.: Constrained total generalized p-variation minimization for few-view X-ray computed tomography image reconstruction. PLoS ONE 11(2), 1–28 (2016)
Zhang, Y., Wang, Y., Zhang, W., Lin, F., Pu, Y., Zhou, J.: Statistical iterative reconstruction using adaptive fractional order regularization. Biomed. Opt. Express (OSA) 7(3), 1015–1029 (2016)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Papadimitriou, K., Sfikas, G. & Nikou, C. Tomographic Image Reconstruction with a Spatially Varying Gamma Mixture Prior. J Math Imaging Vis 60, 1355–1365 (2018). https://doi.org/10.1007/s10851-018-0817-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10851-018-0817-x