Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

Diffusion-weighted MRI is a key technique in studying the neural architecture and connectivity of the brain. It can be utilized as imaging-based biomarkers for investigating several brain disorders such as Alzheimer’s disease, schizophrenia, mild traumatic brain injury, etc. [1]. In many clinical applications such as neurosurgical planning and deep brain stimulation, it is critically important to use high-spatial-resolution diffusion images to accurately localize brain structures, especially those that are very small (such as substantia nigra and sub-thalamic nucleus). Moreover, HR images are critical for tracing small white-matter fiber bundles and to reduce partial volume effects. Further, with high spatial resolution, gray-matter and white-matter structures can be better resolved, especially in neonate and infant brains. The typical voxel size of a dMRI image acquired from a clinical scanner is about \(1.7^3\) to \(2^3~\,mm^3\) which is too large to study certain brain structures that are a few millimeters thick. Due to signal loss from \(T_2\) decay with longer echo times, reducing the voxel size leads to a proportionate decrease in the signal-to-noise ratio (SNR). Though SNR could be enhanced by averaging multiple acquisitions, the total acquisition time may be too long to apply this approach in clinical settings.

Existing techniques that obtain high-resolution (HR) dMRI can be classified into two categories based on their data acquisition scheme. The first group of methods obtain HR data using a single LR image via intelligent interpolations or regularizations. These types of methods have been investigated in structural MRI [2] and in dMRI to enhance anatomical details as in [3]. However, as pointed out in [3], the performance of some of these methods still largely relies on the information contained in the original LR image. The second group of methods require multiple LR images acquired according to a specific sampling scheme to reconstruct a HR image. Each of the LR image is modeled as the measurement of an underlying HR image via a down-sampling operator. Then the HR image is estimated by solving a linear inverse problem with suitable regularization. These methods use the classical concept of super-resolution reconstruction (SRR) [4]. In dMRI, these methods were used in [5, 6] to reconstruct each diffusion weighted volume independently. A different acquisition scheme with the LR images having orthogonal slice acquisition direction was proposed in [7]. However, the distortions from LR scans with different slice directions need to be corrected prior to applying the SRR algorithm, which involves complex non-linear spatial normalization. Moreover, each diffusion-weighted image (DWI) volume is reconstructed independently, requiring the LR images to be acquired or interpolated on the same dense set of gradient directions. Thus, these methods require the same number of measurements (e.g., 60 gradient directions) for each LR acquisition. To address this problem, more recently, [8] introduced a method that used the diffusion tensor imaging (DTI) technique to model the diffusion signal in q-space. However, a very simplistic diffusion tensor model was assumed, which is not appropriate for modeling more complex diffusion phenomena (crossing fibers). A similar method has been proposed in [9] to improve distortion corrections for DWI’s using interlaced q-space sampling.

Fig. 1.
figure 1

An illustration of the CS-SRR scheme: a high-resolution image is reconstructed using three overlapping thick-slice volumes with down-sampled diffusion directions.

Our Contributions: We propose a compressed-sensing-based super-resolution reconstruction (CS-SRR) approach for reconstructing HR diffusion images from multiple sub-pixel-shifted thick-slice acquisitions. As illustrated in Fig. 1, we use three LR images with anisotropic voxel sizes that are shifted in the slice acquisition direction. Each LR image is acquired with a different (unique) set of gradient directions to construct a single HR image with isotropic voxel size and a combined set of gradient directions. This is in contrast to classical SRR techniques which require each LR scan to have the same set of gradient directions. In the proposed framework, only a subset (one-third) of the total number of gradient directions are acquired for each LR scan, which reduces scan time significantly while making the technique robust to head motion. To account for the correlation between diffusion signal along different gradients, we represent the signal in the HR image in a sparsifying basis of spherical ridgelets. To obtain the HR image, we propose a framework based on the ADMM algorithm, where we enforce sparsity in the spherical-ridgelet domain and spatial consistency using total variation regularization. We perform quantitative validation of our technique using very high resolution data acquired on a clinical 3 T scanner.

2 Background

We consider diffusion images acquired on single spherical shell in q-space [10]. Hence, the problem of super resolution reduces to that of reconstructing the spherical signal at each voxel of the HR image. Before introducing the proposed method, we provide a brief note on the concepts of compressed sensing and spherical ridgelets that are used to model and estimate the diffusion signals.

Compressed Sensing: The theory of compressed sensing provides the mathematical foundation for accurate recovery of signals from set of measurements far fewer than that required by the Nyquist criteria [11]. In the CS framework, a signal of interest \({\varvec{s}}\in {\mathbb R}^n\) is represented in an over-complete basis \(\varPhi \in {\mathbb R}^{n\times m}\) with \(n\ll m\) such that \({\varvec{s}}=\varPhi {\varvec{c}}\) with \({\varvec{c}}\in {\mathbb R}^m\) being a sparse vector, i.e. \({\varvec{c}}\) only has few non-zero elements. Let \({\varvec{y}}\in {\mathbb R}^p\) denote a measurement vector given by \( {\varvec{y}}=\varPsi {\varvec{s}}+{\varvec{\mu }}=\varPsi \varPhi {\varvec{c}}+{\varvec{\mu }}\) where \({\varvec{\mu }}\) is the noise component and \(\varPsi \in {\mathbb R}^{p\times n}\) is a down-sampling operator. If the matrix \(\varPsi \varPhi \) satisfies certain incoherent properties, then the CS theory [11] asserts that the sparse vector \({\varvec{c}}\) can be robustly recovered by solving the \(\ell _1\)-norm regularized problem

$$\begin{aligned} \min _{\varvec{c}}\frac{1}{2} \Vert {\varvec{y}}-\varPsi \varPhi {\varvec{c}}\Vert ^2+\lambda \Vert c\Vert _1 \end{aligned}$$
(1)

for a suitable \(\lambda \ge 0\). Hence, the signal \({\varvec{s}}\) can be accurately recovered as \(\varPhi {\varvec{c}}\) using the lower dimensional measurement \({\varvec{y}}\). This CS methodology has been used to estimate the dMRI signal from reduced number of measurements [1214] with the basis given by the spherical ridgelets.

Spherical Ridgelets: Spherical ridgelet functions were introduced in [12] as a frame to represent \(L_2\) functions on the sphere. Each spherical ridgelet function is specially designed to represent diffusion signals with a particular orientation and certain degree of anisotropy. A sparse combination of such basis functions is suitable to model the diffusion data with complex fiber orientations [13, 14]. Specifically, the spherical ridgelets are constructed as follows: For \(x\in {\mathbb R}_+\) and \(\rho \in (0, 1)\), let \(\kappa (x)=\exp \{-\rho x(x+1)\}\) be a Gaussian function. Further, we let

$$\begin{aligned} \kappa _j(x)=\kappa (2^{-j}x)=\exp \left\{ -\rho \frac{x}{2^j}\left( \frac{x}{2^j}+1\right) \right\} \text{ for } j=0,1,2, \ldots . \end{aligned}$$

Then the spherical ridgelets with their energy spread around the great circle supported by a unit vector \({\varvec{v}}\in {\mathbb S}^2\) is given by

$$\begin{aligned} \psi _{j,{\varvec{v}}}({\varvec{u}})=\frac{1}{2\pi }\sum _{n=0}^{\infty }\frac{2n+1}{4\pi }\lambda _n(\kappa _{j+1}(n)-\kappa _j(n))P_n({\varvec{u}}\cdot {\varvec{v}}), \forall {\varvec{u}}\in {\mathbb S}^2 \end{aligned}$$
(2)

where \(P_n\) denotes the Legendre polynomial of order n, \(\kappa _{-1}(n)=0\), \(\forall n\) and \(\lambda _n=2\pi (-1)^{n/2}\frac{1.3...(n-1)}{2.4\ldots n}\) if n is even, otherwise \(\lambda _n=0\).

To construct a finite over-complete dictionary, we follow the method in [13, 14] and restrict the values for the resolution index j to the set \(\{-1, 0, 1\}\). For each resolution index j, the set of all possible orientations \({\varvec{v}}\in {\mathbb S}^2\) is discretized to a set of \(M_j\) directions on the unit sphere with \(M_{-1}=25\), \(M_{0}=81\) and \(M_1=289\). To this end, the set of over-complete spherical ridgelets is given by \(\varPsi _\mathrm{{SR}}=\left\{ \phi _{j, {\varvec{v}}_{j}^i} \mid j=-1, 0, 1, i=1, 2, \ldots , M_j \right\} \). Since the spherical ridgelets are anisotropic, the isotropic diffusion signal in the CSF areas is not efficiently modeled as a sparse combination of the basis functions. To resolve this issue, we expand the basis functions by adding one isotropic element as \(\varPsi :=\left\{ \psi _\mathrm{iso}, \varPsi _\mathrm{SR}\right\} \) where \(\psi _\mathrm{iso}\) denotes a constant function on the unit sphere. For convenience, we denote the functions in \(\varPsi \) as \(\psi _1, \psi _2, \ldots , \psi _M\) with \(\psi _1=\psi _\mathrm{iso}\). Given the diffusion signal measured along N diffusion directions \(\{{\varvec{u}}_n\}_{n=1}^N\), we construct a basis matrix A with \(A_{nm}=\psi _m({\varvec{u}}_n)\) for \(m=1, \ldots , M\) and \(n=1, \ldots , N\).

3 Method

Let \(L_k\), \(k=1, \ldots , K\), denote the K low-resolution diffusion weighted imaging (DWI) volumes. For each \(L_k\), the diffusion signal is acquired along a set of gradient directions \(U_k=\{{\varvec{u}}_1^{(k)}, \ldots , {\varvec{u}}^{(k)}_{N_k}\}\) at the same b-value. The set of gradient directions for each LR image \(L_k\) is assumed to be different, i.e. \(U_k \cap U_\ell =\emptyset \) for \(k\ne \ell \). The HR image that has \(n_x\times n_y\times n_z\) voxels and \(N=\sum _{k=1}^K N_k\) gradient directions is denoted by a matrix H of size \(n_x n_y n_z\times N\). Then, each LR image \(L_k\) is given by

$$\begin{aligned} L_k=D_kB_k H Q_k^T+ \mu _k \text{ for } k=1, \ldots , K, \end{aligned}$$
(3)

where \(D_k\) denotes the down-sampling matrix that averages neighboring slices, \(B_k\) denotes the blurring (or point-spread function) and \(Q_k\) is the sub-sampling matrix in q-space. The main difference between the above model and the one used in [6, 15] is given by the \(Q_k\)’s that allow the LR images to have different sets of gradient directions. Further, we use the spherical ridgelets basis to model the diffusion signal at each voxel of H. To this end, H is assumed to satisfy \(H=VA^T\) with A being the basis matrix of spherical ridgelets (SR) constructed along the set of gradient directions \(\{U_1, \ldots , U_K\}\) and each row vector of V being the SR coefficients at the corresponding voxel. Since each basis function is designed to model a diffusion signal with certain degree of anisotropy, a non-negative combination of spherical ridgelets provides more robust representation for diffusion signals especially when the SNR is low. Hence, the coefficients V are restricted to be non-negative. Following (1), we estimate H by solving

$$\begin{aligned} \min _{H,V\ge 0} \left\{ \frac{1}{2} \sum _{k=1}^K \Vert L_k-D_kB_k H Q_k^T\Vert ^2+ \lambda \Vert W\circ V\Vert _1\right\} \text{ s.t. } H=VA^T \end{aligned}$$
(4)

where \(\circ \) denotes the element-wise multiplication, \(\Vert M\Vert ^2\) is the squared Frobenius norm of a matrix M and \(\Vert M\Vert _1\) is the summation of the absolute values of all the entries of M. W is the weighting matrix with each row having the form \([w, 1, \ldots , 1]\) where w is used to adjust the penalty for choosing the isotropic basis function in a voxel. The value of w can be obtained a-priori from a probabilistic segmentation of the T1-weighted image of the brain. Thus, w is small in the CSF and large in white and gray-matter areas.

3.1 Total-Variation (TV) Regularization

Let \({\varvec{h}}_{{\varvec{r}}}({\varvec{u}}_n)\) denote the diffusion signal along the direction \({\varvec{u}}_n\) at the voxel \({\varvec{r}}\in \varOmega \) with \(\varOmega \) being the set of all voxels of the HR image. The diffusion signal along the gradient \({\varvec{u}}_n\) in all voxels forms a 3D image volume denoted by \(H_n\). The correlation of the diffusion signal in neighboring voxels implies that \(H_n\) is spatially smooth. A standard technique to utilize this fact is to minimize the TV semi-norm of \(H_n\) defined as \( \Vert H_n\Vert _\mathrm{TV}=\sum _{{\varvec{r}}\in \varOmega }\left[ \sum _{{\varvec{p}}\in {\mathcal N}({\varvec{r}})}|{\varvec{h}}_{{\varvec{r}}}({\varvec{u}}_n)- {\varvec{h}}_{{\varvec{p}}}({\varvec{u}}_n)|^2 \right] ^{1/2}, \) where \({\mathcal N}({\varvec{r}})\) denotes a set of neighbors around \({\varvec{r}}\). For a collection of image volumes \(H=\{H_n\}_{n=1}^N\), the TV semi-norm of H is defined as the sum of the TV semi-norm of each image volume \(H_n\), i.e. \(\Vert H\Vert _\mathrm{TV}=\sum _{n=1}^N \Vert H_n\Vert _\mathrm{TV}\). By adding a regularization term for \(\Vert H\Vert _\mathrm{TV}\) in (4), we rewrite the optimization problem as

$$\begin{aligned} \min _{V\ge 0,H} \left\{ \frac{1}{2} \sum _{k=1}^K \Vert L_k-D_kB_k H Q_k^T\Vert ^2+\lambda _1 \Vert W\circ V\Vert _1+\lambda _2\Vert H\Vert _\mathrm{TV}\right\} \text{ s.t. } H=VA^T \end{aligned}$$
(5)

where the positive parameters \(\lambda _1, \lambda _2\) determine the relative importance of the data fitting terms versus the sparsity and the total regularization terms. Next, we introduce an efficient algorithm for solving (5) based on the alternating directions method of multipliers (ADMM) [16].

3.2 Optimization Algorithm

The optimization problem (5) typically involves high dimensional optimization variables. A suitable implementation of the ADMM algorithm may distribute the computational cost and decompose the optimization into a sequence of simpler problems. First, we note that (5) can be equivalently written as

$$\begin{aligned} \min _{V\ge 0,H,Z}&\bigg \{\frac{1}{2}\sum _{k=1}^K \Vert L_k-D_kB_kHQ_k^T\Vert ^2+\lambda _1\Vert W\circ V\Vert _{1}+\lambda _2\Vert Z\Vert _\mathrm{TV}\nonumber \\&+\frac{\rho _1}{2}\Vert H-VA^T\Vert ^2+\frac{\rho _2}{2}\Vert H-Z\Vert ^2\bigg \} \text{ s.t. } H-VA^T=0,~H-Z=0, \nonumber \end{aligned}$$

where Z is an auxiliary variable that equals to H, and the augmented terms \(\frac{\rho _1}{2}\Vert H-VA^T\Vert ^2+\frac{\rho _2}{2}\Vert H-Z\Vert ^2\) with \(\rho _1, \rho _2\ge 0\) do not change the optimal value. Let \(\varLambda _1, \varLambda _2\) be the multipliers for \(H-VA^T=0\) and \(H-Z=0\), respectively. Then, each iteration of the ADMM algorithm consists of several steps of alternately minimizing the augmented Lagrangian \(\bigg \{\frac{1}{2}\sum _{k=1}^K \Vert L_k-D_kB_kHQ_k^T\Vert ^2+\lambda _1\Vert W\circ V\Vert _{1}+\lambda _2\Vert Z\Vert _\mathrm{TV}+\frac{\rho _1}{2}\Vert H-VA^T+\varLambda _1\Vert ^2+\frac{\rho _2}{2}\Vert H-Z+\varLambda _2\Vert ^2\bigg \}\) over HVZ and one step of updating \(\varLambda _1, \varLambda _2\). More specifically, let \(V^t, H^t, Z^t\) and \(\varLambda _1^t, \varLambda _2^t\) denote the values of these variables at iteration t. Then, iteration \((t+1)\) consists of two steps of estimating \(\{V^{t+1}, Z^{t+1}\}\) and \(H^{t+1}\) by solving

$$\begin{aligned}&\underset{V\ge 0,Z}{\min } \bigg \{\lambda _1\Vert W\circ V\Vert _{1}+\lambda _2\Vert Z\Vert _\mathrm{TV}+\frac{\rho _1}{2}\Vert H^{t}-VA^T+\varLambda _1^{t}\Vert ^2+\frac{\rho _2}{2}\Vert H^{t}-Z+\varLambda _2^{t}\Vert ^2 \bigg \} \end{aligned}$$
(6)
$$\begin{aligned}&\underset{H}{\min } \bigg \{\frac{1}{2}\sum _{k=1}^K \Vert L_k-D_kB_kHQ_k^T\Vert ^2+\frac{\rho _1}{2}\Vert H-V^{t+1}A^T+\varLambda _1^t\Vert ^2+\frac{\rho _2}{2}\Vert H-Z^{t+1}+\varLambda _2^t\Vert ^2\bigg \} \end{aligned}$$
(7)

and one step of updating the multipliers as

$$\begin{aligned} \varLambda _1^{t+1}=\varLambda _1^t+(H^{t+1}-V^{t+1}A^T),~\varLambda _2^{t+1}=\varLambda _2^t+(H^{t+1}-Z^{t+1}) . \end{aligned}$$

A typical stopping criteria is to check if \(\varLambda _1^t\) and \(\varLambda _2^t\) have “stopped changing”, i.e. \(\Vert \varLambda _1^{t+1}-\varLambda _1^t\Vert ^2\le \epsilon _1\) and \(\Vert \varLambda _2^{t+1}-\varLambda _2^t\Vert ^2\le \epsilon _2\) for some user defined choice of \(\epsilon _1, \epsilon _2\).

We point out some important notes for the above iterative algorithm: (1) Problem (6) can be decomposed into two independent optimization problems on V and Z, respectively. The update for \(V^t\) is obtained by solving an \(\ell _1\)-minimization problem. In particular, the solution for each voxel can be obtained independently (parallely). We have also developed an ADMM-based algorithm for solving the non-negative weighted \(\ell _1\)-regularization problem, which is not presented here due to page limitations. (2) The update for \(Z^t\) is a standard TV denoising problem. (3) Problem (7) is a least-squared problem which needs a matrix inversion to compute the closed-form solution. Though, in general a huge matrix needs to be inverted, in this particular situation where the LR images are sub-pixel-shifted scans with \(B_k\)’s representing a blurring operation along the slice-selection direction and \([Q_1^T, \ldots , Q_K^T]\) is a permutation matrix, the computation reduces to inverting K matrices of size \(n_z \times n_z\), i.e. \(\left( B_k^TD_k^TD_kB_k+(\rho _1+\rho _2)I\right) \) for \(k=1, \ldots , K\), which can be easily done on standard workstations. We also note that in extreme situations, when the matrix size is too large, the steepest descent iterative method in [6] or the conjugate gradient method can be used alternatively.

4 Experiments

We tested the performance of the proposed method using two experiments. In the first one, we used artificially generated thick-slice acquisitions based on the high resolution data set from the Human Connectome Project (HCP) while the second experiment consisted of an actual validation setup where both the LR and HR images were acquired on a 3T Siemens clinical scanner. We compared the recovered CS-SRR result with the corresponding gold-standard (GS) HR data using the following metrics:

  • FA and Trace difference: Whole brain multi-tensor tractography [17] was computed on the GS and the reconstructed data, and several fiber bundles were extracted. The average fractional anisotropy (FA) and trace for tensors along the fiber bundles were computed.

  • Fiber-bundle overlap: The Bhattacharyya coefficient described in [18] was used to compare the fiber bundle overlap. This measure ranges between [0, 1], with 0 being no overlap, while 1 being complete overlap of fibers.

4.1 Evaluation on HCP Data

In this experiment, we used HCP data, which has spatial resolution of \(1.25\times 1.25\times 1.25~\,mm^3\) with 90 gradient directions at \(b=2000~s/mm^2\). To construct LR images, we first artificially blurred the DWI volumes along the slice direction using a Gaussian kernel with full width at half maximum (FWHM) of \(1.25~\,mm\). Then the data was down-sampled by averaging three contiguous slices to obtain a single thick-slice volume with spatial resolution of \(1.25 \times 1.25 \times 3.75~\,mm^3\). Similarly, two additional LR volumes were obtained so that all the three LR volumes were slice-shifted in physical space (see Fig. 1). The thick-slice volumes were also sub-sampled in q-space so that each set had 30 unique gradient directions. We also obtained a segmentation of the brain into three tissue types, namely, gray, white and CSF, from the \(T_1\)-weighted MR images. This tissue classification was used as a prior to set \(w=10^{-4}\) in CSF and \(w=1\) in gray and white matter areas. The auxiliary parameter \(\rho _1\) was set to \(\rho _1=1\).

To compare the tractography results, we first obtained the whole-brain tractography using the method in [17] for the GS and the reconstructed data sets, respectively. Next, we extracted the cingulum bundle (CB), the corticospinal tract (CST), the superior longitudinal fasiculus II (SLF-II) and a sub-part of the corpus callosum called caudal middle frontal bundle (CC-CMF), respectively, using the white matter query language (WMQL) [19] which uses Freesurfer cortical parcellations. The four pairs of fiber bundles are shown in Fig. 2a to d, respectively. The fiber-bundle overlap measure is very close to one, indicating a significant overlap between the fiber bundles obtained using the GS and CS-SRR data. Further, the estimated average FA and trace are very similar for both data sets.

Fig. 2.
figure 2

(a), (b) (c) and (d) are the tractography results for the cingulum bundle (CB), the corticospinal tract (CST), the superior longitudinal fasciculus II (SLF-II) and the corpus callosum caudal middle frontal (CC-CMF) bundle (with red and yellow tracts obtained using the GS and CS-SRR data sets respectively). The comparison metrics for the tractography results are shown in the following table.

 

GS FA

CS-SRR FA

GS Trace

CS-SRR Trace

Fiber-bundle overlap

CB

0.54

0.52

\(2.1\times 10^{-3}\)

\(2.0\times 10^{-3}\)

0.97

CST

0.69

0.66

\(2.0\times 10^{-3}\)

\(2.0\times 10^{-3}\)

0.97

SLF-II

0.63

0.60

\(2.0\times 10^{-3}\)

\(2.0\times 10^{-3}\)

0.95

CC-CMF

0.63

0.61

\(2.2\times 10^{-3}\)

\(2.1\times 10^{-3}\)

0.95

Another important goal of this work is to demonstrate the advantage of using high-resolution DWI image in studying small white-matter fascicles. For this purpose, we generated a low-resolution DWI volume by averaging every \(2\times 2\times 2\) neighboring voxels from the GS data to obtain a LR dMRI data with isotropic voxel size of \(2.5^3 \; mm^3\). Figure 3a–c shows the single tensor glyphs color coded by the direction of the dominant eigenvector for the GS, LR and SRR data sets in the same brain region from a coronal slice. Linear interpolations were used to increase the spatial resolution of the LR image. The background of the glyphs is the corresponding \(T_1\) weighted image. As pointed out in the rectangular area in Fig. 3b, due to partial-volume effects, the LR image was not able to capture the fine curvature of of fiber bundles near the gray matter areas. Figure 3d–f show the tractography results for tracts originating from the sub-thalamic nucleus with fibers being color-coded by orientation. While the tracts obtained from the GS and CS-SRR data are very similar, as pointed out by the arrow in Fig. 3e, most of the green fibers are missing in the LR image, though the number of seeds in the LR image is 8 times higher than the HR images.

Fig. 3.
figure 3

(a), (b) and (c) show the single tensor glyphs for the GS, LR and CS-SRR HCP data sets, respectively. The rectangle in (b) shows the partial-volume effects in the LR image where the orientations of the glyphs are not estimated correctly. (d), (e) and (f) are the fiber tracts, color-coded with tract orientation for the three data sets with seeds in sub-thalamic nucleus. The arrow points out some missing fiber tracts in the LR image.

4.2 True CS-SRR Scenario

The second experiment was based on a data set acquired on a 3T Siemens clinical scanner. We acquired three overlapping thick-slice scans with spatial resolution \(1.2\times 1.2 \times 3.6~\,mm^3\). The DWI volumes were sub-pixel shifted by \(1.2~\,mm\) along the slice-selection direction. Each LR DWI had a set of 30 unique gradient directions with \(b=1000\). For comparison, we also acquired 9 acquisitions (with 90 gradient directions each) of the same subject with a spatial resolution of \(1.2 \times 1.2 \times 1.2~\,mm^3\), which was used as the “gold standard” data. Due to time limitations, these high resolution scans had partial brain coverage (it took about 1.5 hours to obtain these 9 scans). The average of these 9 scans (after motion correction) was considered as the gold standard. We also acquired a high resolution \(B_0\) image, i.e. the \(b=0\) image with no diffusion encoding, and a \(T_1\)-weighted image to obtain tissue classifications for prior-information used in our algorithm. To ensure that the LR DWI’s were in the same spatial co-ordinate system, we first down-sampled the whole-brain \(B_0\) image to produce three thick-slice volumes which were considered as the reference images. These images were only used for spatial normalization and not in the actual reconstruction algorithm. Then, the three acquired thick-slice LR DWI scans were registered to the corresponding reference volumes. The \(T_1\) image was registered to the whole-brain \(B_0\) image using a nonlinear transformation. From the registered \(T_1\) image, the tissue brain segmentation was used for adjusting the parameters of the algorithm in different tissue types. We set the FWHM of the blurring kernel to \(1.2~\,mm\), \(\lambda _1=0.005\), \(\lambda _2=0.05\), and \(w=0.01\) in CSF area and \(w=1\) in gray and white matters. These parameters were learned using exhaustive search experiments based on one slice of the image so that the reconstruction error was small compared with the gold-standard. For computing quantitative metrics, we registered the reconstructed whole-brain data to the partial-brain GS data set.

We first obtained the tractography results from the reconstructed (CS-SRR) and GS partial-brain data sets using the method in [17]. Since whole-brain Freesurfer cortical parcellation was not available, we could not employ WMQL for tract extraction. Consequently, we use manually selected ROI’s to extract fiber bundles in the corpus callosum (CC) and the corticospinal tract (CST), respectively. The extracted fiber bundles are shown in Fig. 4a and b, respectively. We note that there is minor difference between FA and Trace obtained from both the data sets. The value of the Bhattacharyya coefficient also indicates a significant overlap between the fiber bundles.

Fig. 4.
figure 4

(a), (b) are the tractography results with manually selected seeds in corpus callosum (CC) and the corticospinal tract (CST) (with red and yellow tracts obtained using the GS and CS-SRR data sets respectively). The comparison metrics for the tractography results are shown in the following table.

 

GS FA

CS-SRR FA

GS Trace

CS-SRR Trace

Fiber-bundle overlap

CC

0.64

0.65

\(2.2\times 10^{-3}\)

\(2.4\times 10^{-3}\)

0.96

CST

0.62

0.64

\(2.0\times 10^{-3}\)

\(2.2\times 10^{-3}\)

0.94

To demonstrate the difference between HR and LR images, we also generated a LR image by averaging every \(2\times 2\times 2\) neighboring voxel in the GS data set to obtain a LR dMRI data set with isotropic voxel size of \(2.4^3 \; mm^3\). Figure 5 shows the single tensor glyphs color coded by direction for the GS, LR and the CS-SRR data sets in the same brain region from a coronal slice. The background of the glyphs is the corresponding \(T_1\) weighted image. As pointed out in the rectangular region in Fig. 5b, due to partial-volume effects, the LR image was not able to capture the correct fiber orientations near the gray matter areas. Figure 5c is similar to Fig. 5a, indicating that the proposed method was able to correctly reconstruct the fiber orientations near gray-matter areas.

Fig. 5.
figure 5

(a), (b) and (c) show the single tensor glyphs colored coded by fiber orientation for the GS, LR and CS-SRR data sets, respectively. The rectangle in (b) points out the partial-volume effects in the LR image where the directions of the glyphs are not estimated correctly (shown by different colors from the GS glyphs).

5 Conclusion

We introduced a novel method for reconstructing very high-resolution diffusion data on a standard clinical scanner. By combining the concepts of compressed sensing and super-resolution, we were able to with significantly reduce scan time while increasing the spatial resolution. Preliminary results show that the our method is capable of accurately recovering complex fiber orientations in white and gray matter regions at a high spatial resolution, similar to a physically acquired gold-standard data. Future work involves doing extensive validation on several subjects in different brain regions.