Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

Patient head movement during high resolution brain positron emission tomography (PET) scanning will cause blurring and ghosting [3]. The low count rate and resulting low contrast makes it almost impossible to perform motion correction on raw PET data, and therefore most methods rely on external tracking of the head movement [1012, 16]. The Polaris Vicra (Northern Digital Inc.) tracking system has been used as the reference on many PET installations [6]. While the Polaris system is well tested and accurate, it suffers from problems related to attaching optical markers to the patient’s head. Experience shows that in a clinical setting, the markers are difficult to attach such that they stay in position during the entire scan. A markerless system that fits into the narrow PET tunnel will improve the clinical acceptance and the diagnostic value of PET brain scans.

We have previously described a structured light based system that is based on a small projector and two small cameras [7] for tracking patient head pose. This system has been modified as described below and it is referred to as Tracoline. The Tracoline system has been designed to fit into the patient tunnel of the Siemens High Resolution Research Tomograph (HRRT) PET brain scanner. The HRRT PET scanner has a spatial resolution down to 1.4 mm [8] and is therefore well suited for testing new motion correction methods. The Tracoline system is based on the progressive reconstruction of 3D surfaces of the upper face region of the patient in the scanner. The pose changes are found by computing the rigid transformation between the current scan and the initial surface scan. The system described in [7] was based on visible light and did not operate in real time. Visible light scanners are not suited for repeated human facial scans. Furthermore, to be functional the system needs to acquire frames sufficiently fast to faithfully capture patient head movements. In this paper, we describe a system using invisible light with a camera acquisition rate of 30 frames per second.

While the previous paper focused on measuring the accuracy of the structured light tracking system using a rotation stage as ground truth [7], the real interest is the improvement of the PET scans. To be able to evaluate the quality improvement of the PET scan, a scan using a radioactive tracer must be performed. A common approach is to use a phantom and compare the resulting PET scan with the known geometry of the phantom [8]. We have therefore designed a customized phantom with a radioactive source and used this in the evaluation of the Tracoline based motion correction. Our system setup with the Tracoline system and the custom phantom can be seen in Fig. 1. Compared to other external tracking systems, where the geometric alignment between the tracking system and the PET scanner can be problematic, we investigate a novel alignment approach based on aligning the Tracoline system scan directly to the PET transmission scan.

Fig. 1
figure 1

Left: The patient tunnel of the HRRT PET scanner with the Tracoline system mounted. The phantom is mounted on a rotating stage rotated to the right (−10 degrees). Right: The phantom including one of the two radioactive line sources. It is placed in the head in the same angle as shown here to simulate the brain

2 Experiments and Methods

The Tracoline system consists of two Point Grey Flea2 cameras (1288×964 pixels), each running at 30 frames per second. The Pico Digital Light Processing (DLP) projector from Texas Instruments is used to project phase-modulated patterns. One of the light diodes of the Pico projector has been replaced with a near infrared (NIR) diode resulting in a non-visible projected pattern. The projector is controlled by a GFM Pico developer kit board that also sends trigger pulses to the cameras, in order to synchronize the projected patterns and the shutter of the cameras. A multi-threaded C++ program running on a standard portable computer acquires the real time camera data and stores them as image files. The 3D point cloud generation, surface reconstruction, and alignment are done in a post-processing step. The Tracoline system and HRRT PET acquisition computer are synchronized through an internal network time protocol (NTP) server.

The 3D point cloud generation is based on phase-shifting interferometry (PSI) [5] where a set of 2D interferograms are projected and projector-camera correspondence can be found using phase unwrapping. This is explained in detail in [5, 7]. While three patterns are used in [7], the system is now extended to use six patterns with varying wavelengths to make the phase unwrapping more robust to discontinuities in the surface. Since each point cloud computation requires six frames, the effective tracking frequency is 5 Hz. A surface is reconstructed using a modern algorithm [9] based on the point cloud acquired in the initial position. The facial pose changes are then found by rigidly aligning the following surface scans to this reference surface using an optimized, iterative, closest point (ICP) algorithm [15].

To correct for motion, we need to know the transformation between the HRRT PET scanner coordinate system and the Tracoline system. To estimate this transformation, we use the transmission scan of the HRRT PET scanner, which is also used for the attenuation and scatter correction within normal PET reconstruction. The transmission scan is a voxel volume similar to a computed tomography (CT) scan. The initial reference surface scan is captured by the Tracoline system during the transmission scan, thus creating correspondence. The transformation is computed using a pseudo-ICP algorithm [13]. The surface scan is scaled to fit the volume, and manually rotated and translated into an initial position. To find correspondence between surface sample points and the volume, the volume is sampled in the normal direction (both positive and negative) of the surface scan to find the point with maximum gradient. Knowing the general orientation of the patient in the PET scanner, we use the absolute gradient in the x (left/right) and z (axial) direction and the negative gradient for the y (anterior/posterior) direction:

$$\Delta f = \biggl \vert \frac{\partial f}{\partial x} \biggr \vert - \frac{\partial f}{\partial y} + \biggl \vert \frac{\partial f}{\partial z} \biggr \vert $$

With a point correspondence, a rigid transformation is found using the closed-form loop to estimate the absolute transformation [4]. With an initialization, transformation, this process is iterated until the transformation of the Tracoline scan converges to the volume data.

In order to apply the motion correction to the PET data, we apply the multiple acquisition frames (MAF) method [10]. In [8] the MAF method was demonstrated on the Siemens HRRT PET scanner using the tracking input from the Polaris Vicra system. We divide the PET emission list mode data into equal time length intervals and, for each interval, a PET frame is reconstructed using the 3D ordered subset expectation maximization (3D-OSEM) algorithm with resolution modeling and incorporating a spatially invariant point spread function [14]. These frames are then repositioned to a reference position using the Tracoline tracking system.

A custom phantom with known geometry was designed. It consists of a hollow plastic mannequin head with a very low attenuation coefficient. Two radioactive line sources are inserted into the head to provide activity for the HRRT PET scanner. The activities of the line sources are 2×7.8 MBq each, created by a positron-emitting germanium-68/gallium-68 generator. As can be seen in Fig. 1, the line sources go through the head from the back of the skull to the forehead. The phantom was mounted onto a rotation stage by Thorlabs and placed in the patient tunnel of the HRRT PET scanner.

Two experiments where performed using the stage to rotate the head. In experiment one, the head was rotated from −20 to 20 degrees in steps of 5 degrees. At each position a 30 s frame was PET reconstructed and repositioned. Data with motion was excluded from the reconstruction. In experiment two, the head was rotated from −10 to 10 degrees in a continuous motion with a maximum speed of one degree per second. The PET data was reconstructed using one second frames in experiment two. Experiment one simulates the clinical situation where the patient is performing a rapid head motion followed by a stationary period. State of the art practice is to discard PET data during such rapid motions. The second experiment simulates e.g. a patient falling asleep, where the head is slowly drifting from side to side.

We evaluate the effect of the motion correction on the reconstructed PET images by calculating Dice’s coefficient (percent volume overlap) [1] between a reference image recorded without phantom motion, the motion distorted image, and the Tracoline based motion-corrected image. The number of voxels, N, included in the calculation is set to a value corresponding to the number of voxels inside the tubes 2.5 times the diameter of the PET sources used (outer diameter 3.2 mm and active length 168 mm) [8]. The extended volume is used in order to compensate for partial volume effects. In each image to be studied, the set of the N most intense voxels is extracted and used for the Dice’s coefficient computation, presented as the percentage of overlapping voxels. In addition we compute the normalized cross correlation between the reference image and each image frame, either motion-corrected or uncorrected [2].

3 Results and Discussion

The rigid transformation between the coordinate system of the Tracoline system and the PET image frame is obtained from using the described surface-to-volume alignment. Figure 2(a) shows the reference surface aligned to maximum gradient points in the transmission scan.

Fig. 2
figure 2

(a) shows the alignment between maximum gradient points in the transmission scan shown as red dots and the Tracoline face scan shown as a blue surface. (b) shows quantitative results of the stepwise experiment. Top: the percentage of overlapping points in the reference PET image compared to the unaligned/aligned PET images based on either the right or the left camera. Bottom: the cross correlation between the reference PET image and the unaligned/aligned PET images

Figure 2(b) shows the results of the first experiment with stepwise rotation of the phantom. The top figure shows the percentage overlap between PET frames of the line sources in the reference position and a scan position as a function of the performed rotation of the head. Similarly, the bottom plot of Fig. 2(b) shows the correlation coefficient between the reference image and a motion-corrected/uncorrected image for the different scan positions. Results based on tracking information from the left and right camera of the Tracoline system are shown in green and blue colors respectively, while the red curve represents the uncorrected image results. The overlap and the correlation measures are in agreement. The results of the uncorrected frames decrease with the size of performed rotation from an overlap of 100% down to 2% at ±20 degrees. The overlap of the motion-corrected reconstruction is improved significantly for all positions with percentage overlap of 71–93%. The overlap is not 100%, which is mainly due to the internal calibration of the Tracoline system, the ICP alignment, and the geometrical alignment with the HRRT PET scanner. In addition, the interpolation error, combined with the straight and narrow line sources of the phantom (with a diameter similar to the voxel size of 1.2 mm), induces partial volume effects and thereby decreases the overlap and correlation measure. The differences between the left camera and the right camera could be explained by the construction of the reference surface scan, where left camera was chosen as the basis. The result is similar to [8], where the overlap was 65–85% for a 10 degrees corrected rotation. However, the two studies cannot be directly compared since the phantom designs are different.

A visual evaluation of the motion correction is shown in Fig. 3 for the maximum rotation of 20 degrees. The PET images are summed along one dimension and visualized on top of each other pairwise in the red and green color channels. The overlapping pixels of the two PET images appear yellow. The top row of the figure shows the reference image and the uncorrected image as two sets of rods rotated approximately 2 cm at the end points. These correspond to the relevant brain regions: the frontal lobe and cerebellum. The bottom row shows the reference image and the corrected image seen as two yellow rods, demonstrating a near-perfect motion correction. This position has an overlap of 89% in contrast to the rotation of −20 degrees with an overlap of 71%.

Fig. 3
figure 3

The figure shows the summation of the PET images along 3 different axes for a reference image (shown in green) and a target image (shown in red) with a 20 degree rotation (overlap = 89%, shown in yellow). The uncorrected image is shown as captured in the first row, while the second row shows the image after motion correction

The results of the second experiment, with a continuous rotation of 20 degrees demonstrating the realtime pose registration of the Tracoline system, are presented in Fig. 4. The one second PET frames are summed and fused with the transmission image of the phantom. The top of the figure shows a row of uncorrected images, where the motion of the line sources is seen as blurred circle parts. The bottom row shows the motion-corrected image, where the previously blurred parts appear in focus and with high intensity. The cross section of the line sources shows dots with a diameter of only a few pixels. Long drift motion is a very complex problem to overcome using image registration methods for motion estimation, and this is why an external tracking system is of great value. Our latest results show that continuous motion can be tracked in real time and PET frames successfully corrected.

Fig. 4
figure 4

Results of the dynamic PET scan. One-hundred, one-second frames uncorrected (top) and MAF motion corrected (bottom) are summed and fused with a transmission scan. The frame repositioning is based on the left camera alignment

4 Summary and Conclusions

This paper describes a complete system for motion correction in high resolution PET brain imaging. It is based on a small and flexible structured light scanner mounted above the patient tunnel of the PET scanner. The scanner is equipped with a near infrared light source, making it suitable for future patient examinations. Furthermore, the system tracks the head pose changes with a frequency of 5 Hz, which is suitable for the head movement experienced during real clinical PET scanning. In order to align the two systems, a novel algorithm using the HRRT PET transmission scan and the initial surface scan was presented. The performance of the system was evaluated using a custom-designed phantom with two radioactive line sources mounted on a programmable rotation stage. The results of the two experiments are very promising. The first experiment simulates rapid but short head movements and the second experiment simulates slow but longer head movements. Quantitative analysis shows that the combined system is able to robustly reduce motion artifacts and greatly improve PET scans for scenarios involving both slow and rapid movements. The system is nearly ready for actual clinical testing.