Abstract
Several chronic lung diseases, like idiopathic pulmonary fibrosis (IPF) are characterised by abnormal dilatation of the airways. Quantification of airway features on computed tomography (CT) can help characterise disease severity and progression. Physics based airway measurement algorithms that have been developed have met with limited success, in part due to the sheer diversity of airway morphology seen in clinical practice. Supervised learning methods are not feasible due to the high cost of obtaining precise airway annotations. We propose synthesising airways by style transfer using perceptual losses to train our model: Airway Transfer Network (ATN). We compare our ATN model with a state-of-the-art GAN-based network (simGAN) using a) qualitative assessment; b) assessment of the ability of ATN and simGAN based CT airway metrics to predict mortality in a population of 113 patients with IPF. ATN was shown to be quicker and easier to train than simGAN. ATN-based airway measurements showed consistently stronger associations with mortality than simGAN-derived airway metrics on IPF CTs. Airway synthesis by a transformation network that refines synthetic data using perceptual losses is a realistic alternative to GAN-based methods for clinical CT analyses of idiopathic pulmonary fibrosis. Our source code can be found at https://github.com/ashkanpakzad/ATN that is compatible with the existing open-source airway analysis framework, AirQuant.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
- Generative model evaluation
- Style transfer
- Computed tomography
- Airway measurement
- Bronchiectasis
- Idiopathic pulmonary fibrosis
1 Introduction
Chronic lung disease is one of the leading causes of morbidity and mortality across the world. As smoking rates in the developing world increase, the prevalence of chronic lung disease is set to rise. Interstitial lung diseases (ILD) are characterised by inflammation and scarring of the lung and the incidence of ILD continues to increase [25].
A subset of ILDs are characterised by lung fibrosis, with idiopathic pulmonary fibrosis (IPF) having the worst prognosis of all the fibrosing ILDs [4]. In IPF the airways are pulled open by fibrotic contraction of the surrounding connective tissue. Computed tomography (CT) imaging is used to visualise airway structure. In IPF the presence of dilated airways in the lung periphery on CT, termed traction bronchiectasis, is a disease hallmark.
When assessing disease severity in IPF, physiologic measurements are typically used. However these are associated with a degree of measurement variability. It has been postulated that combining imaging measures of airway abnormality with lung function measurements could help improve estimation of disease severity in IPF [18]. Importantly, better measures of disease severity would benefit cohort enrichment of subjects into therapeutic trials.
Lung damage in IPF progresses from the distal lung towards the centre of the lung [15]. As a result, the earliest signs of lung damage are seen in the smaller airways. Yet these airways are typically the most challenging to quantify. Airway measurement is complicated by partial volume effects that result in smaller airways having a blurred contour to their walls. Measurement challenges are compounded by variations in CT image acquisition including different reconstruction kernels, scan parameters and scanner models as well as the underlying pathology affecting the lung.
Physics based airway measurement algorithms tend to perform sub optimally when measuring the lumens of small airways [3, 12]. Identifying airway walls can also be challenging. Airway paths often run in tandem with those of the pulmonary artery. Consequently, in regions when the pulmonary artery abuts the airway wall, identification of the contour of the outer airway wall is compromised.
1.1 Related Work
Deep learning frameworks have been applied to the measurement of airways in the lung in a bid to improve measurement accuracy. However, these machine learning methods are extremely data hungry and can be biased towards the training data sample [10]. Synthetic data by way of generative models has been employed to improve the training of deep learning models. This helps overcome the data limitations that are ubiquitous to medical imaging studies [24].
A state of the art method in measuring airway lumen radius and wall thickness on CT imaging, simGAN [16, 21], takes labelled simplistic representations of airway patches (synthetic images) and aims to transforms them in to the emulations of real airways by generative adversarial training (GAN) [6]. These are then used for supervised training of a convolutional neural regressor (CNR) which learns to measure airway radius and wall thickness and ultimately to run inference on real CT images.
The driving loss for realism in simGAN is cross-entropy loss computed on the classifications of the discriminator. For successful synthetic refinement by image transformation, the synthetic and refined images must have good correspondence in their shared label. To this end, a per-pixel \(\Vert l\Vert _{1}\) regularisation loss is applied between input and output of the refiner.
GAN training is inherently unstable with mode collapse complicating and lengthening training times. As an alternative strategy, in this paper we propose the first use of perceptual losses to generate labelled synthetic airway images. Perceptual loss functions have been applied to image style transfer and super-resolution tasks [11]. We explore the clinical benefits of learning from perceptual loss generated synthetic data in mortality prediction.
2 Methods
In the first part of our study we generate synthetic airway patches that demonstrate realistic airway characteristics. In tandem, we segment the airways on clinical CT scans of a cohort of IPF patients. We train our Airway Transfer Network (ATN) to transform our synthetic images to refined images across our synthetic and real datasets by optimising for perceptual losses. We then compare the results of ATN with simGAN. A CNR is trained on the resultant refined datasets for the purpose of inference on real CT airways. We compare the two refiner models qualitatively. We compare ATN and simGAN against the full width at half maximum edgecued segmentation limited (FWHMesl) technique as implemented in [20], originally by [12]. The FWHMesl technique is widely used in the literature as the reference for comparison of previous airway measurement methods [7, 16, 26]. In our clinical comparison, we examine which of the three methods of airway measurement provides the best and most consistent association with mortality on CT scans of patients with IPF.
Airway segmentation was performed using a 2D dilated U-Net [27] trained on CT scans in 25 IPF and healthy individuals [17]. We extract orthogonal airway patches for all segmented airways. We parameterise airway labels as two ellipses that share centre and rotation, resulting in 7 parameters for each patch: inner airway wall major and minor axis radii \(R_{A}\) and \(R_{B}\); outer airway wall major and minor axis radii \(W_{A}\) and \(W_{B}\); centre coordinates \(C_{x}\) and \(C_{y}\); and rotation \(\theta \). Due to the phase in \(\theta \), for the purposes of CNR training the rotation angle is converted into a double angle representation [13].
Once the refiner model has been trained, its output is used to train a CNR by supervised learning to regress to target airway labels. The inner and outer airway wall measures are then derived. All deep learning methods were implemented in pytorch [19] and CT image processing was done using the open source airway analysis framework known as AirQuant [17]. We release our code open sourceFootnote 1.
2.1 Airway Synthesis
Details of airway parameters and synthesis pipeline have been previously described [16]. Airway characteristics are sampled from a set of distribution parameters informed by [23]. We deviate from these parameters in two ways. First, we use an airway lumen radius (LR) interval of [0.3, 6] to permit measurement of smaller airways. Second, we use an airway wall thickness [\(0.1 \cdot LR + 0.2\), \(0.3\cdot LR+0.8\)] mm to reflect the lack of airway wall thickening in IPF. We add four further parameters: (i) parameters for the airway centre determined by a normal distribution \(X\sim N(0,1)\) mm to account for airway skeletons that are not perfectly positioned within the centre of the airway lumen. (ii) \(p=0.4\) that an adjacent airway of similar diameter is randomly added. This is performed to accommodate airway patches close to airway bifurcations and to train the CNR to correctly identify the airway in the centre of the patch. (iii) We model our airways as ellipsoids, we achieve this by an ellipsoidness characteristic, sampled from a uniform distribution, \(X\sim U(0.9,1)\) which determines the ratio in major and minor radii of the ellipse. (iv) Uniformly random rotation applied to the airway in the horizontal axis. We include our synthetic airway generator and configuration parameters in the open-source code repository.
2.2 Perceptual Losses
We implement perceptual losses for computing high level perceptual differences between synthetic and real images as described by [11]. These losses are computed by comparing the activations in particular layers, j of a pretrained convolutional neural network (CNN), \(\phi \) between a pair of images. Different activation layers of a trained CNN learn to represent different image features on the same sampled patch. In minimising for perceptual losses we are looking to reduce the differences in the activation of these layers between the refiner output and some objective image. For each calculation of perceptual losses on a synthetic input image, x we have a refiner prediction, \(\hat{y}\). As a modification of the original style transfer implementation [11], a randomly chosen real image is selected as the style target, \(y_{s}\). Perceptual losses are then calculated and summed for different layers \(\phi _{j}\).
We utilise feature reconstruction loss. This is defined as the mean euclidean distance between activations of the input and output images of the refiner, where C, H, and W are the number of channels, height and width of layer j respectively. We use a VGG-16 [22] network pretrained on the ImageNet dataset [2] in our calculations of style and feature losses.
We also employ style reconstruction loss, which considers those features that tend to be activated together between the refiner output and the given style target image, a random real airway, where \(G^{\phi }_{j}\) is the gram matrix for a given layer j of \(\phi \) as described in [5].
2.3 Clinical Data
We examined CT images from 113 IPF patients diagnosed at the University Hospitals Leuven, Belgium. CTs were evaluated by an experienced chest radiologist (author JJ) for quality i.e. absence of breathing artefacts and infection. The quality of the automated segmentation was also visually inspected to ensure contiguous airway segmentations without oversegmentation blowouts. Airway segmentations were also required to reach the sixth airway generation in the upper and lower lobes to be selected for analysis. Pulmonary function tests were considered if they occurred within 90 days of the CT scan: Forced Vital Capacity (FVC, n = 111)); diffusing capacity of the lung for carbon monoxide (DLco, n=103).
The trachea and first generation bronchi were excluded from analysis. We define an airway segment as the length of airway that runs between airway branching points or an airway endpoint. All airway segments were pruned by 1 mm at either end to avoid bifurcating patches. \(80\times 80\) pixel size orthogonal airway patches were linearly interpolated with a pixel size of \(0.5\times 0.5\) mm from the CT at 0.5 mm intervals along each segment. This resulted in a final set of 546,790 real CT-derived airway patches. A synthetic dataset of 375,000 patches was generated to train our refiner.
27% of patients were female. 74% of patients had smoked previously. The median patient age was 71, with 57% of patients having died. All patients had received antifibrotic drug treatment.
Measures of intertapering, intratapering [14] and absolute airway volume were derived from the airway measurements for each airway segment. Segmental intertapering represents the relative difference in diameter of an airway segment when compared to its parent segment. Segmental intertapering is calculated as the difference in mean diameter, \(\bar{d}\) of an airway segment and its parent segment, \(\bar{d_{p}}\), divided by the mean diameter of the parent segment. Segmental intratapering is the gradient of change in diameter of the airway segment relative to the diameter of the origin of the segmentFootnote 2. Segmental intratapering is computed by dividing the gradient, m by the zero-intercept, c of a line \(y=mx+c\) fitted to the diameter measurements of an airway segment. Segmental volume is computed by summing area measurements along an airway segment, and multiplying this value by the measurement interval, i.e. an integration of area along the segment’s length.
Univariable and multivariable Cox proportional hazards models were used to examine patient survival. Multivariable models included patient age (years), gender, smoking status (never vs ever) and either FVC or DLco (as measures of disease severity) as covariates. The goodness of fit of the model was denoted by the concordance index [8]. A p-value of <0.05 was considered statistically significant.
2.4 Implementation Details
We use the same refiner architecture as in [16, 21], the refiner is a purely convolutional network with four repeating 3 \(\times \) 3, 64 feature ResNet blocks [9]. The measurement CNR, described in [16], is a convolutional network that feeds into two fully connected layers to learn the airway ellipse parameters. Instead of the custom CNR loss described in [16], we implemented a mean square error (MSE) loss for regressing to the airway ellipse parameters.
Synthetic images were generated to \(0.5\times 0.5\) mm pixel size making \(80\times 80\) pixel patches, corresponding to the real patch generation noted in Sect. 2.3. All images were standardised and augmented on the fly, adding random Gaussian noise [25, 25] Hounsfield units, random levels of Gaussian blurring with standard deviation scalled in the interval [0.5, 0.875] and random flipping (\(p=0.2\)). We apply random scaling on real images only, in the interval [0.75, 1.25] to increase diversity in airway size. Finally, a centre crop was applied to make a \(32\times 32\) pixel input patch.
Both simGAN and ATN models were trained for 10000 steps, where the simGAN refiner had 50 training iterations and the discriminator 1 iteration for every 1 step. The simGAN discriminator was implemented as described in the original method, with a memory buffer and local patch discrimination [21]. We used Weights & Biases for experiment tracking [1].
Figure demonstrates the overall method employed here as well as the ATN and CNR architecture.
3 Results
We implemented all training on an NVIDIA GeForce RTX 2070 graphical processing unit with a batch size of 256, learning rate of 0.001, \(\Vert l\Vert _{1}\) regularisation factor in range of [0.0001, 0.1]. simGAN and ATN took 14 and 0.6 h respectively to converge during training. We qualitatively found that both simGAN and ATN produced refined images of optimal quality with a \(\Vert l\Vert _{1}\) regularisation factor of 0.01.
Style-transfer from paintings to natural images show that larger-scale structure is transferred from the target image when training on losses of higher layers [11]. In order to maintain label correspondence between refiner input and output, we similarly only use the feature loss using the relu3_3 activation layer. Style loss is computed from the two lower relu1_2, relu2_2 activation layers onlyFootnote 3. Figure 2 demonstrates qualitative results of our airway refinement method.
The CNR was trained with batch size in the interval [256, 2000] and learning rate of 0.001. Batch size of 2000 was chosen for its speed, and converged at around 40 epochs within one hour. The CNR achieves comparable results on ATN and simGAN refined images.
Figure 3 demonstrates qualitative results of our ATN method on real CT data. Table 1 shows results of the Cox regression survival analyses. The CNR when regressing to an airway feature demonstrated a strong association with mortality. This was despite the CNR label not perfectly aligning to the exact airway boundary.
4 Conclusion
We present a learning based airway measurement method trained on a transformation network that refines synthetic data using perceptual losses. Our model ATN was compared with a state-of-the-art model simGAN [16] and a physics based method FWHMesl. When assessing the clinical utility of ATN, we found that it was the strongest predictor of survival across all three airway biomarkers. We found that our method trains faster and with minimal complications, unlike a GAN framework. We expect future work to consider the generalisation of such a method, for example examining airways in patients with different diseases, images acquired on different scanner parameters and potentially on higher scale imaging such as micro-CT studies of the lungs.
Notes
- 1.
- 2.
Segments are considered to be oriented from the centre of the lung to the periphery. Accordingly, measurement of the airway origin beings at the end closest to the trachea.
- 3.
higher activation layers are considered in the supplementary material.
References
Biewald, L.: Experiment tracking with weights and biases (2020). https://www.wandb.com/, software available from wandb.com
Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255, June 2009. https://doi.org/10.1109/CVPR.2009.5206848
Estépar, R.S.J., Washko, G.G., Silverman, E.K., Reilly, J.J., Kikinis, R., Westin, C.-F.: Accurate airway wall estimation using phase congruency. In: Larsen, R., Nielsen, M., Sporring, J. (eds.) MICCAI 2006. LNCS, vol. 4191, pp. 125–134. Springer, Heidelberg (2006). https://doi.org/10.1007/11866763_16
Flaherty, K.R., et al.: Idiopathic pulmonary fibrosis. Am. J. Respir. Crit. Care Med. 174(7), 803–809 (2006). https://doi.org/10.1164/rccm.200604-488OC
Gatys, L.A., Ecker, A.S., Bethge, M.: A neural algorithm of artistic style (2015). https://doi.org/10.48550/ARXIV.1508.06576, https://arxiv.org/abs/1508.06576
Goodfellow, I.J., et al.: Generative adversarial networks. arXiv:1406.2661 [cs, stat] (June 2014). http://arxiv.org/abs/1406.2661
Gu, S., et al.: Computerized identification of airway wall in CT examinations using a 3D active surface evolution approach. Med. Image Anal. 17(3), 283–296 (2013). https://doi.org/10.1016/j.media.2012.11.003
Harrell Jr., F.E., Lee, K.L., Mark, D.B.: Multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Stat. Med. 15(4), 361–387 (1996). https://doi.org/10.1002/(SICI)1097-0258(19960229)15:4<361::AID-SIM168>3.0.CO;2-4
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition, December 2015. https://doi.org/10.48550/arXiv.1512.03385
Hofmanninger, J., Prayer, F., Pan, J., Röhrich, S., Prosch, H., Langs, G.: Automatic lung segmentation in routine imaging is primarily a data diversity problem, not a methodology problem. Eur. Radiol. Exp. 4(1), 1–13 (2020). https://doi.org/10.1186/s41747-020-00173-2
Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution, March 2016. https://doi.org/10.48550/arXiv.1603.08155
Kiraly, A.P., Reinhardt, J.M., Hoffman, E.A., McLennan, G., Higgins, W.E.: Virtual bronchoscopy for quantitative airway analysis. In: Amini, A.A., Manduca, A. (eds.) Medical Imaging 2005: Physiology, Function, and Structure from Medical Images, vol. 5746, p. 369. International Society for Optics and Photonics, April 2005. https://doi.org/10.1117/12.595283
Kluvanec, D., Phillips, T.B., McCaffrey, K.J.W., Moubayed, N.A.: Using orientation to distinguish overlapping chromosomes, March 2022. https://doi.org/10.48550/arXiv.2203.13004
Kuo, W., Perez-Rovira, A., Tiddens, H., de Bruijne, M.: Airway tapering: an objective image biomarker for bronchiectasis. Eur. Radiol. 30(5), 2703–2711 (2019). https://doi.org/10.1007/s00330-019-06606-w
Lederer, D.J., Martinez, F.J.: Idiopathic pulmonary fibrosis. N. Engl. J. Med. 378(19), 1811–1823 (2018). https://doi.org/10.1056/NEJMra1705751
Nardelli, P., Ross, J.C., San José Estépar, R.: Generative-based airway and vessel morphology quantification on chest CT images. Med. Image Anal. 63, 101691 (2020). https://doi.org/10.1016/j.media.2020.101691
Pakzad, A., et al.: Evaluation of automated airway morphological quantification for assessing fibrosing lung disease. Technical report, November 2021. arXiv:2111.10443, arXiv. https://doi.org/10.48550/ARXIV.2111.10443
Pakzad, A., Jacob, J.: Radiology of bronchiectasis. Clin. Chest Med. 43(1), 47–60 (2022). https://doi.org/10.1016/j.ccm.2021.11.004
Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. In: Wallach, H., Larochelle, H., Beygelzimer, A., d’ Alché-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 32, pp. 8024–8035. Curran Associates, Inc. (2019). http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
Quan, K., et al.: Tapering analysis of airways with bronchiectasis. In: Angelini, E.D., Landman, B.A. (eds.) Medical Imaging 2018: Image Processing, vol. 10574, p. 87. SPIE, March 2018. https://doi.org/10.1117/12.2292306
Shrivastava, A., Pfister, T., Tuzel, O., Susskind, J., Wang, W., Webb, R.: Learning from simulated and unsupervised images through adversarial training (2017). https://doi.org/10.48550/ARXIV.1612.07828, https://arxiv.org/abs/1612.07828
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 [cs] (April 2015). https://doi.org/10.48550/ARXIV.1409.1556
Weibel, E.R.: Morphometry of the Human Lung. Springer, Berlin, Heidelberg (1963). https://doi.org/10.1007/978-3-642-87553-3_6
Willemink, M.J., et al.: Preparing medical imaging data for machine learning. Radiology 295(1), 4–15 (2020). https://doi.org/10.1148/radiol.2020192224
Xie, M., Liu, X., Cao, X., Guo, M., Li, X.: Trends in prevalence and incidence of chronic respiratory diseases from 1990 to 2017. Respir. Res. 21(1), 1–13 (2020). https://doi.org/10.1186/s12931-020-1291-8
Xu, Z., Bagci, U., Foster, B., Mansoor, A., Udupa, J.K., Mollura, D.J.: A hybrid method for airway segmentation and automated measurement of bronchial wall thickness on CT. Med. Image Anal. 24(1), 1–17 (2015). https://doi.org/10.1016/j.media.2015.05.003
Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions. arXiv:1511.07122 [cs] (April 2016). https://doi.org/10.48550/ARXIV.1511.07122
Acknowledgements
This research was funded in whole or in part by the Wellcome Trust [209553/Z/17/Z]. For the purpose of open access, the author has applied a CC-BY public copyright licence to any author accepted manuscript version arising from this submission. AP is funded jointly by the Cystic Fibrosis Trust and EPSRC i4health, centre for doctoral training studentship. JJ was supported by a Wellcome Trust Clinical Research Career Development Fellowship and the NIHR UCLH Biomedical Research Centre, UK.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Pakzad, A. et al. (2022). Airway Measurement by Refinement of Synthetic Images Improves Mortality Prediction in Idiopathic Pulmonary Fibrosis. In: Mukhopadhyay, A., Oksuz, I., Engelhardt, S., Zhu, D., Yuan, Y. (eds) Deep Generative Models. DGM4MICCAI 2022. Lecture Notes in Computer Science, vol 13609. Springer, Cham. https://doi.org/10.1007/978-3-031-18576-2_11
Download citation
DOI: https://doi.org/10.1007/978-3-031-18576-2_11
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-18575-5
Online ISBN: 978-3-031-18576-2
eBook Packages: Computer ScienceComputer Science (R0)