Abstract
Lung cancer is one of the fatal cancer types, and its identification at early stages might increase the patients’ survival rate up to 60–70%. Predicting patients’ lung cancer survivability has become a trendy research topic by scholars from medicine and computer science domains. In this article, a novel approach to the deep convolutional neural network (DCNN) is proposed for the precise and automatic classification of lung cancer. Specifically, the auto weight dilated convolutional unit utilized multi-scale convolutional feature maps to acquire lung cancer features at different scales and employed a learnable set of parameters to fuse convolutional feature maps in encoding layers. The AD unit is an effective architecture for feature extraction in the encoding stage. We used the advantages of the U-Net network for deep and shallow features, combined with AD units to multimodal image classification. In this model, the four channel model inputs correspond to the CT images of four modes, respectively. The main body of the network is composed of auto-weight dilated (AD) unit, Residual (Res) unit, linear upsampling, and the first and last convolution units.. The network that applied Block-R3 had higher segmentation performance than the networks of Block-R1 and Block-R2 experimental results indicated that DCN had outperformed all the six classical machine learning algorithms in predicting the survival period of lung cancer patients with an accuracy of 88.58%. The results are believed to support healthcare professionals to manage costs and provide treatment at the appropriate time.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Lung cancer patients’ relative five-year survival rate is around 20%, making it the deathliest cancer type. A man’s probability of lung cancer is 1/15, while it is 1/17 for a woman. This ratio is increased for smokers [1]. The major issue with lung cancer is the difficulty of its treatment because it is identified at later stages [2]. Identifying lung cancer at early stages might boost the patients’ survival rate up to 60–70% [3]. Predicting the estimated survival period after identification improves the prognostic accuracy, which leads to physicians’ and patients’ families’ better decision-making [4]. Therefore, predicting patients’ lung cancer survivability has become a trendy research topic by scholars from medicine and computer science domains. The advent of artificial intelligence (AI) techniques, specifically machine learning algorithms, has improved the diagnosis and treatment of cancer patients.
Machine learning develops models capable of learning and providing informed decisions based on large amounts of historical data. The training data used by machine learning algorithms also enable professionals to get an insight into early cancer diagnosis, variation of treatments, and drug discovery [5]. This automatic process of identification and exploration has proved its performance and efficiency in classifying patients’ lung cancer images using various machine learning algorithms [6]. The existing literature has extensively relied on computer tomography (CT) or X-ray images for identifying lung cancer patients. For instance, Vas and Dessai [7] used an artificial neural network (ANN) to categorize the cancer stages based on CT scan images, while Gang et al. [8] employed the parallel immune algorithm for cancer diagnosis based on X-ray images. However, these techniques were criticized for several reasons. First, the errors in the images taken through the CT and X-ray techniques lead to false-negative reports, causing delays in cancer treatment. Second, these techniques can sometimes be difficult for screening patients because of the low number of available devices, high costs, and radiation doses. Sathesh A 2020 proposed a lung nodule segmentation algorithm that uses adaptive weights as a feature for the recurrent neural network. The algorithm initially detects the lung parenchyma from which the background region is minimized. However, the boundaries of the obtained nodule candidate region are not accurate. The evaluation was done with the LIDC datasets using the metrics such as Hausdorff distance, probability rand index (PRI), accuracy, recall, and precision. The scheme provides accuracy, recall, and precision of 94.08%, 89.3%, and 94.1%, respectively [19, 20].
To overcome the limitations of CT and X-ray images, machine learning algorithms have been used to diagnose lung cancer patients by relying on patients’ clinical features. However, there is a lack of research on predicting lung cancer patients’ survival or death period, specifically with the use of deep learning techniques. Unlike classical machine learning, deep learning techniques learn by building a more abstract representation of data as the network expands deeper [10, 11]. This, in turn, helps maximizing the prediction accuracy of deep learning models compared to their counterparts of classical machine learning. Therefore, this research aims to predict lung cancer patients’ survival or death period through a comparison between classical machine learning and deep learning techniques using patients’ demographic and clinical features. The early prediction of the survival period of lung cancer helps patients and healthcare professionals to manage costs better and provide treatment at the appropriate time [12,13,14,15].
2 Proposed Methodology
In recent years, deep learning methods gained significant interest in the segmentation of lung cancer. Deep learning method has gained a significant advantage compared to other approaches. Deep learning puts forward a way to let computers learn the features automatically based on data-driven to reduce the complexity of artificial design features. The deep learning model with essentially enlarged depth advances segmentation performance. Figure 1 depicts the summary of the proposed methodology.
2.1 Pre-processing
We used the randomization strategy as image preprocessing, which could ensure that the deep learning model still maintains strong generalization performance after a large number of repeated training. Multimodal brain images of the same patient use the same processing in one epoch training and different random measures in different epochs. It helps to learn the image features of different modes in the same brain while obtaining generalization. The figure shows the image preprocessing methods: 3D random clipping, 3D random rotation, 3D image intensity random enhancement, 3D image random mirror inversion, and normalization.
Image normalization is a widely used technique in computer vision, pattern recognition and other fields. The z-score normalization was applied in this work. It is defined as per Eq. (1):
where \(\sigma\) is the standard deviation, and \(\mu\) is the mean value. Then, the 3D random clipping method randomly cuts the MRI image (240, 240, 155) into a matrix (144, 144, 128). The 3D random rotation method rotates the reduced image by the angle \(U(-10, +10)\). The random intensity enhancement method of 3D image sets the image pixel value is defined as per Eq. (2):
where \(U\) is the uniform distribution. The random mirror processing symmetrizes the image according to deep, height and width directions. We applied these image enhancement routines to extend the training data set to improve the performance and generalization ability of the deep neural network.
2.2 Deep Learning Model
Deep learning which is a subset of Artificial intelligence is gaining momentum each day by making different tasks much easier and more efficient. CNN which is a type of deep learning mechanism is an inevitable part of image vision problems. In recent years, deep learning methods gained significant interest in the segmentation of lung cancer. The performance of deep learning segmentation methods usually depends on the size of training data. However, it is always difficult to acquire a large number of image data with pixel-level annotation in clinical practice. In order to address the challenge of scarce annotation, studies [12,13,14,15,16,17] have applied few-shot learning on medical image analysis, where the labeled data is a small portion of the whole dataset. For these methods, it is often difficult to effectively utilize the unlabeled data. In recent years, unsupervised learning has been performed on medical image, using unlabeled data for model optimization. The U-shaped model is an efficient and straightforward segmentation network in 3D medical images especially in lung cancer, learning features from deep and shallow neural units. The UNet model consists of four encoder layers and four decoder layers. The proposed model is shown in Fig. 2. In this model, the four channel model inputs correspond to the MRI images of four modes, respectively. The main body of the network is composed of auto-weight dilated (AD) unit, Residual (Res) unit, linear upsampling, and the first and last convolution units. In the downsampling stage (feature coding extraction), we use 8 AD units to obtain multi-scale feature maps. In the upsampling stage (feature decoding), we use the AD unit, Res unit and a linear upsampling layer to form a primary decoding layer. Finally, a convolution unit outputs the results of the network model. Moreover, each convolution unit, AD unit and Res unit contains batch normalization and ReLU functions. We used extended convolution to extract fine-grained and multi-scale glioma features, and employed residual structure to obtain long-dependent glioma features.
As for the Res Unit layer, we used two convolution units to reduce and then enlarge the number of convolution kernels so as to realize feature learning and feature map reorganization. From an experimental point of view, this is an efficient coding method. Then, we used two group convolution units with stripe 1 and group 16, and the kernel size is 3 × 3 × 3. Finally, we used a convolution residual element to obtain the characteristic graph of long dependence.
As for the AD Unit layer, we used two convolution units firstly (like the Res unit). Then, we used three extended convolution units (the divided parameters are 1 and 2, respectively) and used two learnable parameters to adjust and fused the characteristics of the two group extended convolution units. Finally, a group convolution unit was used to output the result of the AD unit. We also set up residual calculations in the AD unit. The dilated convolution could expand the receptive field of the convolutional kernel without sacrificing computational resources, while normal convolution could provide a more accurate feature map. The fusion of the two types of convolutions could strengthen the ability of the network to extract features.
In the encoder stage, each residual block is a dual-pathway structure. In this stage, we set channel depth to 32, 64, 128, and 256. The residual block is the critical structure of down-sampling. In the decoder stage, we connect a convolutional unit and a de-convolutional unit for upsampling. The stripe of the de-convolutional unit is 2 \(\times\) 2 \(\times\) 2, and the kernel size is 3 \(\times\) 3 \(\times\) 3. Batch Normalization and RReLU activation functions are connected behind the convolutional unit or the de-convolutional unit for all convolutional units and all de-convolutional units in this stage. Similarly, we set the channel depth in the decoder stage to 32 \(\times\) 64 \(\times\) 128 \(\times\) 256. A combined deep neural network with the residual blocks enables the network to obtain more significant gradients in deep layers.
So, the phenomenon of gradient disappearance is relatively rare and gets more practical features of gliomas. The formula of the gradient propagation in the convolutional layer can be defined in Eq. (3),
where \({\sigma }^{^{\prime}}\) means the first derivative of the loss function, \(w\) describes the weight, \(O\) indicates the output matrix vector, and \(l\) is the layer \(l\). Then, the gradient in Block-R1, Block-R2 and Block-R3 can be defined in Eq. (4),
where \(f\) means the activation function, \({\delta }_{1}\) and \({\delta }_{2}\) represent the first and second convolution calculations, respectively. It is worth noting that the difference between Eq. (3) and Eq. (4) lies in the order of normalization, which is not reflected in the equation.
Multiplication is widely used in the calculation of series convolution, such as \({\delta }_{2}(f({\delta }_{1}))\). The cumulative multiplication between \((-1, 1)\) makes it possible for the gradient to appear the result of approximate 0, so that the classical gradient disappears. The residual-connection weakens this problem through weight addition, and enhances the stability of the network. Obviously, it is a very effective way to use residual blocks to build architecture in very deep neural layers, especially in calculating the depth feature map.
We defined the convolution block (BN, RL, Conv) in AD unit as per Eq. (5),
where \({c}_{i}\) is the convolution layer \(i\), and \(k\) describes the kernel size, \(f\) is the activation function. The \(w\) and \(b\) represent the convolution weight and bias, \({I}_{i}\) is the input data. Then, the AD block can be define as per Eq. (6)
where \(\ell_{{d}_{0}}\) is the dilated convolution. In the gradient back-propagation process, \(a\) and \(b\) can automatically adjust the weight ratio of the convolution integral branch in the main path. In addition, the channel parameters of the AD-Net were set to 32, 64, 128, 256, and the skip connection adopted the 3D matrix concatenate method. The residual structure was a necessary element. The residual calculation ensured the stability of gradient in deep feature calculation.
3 Experimental Results and Discussion
This research involved 10,001 subjects from the SEER database collected between 2000 and 2019 [16] (SEER, 2021). The age of the patients was ranged between 19 to 94 years old (Mean = 65.85, SD = 9.72). Further, 51.8% of the patients were males (n = 5181) and the remaining were females (n = 4820). There were 1214 (12.1%) alive cases and 8787 (87.9%) death cases. The age of alive cases ranged between 23 and 90 years old (Mean = 64.40, SD = 8.71), whereas the age of death cases ranged between 19 and 94 years old (Mean = 66.05, SD = 9.83). In terms of race, there were 86.3% white, 10.8% black, and 5.1% other nations (i.e., American Indian/AK Native, Asian/Pacific Islander). The alive cases included 571 (47%) males and 643 (53%) female patients. In addition, the death cases included 4610 (52.5%) male patients and 4177 (47.5%) females. An independent sample t-test was performed to test the difference between male and female patients. The outcomes pointed out that there is a significant difference between male and female subjects (t (9999) = 3.55, p < 0.001). This indicates that the probability of death was higher in males than females. The samples lung cancer CT images collected from SEER database are given in Fig. 3.
The following evaluation metrics are used for validating the classifier performance. It is given in Eq. (7)
where accuracy explains how well the architecture can classify the images. It is given as the ratio of total correct prediction made to the total number of predictions made. Precision is given by the ratio of total number of correctly classified positive cases to the sum of total positive cases predicted. High value for precision is required in order to minimize the number of false positive classifications. Recall is defined as the ratio of total positive cases correctly classified to all correctly classified observations and F1 score gives the harmonic mean of precision and recall. It can be also defined as the weighted average of precision and recall [15,16,17,18]. The performance comparison results shown in Tables 1 and 2 suggest that all applied classifiers have a good performance. Experimental results of different models are visualized in Fig. 4 and the classification rate of various models using different statistics method is given in Fig. 5.
4 Conclusion
In this paper, a new multi-scale approach to segment the lung cancer in CT image was described and evaluated in several publicly available databases. This paper also presents an assessment of the most appropriate scales for the lung cancer segmentation, complementing previous work that defines these scales empirically. Furthermore, it was also demonstrated that a multi-scale analysis can improve the lung cancer segmentation. Although recent research has been focusing on deep learning methods, rule-based methods can also be important for the definition of features, that can significantly improve the outcome of these methods. The achieved results show that the proposed approach is very competitive when compared with the current state of the art methods, particularly in high-resolution images. Our method still needs further improvement in the enhancing tumor region segmentation. It were practical tools in 3D lung cancer segmentation.
References
Adem K, Kiliçarslan S (2021) COVID-19 diagnosis prediction in emergency care patients using convolutional neural network. Afyon Kocatepe Üniversitesi Fen Ve Mühendislik Bilimleri Dergisi 300–309
Boddu RSK, Karmakar P, Bhaumik A, Nassa VK, Bhattacharya S (2021) Analyzing the impact of machine learning and artificial intelligence and its effect on management of lung cancer detection in covid-19 pandemic. Mater Today Proc. https://doi.org/10.1016/J.MATPR.2021.11.549
Jayachandran A, Dhanasekaran R (2013) Brain tumor detection using fuzzy support vector machine classification based on a texton co-occurrence matrix. J Imag Sci Technol 57(1):10507-1–10507-7(7)
Brinati D, Campagner A, Ferrari D, Locatelli M, Banfi G, Cabitza F (2020) Detection of COVID-19 infection from routine blood exams with machine learning: a feasibility study. J Med Syst 44(8):135. https://doi.org/10.1007/s10916-020-01597-4
Cai Z, Xu D, Zhang Q, Zhang J, Ngai SM, Shao J (2015) Classification of lung cancer using ensemble-based feature selection and machine learning methods. Mol BioSyst 11(3):791–800. https://doi.org/10.1039/C4MB00659C
Chen YC, Ke WC, Chiu HW (2014) Risk classification of cancer survival using ANN with gene expression data from multiple laboratories. Comput Biol Med 48(1):1–7. https://doi.org/10.1016/J.COMPBIOMED.2014.02.006
Jayachandran A, Kharmega Sundararaj G (2016) Abnormality segmentation and classification of multi model brain tumor in MR images using Fuzzy based hybrid kernel SVM. Int J Fuzzy Syst 17(3):434–443
Dass MV, Rasheed MA, Ali MM (2014) Classification of lung cancer subtypes by data mining technique. In: Proceedings of the 2014 international conference on control, instrumentation, energy and communication (CIEC), pp 558–562. https://doi.org/10.1109/CIEC.2014.6959151
Doppalapudi S, Qiu RG, Badr Y (2021) Lung cancer survival period prediction and understanding: deep learning approaches. Int J Med Inform 148:104371. https://doi.org/10.1016/J.IJMEDINF.2020.104371
Dutta AK (2022) Detecting lung cancer using machine learning techniques. Intell Autom Soft Comput 31(2):1007–1023
Friedman N, Geiger D, Goldszmidt M (1997) Bayesian network classifiers. Mach Learn. https://doi.org/10.1002/9780470400531.eorms0099
Peng G, Yang X, Liu L (2011) Parallel immune algorithm for lung cancer detection in X-ray images based on object shared space. In: 2011 12th international conference on parallel and distributed computing, applications and technologies, pp 197–200. https://doi.org/10.1109/PDCAT.2011.64
Gao Y, Lyu Q, Luo P, Li M, Zhou R, Zhang J, Lyu Q (2021) Applications of machine learning to predict cisplatin resistance in lung cancer. Int J Gener Med 14. https://doi.org/10.2147/IJGM.S329644
Hsu CH, Manogaran G, Panchatcharam P, Vivekanandan S (2018) A new approach for prediction of lung carcinoma using back propagation neural network with decision tree classifiers. In: 2018 IEEE 8th international symposium on cloud and service computing (SC2), pp 111–115. https://doi.org/10.1109/SC2.2018.00023
Ibrahim DM, Elshennawy NM, Sarhan AM (2021) Deep-chest: Multi-classification deep learning model for diagnosing COVID-19, pneumonia, and lung cancer chest diseases. Comput Biol Med 132:104348. https://doi.org/10.1016/J.COMPBIOMED.2021.104348
Prabhu AJ, Jayachandran A (2018) Mixture model segmentation system for parasagittal meningioma brain tumor classification based on hybrid feature vector. J Med Syst 42(12)
SEER (2021) SEER Incidence Data, 1975 - 2018. National Cancer Institute (NCI). https://seer.cancer.gov/data/
Namboodiri S, Jayachandran A (2020) Multi-class skin lesions classification system using probability map based region growing and DCNN. Int J Comput Intell Syst 13(1):77–84
Sathish A (2020) Adaptive shape based interactive approach to segmentation for nodule in lung CT scans. J Soft Comput Paradigm 2(4):216–225
Sungheetha A, Rajesh SR (2020) Comparative study: statistical approach and deep learning method for automatic segmentation methods for lung CT image segmentation. J Innov Image Process 2:187–193
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Jayachandran, A., Anisha, N. (2023). Lung Cancer Classification System for CT Images using Deep Convolutional Neural Network. In: Joby, P.P., Balas, V.E., Palanisamy, R. (eds) IoT Based Control Networks and Intelligent Systems. Lecture Notes in Networks and Systems, vol 528. Springer, Singapore. https://doi.org/10.1007/978-981-19-5845-8_28
Download citation
DOI: https://doi.org/10.1007/978-981-19-5845-8_28
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-5844-1
Online ISBN: 978-981-19-5845-8
eBook Packages: EngineeringEngineering (R0)