Introduction

Lung cancer is considered as a second most probable cancer in human beings. Moreover, the graph of survival rate is curved toward the very low end. Due to this fatal disease, around 1.69 million deaths have been reported worldwide during the year 2015 along with several new cases [1]. The human cells grow in a well-controlled manner. The uncontrolled growth of these cells leads to cancer. In lung cancer, these abnormal cells grow quickly to form a nodule [2]. As the nodule size increases, lungs functionality is severely disturbed and leads to death within a short time [3]. Therefore, it has become a compulsion to identify lung nodules at earlier stages. Medical imaging helps in early diagnosis of lung cancer. The most popular one is computed tomography (CT) [4]. During a CT scan, more than 100 images of a patient are generated. Radiologists have to examine all these images during the diagnosis. Handling such large number of images is error-prone and time-consuming. Computer-based systems help the radiologists in diagnosis process by providing the second opinion [5]. However, there is a significant challenge of sensitivity and number of FPs/scan. One of the major steps of a diagnosis system is segmentation. Accurate segmentation of lung region and pulmonary nodules is critical that helps the radiologists in decision making [6]. This paper presents a multistep segmentation method to improve the diagnosis procedure. In addition to segmentation, the selection of relevant features and a robust classifier helps to improve accuracy and reduce false positives. Considering aforementioned, GTFD is created for better representation of nodule candidates. Finally, SVM-ensemble classifier is applied, which is based on three base classifiers, namely SVM-linear, SVM-polynomial and SVM-RBF (radial-based function). The proposed method is evaluated over a widely used publicly available standard dataset: LIDC–IDRI. The results show improved performance of the proposed method in terms of accuracy, sensitivity and specificity with low FPs/scan as compared to earlier methods.

Related work

Automated diagnostic systems for lung cancer have remained a key area of research in the recent years. Several such systems have been developed based on image processing and machine learning techniques [7]. However, there is a primary issue of accuracy and false positive rate (FPR) in these diagnosis systems. Lung nodule detection methods are mainly comprised of segmentation, feature extraction and classification [8]. The segmentation is further divided into lung region extraction and nodule candidate detection. Several segmentation methods have been proposed including thresholding [9], template matching [10], 3D template matching [11] to name a few. Thresholding is a simple and popular technique used for lung volume segmentation and nodule candidate detection. Different variations of thresholding techniques are reported in literature including simple thresholding [9], optimal thresholding [12], Otsu thresholding [13], thresholding combined with morphological operations [14], fuzzy thresholding [15] and multiple thresholds [16,17,18,19]. Some researchers also used combinations of different segmentation methods to improve the segmentation quality. ..Tan et al. [20] proposed a segmentation method by combining watershed, active contours and Markov random field. ..Shen et al. [21] used bidirectional chain code for nodule segmentation and then SVM for classification. The use of classifier is also a very significant step in the reduction in false positives. It distinguishes actual nodules from the set of nodule candidates on the basis of features provided for training and testing. Artificial neural networks (ANN) [22, 23] and SVM are popular techniques used for nodule classification. ..Nibali et al. [24] proposed deep residual network-based nodule classification method. Their major focus was on the performance evaluation of deep learning in the field of lung nodule detection. ..de Carvalho Filho et al. [25] applied thresholding combined with a genetic algorithm for nodule segmentation and finally SVM for classification. Taşcı, Uğur [13] presented a segmentation method based on Otsu thresholding and morphological operations. Shape and texture features were used as input to a regression-based classifier. .Teramoto, Fujita [26] proposed the cylindrical nodule enhancement filter to improve FPR. They used thresholding for segmentation of region of interest (ROI) and SVM for nodule classification.

Segmentation of lung nodules is a challenging phase in the detection process due to complex structure of lung. Furthermore, handling of small, boundary attached (juxtapleural) and vessel attached (juxtavascular) nodules is a big challenge. Precise feature extraction and powerful classification algorithms assert their impact on the performance of a diagnosis system. The proposed technique addresses these issues in order to give a better detection method for lung cancer.

Proposed method

An automated diagnosis method to detect lung nodules on the basis of various image processing and machine learning techniques is presented in this paper. Figure 1 shows working model of the proposed method. Segmentation comes first as a vital step in the presented technique which deals with two major parts, i.e., lung region extraction and nodule candidate detection. These two stages are further divided into substeps. Other two stages are the creation of GTFD and SVM-ensemble-based nodule classification, respectively.

Fig. 1
figure 1

Workflow of the proposed method

Nodule segmentation

Pulmonary nodule segmentation is an escalating phase in the performance of an automatic diagnostic system. During the diagnosis procedure, precise and accurate segmentation of nodules impacts the performance. Segmentation of smaller nodules, ranging from 3–30 mm, juxtapleural nodules and vascular attached nodules is a significant test for the segmentation method.

Lung region extraction

In the first step, the image background is removed by using corner-seeded region growing combined with thresholding approach. The thresholding is a straightforward and efficient method for image segmentation. It performs segmentation on the basis of pixel intensities. It is observed that simple thresholding is not useful for lung region extraction due to overlapping between the intensities of background and some sections of ROI. To overcome this issue, the background of input images is removed. Furthermore, the threshold generated for one slice is not useful to other slices because of its variation in the gray level for each slice as shown in Fig. 2. Therefore, the suggested segmentation technique used a combination of differential evolution-based optimal thresholding [27] and corner-seeded region growing.

After the removal of image background, differential evolution-based optimal thresholding is used to identify the lung boundary and extraction of lung region. An initial threshold of − 950 HU is applied because most of the lung region lies in the range of − 950 HU to − 500 HU. It is an iterative step in which the threshold is recalculated in each iteration.

Fig. 2
figure 2

Input CT images (a, b) and their respective histograms (c, d)

In an image, a histogram gives the probability distribution of gray levels. This probability distribution is calculated by Eq. (1).

$$\begin{aligned} p(x)=\mathop {\sum }\limits _{i=1}^K P_i .p_i (x)=\mathop {\sum }\limits _{i=1}^K \frac{P_i }{\sqrt{2\pi \sigma _i }}\hbox {exp}\left[ {\frac{-\left( {x-\mu _i } \right) ^{2}}{2\sigma _i ^{2}}} \right] \nonumber \\ \end{aligned}$$
(1)

The total number of categories within the image is represented by K. \(P_i\) and \(p_i(x)\) are the probability and probability distribution functions within the category i, respectively. \(\mu _i\) and \(\sigma _i\) are the mean and standard deviation. To compute optimal threshold, the overall probability error in two different categories is minimized by Eq. (2).

$$\begin{aligned} e(T_i)=P_i \mathop {\int }\limits _{-\infty }^{T_i } p_i (x)\hbox {d}x+P_{i+1}\mathop {\int }\limits _{T_i }^\infty p_{i+1} (x)\hbox {d}x \end{aligned}$$
(2)

This error is with respect to the threshold \(T_i \). Then the overall error is calculated as given by Eq. (3).

$$\begin{aligned} E(T)=\mathop {\sum }\limits _{i=1}^{K-1} e(T_i) \end{aligned}$$
(3)

Henceforth, a threshold image is generated that contains the lung mask. To refine the lung boundary and extraction of juxtapleural nodules, morphological operations are applied. Morphological closing is performed on the binary mask, where a disk of size 25 is taken as the structuring element. The process of closing also includes some extra regions which can affect the accuracy of segmentation. To solve this problem, morphological opening with the same structuring element, a disk of size 25, is applied. Figure 3 highlights the process of juxtapleural nodule extraction. Finally, the lung region is extracted by using the resultant mask and input CT image. The whole procedure is repeated for all the slices of each scan. Figure 4 shows the step-by-step results of segmentation.

Fig. 3
figure 3

a Input images with juxtapleural nodules, b corresponding lung masks missing plural nodules, c boundary correction using morphology, d extracted lungs with juxtapleural nodules

Fig. 4
figure 4

a Input image, b background removed, c lung mask, d extracted lung

Nodule candidate detection

After ROI extraction, the next step is to extract nodule candidates by eliminating vessels and other unwanted objects. For this purpose, a combination of morphological operations, edge detection, bounding box and shape information are applied to the lung ROI for eliminating vessels and noise objects. The major issue in nodule candidate detection is juxtavascular nodules due to their attachment to the vessels [28]. In most of the previously known techniques, these nodules are eliminated with the removal of vessels. To avoid nodule elimination, we use 3D connectivity information. Since nodule is a 3D object, therefore, it must show its existence in consecutive slices of a scan. Considering this property of nodule, the presence of each object with its predecessor and successor slices is investigated. Let p(ijk) is a candidate for a nodule pixel, where k is the slice number and i, j represent the location of pixel within that slice. Then 3D neighborhood is checked by comparing the intensity of that pixel. Each pixel p(ij) of slice k has nine neighborhoods in the \(k-1\) slice and nine in the \(k+1\) slice as shown in Fig. 5. The similar intensity pixels must be there within the eighteen neighborhood pixels, either in predecessor or in successor slices or in both.

Fig. 5
figure 5

3D connectivity

The morphological opening is used to separate juxtavascular nodules from vessels. The structuring element used for this purpose is a ‘sphere’ of size 2 which is effective in 3D lung structure. The nodule is a round-shaped object and the use of a ‘sphere’ as structuring element preserves the nodules and breaks the connection of nodules and vessels. Afterward, 3D connected component analysis is performed and regions with centers within 5 mm are merged. Canny edge detection method is used to identify the boundaries of each object. It results in a large number of candidate objects. To refine this set of objects, the nodule size and shape information are taken into account. The nodule is round-shaped object ranges from 3 to 30 mm in size within ROI. Considering these properties, the nodule candidate set is refined on the basis of diameter, elongation and circularity. The objects with diameter less than 3 mm and greater than 30 mm are removed and considered to be unwanted objects as their sizes do not match with the size specifications of nodules. In addition to diameter, those objects are excluded which are too elongated i.e., their major axis length is 2.25 times more than the minor axis length. The third parameter is circularity, which defines roundness of the object. The roundness of an object ranges from 0 to 1, where 1 represents perfect round. All those objects which have circularity less than 0.65 are also excluded from the candidate set. The elongation and circularity threshold values are empirically selected based on the observation that if circularity value is increased more than 0.65 and elongation value is decreased less than 2.25, this causes the removal of some of the actual nodules. On these parameters, nodule candidate set is refined and nodules are distinguished from other objects like vessels and noise objects as presented in Fig. 6. Some of the extracted nodule candidates are illustrated in Fig. 7.

Fig. 6
figure 6

a Objects in ROI b Nodule candidate regions

Fig. 7
figure 7

Extracted nodule candidates

Geometric texture feature descriptor (GTFD)

In object detection and classification, precise and relevant features play a significant role. In recent studies, a variety of features including geometric, texture, local binary patterns (LBP), wavelet, intensity and gradient are used for the detection and classification of nodules because a single type of feature vector cannot give an accurate representation of objects resulting in unsatisfactory performance. To overcome these issues, researchers also used hybrid feature vectors by encapsulating more types of features in a single feature vector. In this study, a GTFD by combining texture and shape features in 2D and 3D is devised that successfully delineates the nodules and non-nodules.

Geometric features

Geometric features like circularity, area and volume are the shape descriptors of an object [29]. Therefore, these features are helpful to distinguish nodules from other objects within the lung region. The median slice of a scan contains maximum information of objects present in the scan. Therefore, 2D geometric features are extracted from the median slices. However, 3D geometric features are extracted from 3D segmented objects. The features include area, radius, circularity, volume, compactness and elongation. Equations (4)–(9) give the mathematical representation of these features.

$$\begin{aligned}&\hbox {Area}\,(A)=\mathop {\sum }\limits _{o\epsilon S} p \end{aligned}$$
(4)
$$\begin{aligned}&\hbox {Radius}\,(r)=\frac{{D}}{2} \end{aligned}$$
(5)
$$\begin{aligned}&\hbox {Circularity}=\frac{{A}}{4\pi {r}^{2}} \end{aligned}$$
(6)
$$\begin{aligned}&\hbox {Volume}=\mathop {\sum }\limits _{{v}\in {V}} {v} \end{aligned}$$
(7)
$$\begin{aligned}&\hbox {Compactness}=\frac{{A}}{\frac{4\pi {r}^{3}}{3}} \end{aligned}$$
(8)
$$\begin{aligned}&\hbox {Elongation} = \frac{{\hbox {Minor}\ \hbox {Axis}\ \hbox {Length}}}{{\hbox {Major}\ \hbox {Axis}\ \hbox {Length}}} \end{aligned}$$
(9)

where p is pixel within the segmented object O of slice S of a 3D scan, D represents diameter of the nodule candidate and V accounts for the voxels of a candidate object.

Texture features

Texture features characterize the regions of images on the basis of their pixel intensities [30]. Ten texture features, four 2D and six 3D features, are used in this study. The 2D texture features are given by Eqs. (10)–(13) and 3D texture features are specified from Eqs. (14)–(18).

$$\begin{aligned}&\hbox {Mean}\,(\mu )=\frac{\mathop {\sum }\nolimits _{i=0}^I p(i)}{I+1} \end{aligned}$$
(10)
$$\begin{aligned}&\hbox {Variance}\,(\sigma ^{2}) = \frac{\mathop {\sum }\nolimits _{i=0}^I (\mu (i)-\mu )^{2}}{I+1} \end{aligned}$$
(11)
$$\begin{aligned}&\hbox {Skewness}\,(\mu _3) = \frac{\mathop {\sum }\nolimits _{i=0}^I (\mu (i)-\mu )^{3}}{\sigma ^{3}} \end{aligned}$$
(12)
$$\begin{aligned}&\hbox {Energy}_1\,(E_1) = \mathop {\sum }\limits _{i=0}^I (p(i))^{2} \end{aligned}$$
(13)
$$\begin{aligned}&\hbox {Energy}_2\,(E_2) = \mathop {\sum }\limits _{i=0}^I \mathop {\sum }\limits _{j=0}^I [p(i,j)]^{2} \end{aligned}$$
(14)
$$\begin{aligned}&\hbox {Contrast}=\mathop {\sum }\limits _{i=0}^I \mathop {\sum }\limits _{j=0}^I (i-j)^{2}p(i,j) \end{aligned}$$
(15)
$$\begin{aligned}&\hbox {Correlation}=\mathop {\sum }\limits _{i=0}^I \mathop {\sum }\limits _{j=0}^I \frac{ij p(i,j) -\mu _x \mu _y}{\sigma _x \sigma _y } \end{aligned}$$
(16)
$$\begin{aligned}&\hbox {Homogeneity}=\mathop {\sum }\limits _{i=0}^I \mathop {\sum }\limits _{j=0}^I \frac{p(i,j)}{1+(i-j)^{2}} \end{aligned}$$
(17)
$$\begin{aligned}&\hbox {Entropy}(H) = -\mathop {\sum }\limits _{I=0}^I \mathop {\sum }\limits _{j=0}^I p(i,j)\ \hbox {log} _2 [p(i)] \end{aligned}$$
(18)

where p(i) is ith pixel in a 2D slice, p(ij) is ith voxel in 3D volume at slice j and I represents maximum intensity within the slice.

Ensemble classification

SVM-ensemble classification is used to identify the actual nodules from a set of candidate nodules. An ensemble classifier makes final decision on the test dataset derived from the output of individually trained base classifiers [31]. Let there are n base classifiers as: \(\{ {C_1 ,C_2 ,C_3 \ldots C_n}\}\) over a dataset X and target T. All classifiers are trained and then evaluated over the same test examples. The decisions of individual classifiers are then combined to make a final decision. Majority voting is a useful combiner in which decision about the class of a test example is made through the number of votes, determined by Eq. (19).

$$\begin{aligned} f_e (x)=\max \mathop {\sum }\limits _{j=1, f(j)\in C_i }^n 1 \end{aligned}$$
(19)

where \(f_e(x)\) assigns class labels to the test feature vector x on the basis of maximum votes given to a class and n represents the total number of base classifiers. Let if there are three base classifiers, two predict that the test example is of class A and one predicts it in class B, then majority voting predicts it as class A. In this study, SVM is used as a base classifier with three different kernels including linear, polynomial and Gaussian RBF. These kernels are represented mathematically in Eqs. (20)–(22).

$$\begin{aligned}&\hbox {Linear Kernel} \quad K(x_i ,x_j )= x_i^T . x_j \end{aligned}$$
(20)
$$\begin{aligned}&\hbox {Polynomial Kernel} \quad K(x_i ,x_j )=(1+ x_i^T x_j )^{d} \end{aligned}$$
(21)
$$\begin{aligned}&\hbox {Radial-Basis Function} \quad K(x_i,x_j)\nonumber \\&\quad =\hbox {exp}\left( {-\frac{\Vert x_i -x_j \Vert ^{2}}{2\sigma ^{2}}}\right) \end{aligned}$$
(22)

where \(x_i \) and \(x_j \) are vectors in input space, d represents the degree of polynomial, \(\Vert x_i -x_j \Vert ^{2}\) is the Euclidean distance and \(\upsigma \) is standard deviation, defined as Gaussian distribution. SVM is known to be a good classifier that learns easily and produces good classification results. However, a single classifier may not learn all the parameters to produce good global optimum result. To meet this limitation, an SVM-based ensemble classifier is developed for nodule classification. Figure 8 shows the workflow of SVM-ensemble classifier for the proposed method to distinguish between nodules and non-nodules. During the training phase, each individual SVM-based classifier is trained over the labeled data provided for training. SVM-linear, SVM-polynomial and SVM-RBF are trained as base classifiers. The individual classifiers are aggregated on the basis of voting. In the testing phase, an unlabeled test example is supplied as input to all the individual classifiers simultaneously and final decision is made on the basis of Eq. (19).

Fig. 8
figure 8

Workflow of SVM-ensemble for nodule classification

Table 1 Results with geometric features

Results and discussions

In this study, an authentic publicly available dataset LIDC–IDRI [32] is used to evaluate the performance of proposed method. The tube current range, from 40 to 627 mAs, is used for image acquisition. The tube potential energies have a range from 120 to 140 kV. The low-dose scans are more effective to identify small size lung nodules [33]. The dataset contains 1018 CT scans of 1010 patients collected from 7 academic centers. Each scan contains approximately 100 to 400 slices with slice thickness between 0.6 to 5 mm and dimension of \(512\times 512\) pixels. Each image has a 12-bit grayscale resolution with the pixel size ranging from 0.5 mm to 0.8 mm. Out of 1018 cases, 121 cases have slice thickness \(\ge \) 3 mm which is not suitable for the computer-based diagnosis [34]. In addition to this, there are nine cases which have inconsistent slice spacing. By excluding these cases, a total of 888 scans are left for the automated diagnosis. Every scan of dataset is reviewed by four experienced radiologists for nodule annotation. The annotation information for each scan is provided in XML format. All the radiologists have a consensus on 777 nodules. In this study, those nodules are considered for experimentations which are marked by all the four radiologists. Around 45% scans of dataset contain no nodules which are annotated by all the four radiologists and some scans have multiple nodules. Considering these points, those scans are excluded which have zero nodules and those scans are selected which contain number of nodules in the range of 1–8. As a result, 250 scans are finally selected with 567 nodules. Candidate nodules are classified according to annotations provided by the expert radiologists. If a nodule candidate meets the criteria, it is considered as hit and classified as a nodule. On the other hand, if it does not meet the criteria, it is taken as miss and considered as a non-nodule. A candidate is considered hit if its radius is 0.8 to 1.5 times of the actual nodule or its distance with reference to any corresponding annotated nodule is less than 1.5 times the radius of that nodule .[12]. The performance of classifiers is checked over different distributions of data into training and testing ratio. Following evaluation parameters, as mentioned from Equations (23) to (26), are used to analyze the performance.

$$\begin{aligned}&\hbox {Sensitivity} = \frac{\hbox {TP}}{\hbox {TP}+\hbox {FN}} \end{aligned}$$
(23)
$$\begin{aligned}&\hbox {Specificity}=\frac{\hbox {TN}}{\hbox {TN}+\hbox {FP}} \end{aligned}$$
(24)
$$\begin{aligned}&\hbox {Accuracy} = \frac{\hbox {TP}+\hbox {TN}}{\hbox {TP}+\hbox {FP}+\hbox {TN}+\hbox {FN}} \end{aligned}$$
(25)
$$\begin{aligned}&\hbox {False\ Positive\ Rate}\;\hbox {(FPR)} = \frac{\hbox {FP}}{\hbox {TN}+\hbox {FP}} \end{aligned}$$
(26)

MATLAB 2016a is used as an implementation tool. Three different distributions of data are used as 80–20, 70–30 and 50–50% for training and testing purpose, respectively. The hold-out validation is used for training and testing the classifiers. The process is repeated 10 times to avoid the biasness and average score is computed. Various experiments are executed with different combinations of feature sets. The results are shown in Tables 1, 2 and 3 with geometric, texture and GTFD features, respectively. SVM is trained with three different kernels, namely SVM-linear, SVM-polynomial and SVM-RBF. The output of individual classifiers over the test examples is evaluated. Majority voting is applied to make a final decision by SVM-ensemble. From Tables 1, 2 and 3, it is clear that SVM-ensemble has produced much better results as compared to individual classifiers.

Table 2 Results with texture features
Table 3 Results with GTFD

Performance of the proposed method is evaluated with the individual as well as combined feature sets. The results show that combined feature vectors produced a better performance as compared to single feature type. The performance with texture features is lower than the geometric features. However, when GTFD is used by encapsulating both geometric and texture features, much better results are obtained as compared to single type of feature vector. Furthermore, best results are shown when experimentation is done over the combination of 70–30 i.e., 70% training and 30% testing. Table 3 depicts that GTFD presented better results than individual features. In comparison with Tables 1 and 2, it gives a clear picture that geometric features are better descriptors of nodules as compared to texture features. However, on an individual basis, they cannot produce best results. Figure 9 illustrates a receiver operating characteristic (ROC) curve for the results of classification. It is evident that ensemble-based SVM shows better results as compared to RBF, polynomial and linear kernels.

Fig. 9
figure 9

ROC curve with GTFD feature vector

Fig. 10
figure 10

a FROC curve shows false positive reduction by each classifier, b FROC curve illustrates sensitivity comparison of each classifier at 3.4 FPs/scan

Table 4 Comparison of datasets used in existing and proposed systems

The nodule candidate detection process comprises of two stages. In the first stage, 66021 nodule candidates are detected with 261.8 FPs/scan. At this step, all 567 actual nodules are also recognized with 100% sensitivity. The second stage is related to candidate refinement on the basis of nodule size and shape information. In this stage, the number of candidates is reduced to 46176 and FPs/Scan is decreased to 182.5 with 45614 total false positives. However, this refinement phase has missed 5 actual nodules and sensitivity is reduced to 99.1%. The classification phase, based on SVM-ensemble, reduces the number of false positives to 838 which is 3.4/scan. The free-response receiver operating characteristic (FROC) curve in Fig. 10a illustrates the number for FPs/scan produced by four classifiers. The SVM with linear kernel delivers 9.6 FPs/scan with 89.1% sensitivity, SVM-polynomial produces 6.4 FPs/scan with 93.6% sensitivity and SVM-RBF shows comparatively good results with 4.2 FPs/scan and sensitivity of 97.7%. However, SVM-ensemble reduces the FPs/scan to 3.4% with the highest sensitivity of 98.6%. In Fig. 10b, another FROC curve is given which illustrates the comparison of classifiers’ sensitivity at 3.4 FPs/scan. It depicts that SVM-ensemble has highest sensitivity i.e., 98.6% and SVM-linear has the lowest. SVM-RBF has better sensitivity than SVM-polynomial at 3.4 FPs/scan. It is evident from FROC curves that minimum FPs/scan with SVM-ensemble is reduced to 1 but at the cost of sensitivity degradation with 92.1%.

Table 5 Comparison of results with existing methods

For estimating the significance of proposed method in relation to reported techniques in the literature, a comparison of datasets and their results is presented in Tables 4 and 5, respectively.

Ye et al. [15] used clinical data of 108 scans containing 220 nodules of size 3–20 mm with a slice thickness of 0.5–2 mm. Cascio et al. [22] and Choi, Choi [12] used 84 CT scans of LIDC dataset consisting of 148 nodules marked by any of the radiologists. Dehmeshki et al. [10] made use of 70 CT scans of a clinical dataset containing 179 nodules of size 3–20 mm with a slice thickness of 0.5–1.25 mm. Kuruvilla, Gunavathi [23] utilized a clinical dataset of 155 scans with a slice thickness of 0.75–1.25 mm including 110 nodules. In addition, they have also used some scans of LIDC dataset without LIDC annotations. Han et al. [19] used 205 CT scans of LIDC–IDRI dataset containing 490 nodules marked by any of the radiologists. The nodules included in their study were of 3–30 mm size. Teramoto, Fujita [26] made use of 84 CT scans of LIDC dataset with 103 nodules of size 5–20 mm. The nodules included in the work were marked by any of the radiologists. Nibali et al. [24] utilized LIDC–IDRI dataset for experimentations but details are not provided. Tan et al. [9] made use of 125 CT scans of LIDC dataset containing 80 nodules of size 3–30 mm marked by all the four radiologists. Riccardi et al. [14] used 154 CT scans of LIDC dataset consisting of 117 nodules with size 3–30 mm marked by all the four radiologists. Proposed method is evaluated over 250 scans of LIDC–IDRI dataset containing 567 nodules with size 3–30 mm annotated by all the four radiologists. The statistics in Table 4 highlight the fact that the proposed method is tested over a more significant number of scans and nodules as compared to other research works reported in the literature.

Table 5 compares the results of the proposed and existing methods. Kuruvilla, Gunavathi [23] produced 91.4% sensitivity but with a very high FPs/scan of 30. Moreover, they did not provide specificity. Dehmeshkiet.al [10] also produced very high FPs/scan of 14.6 with 90% specificity. These results were taken on clinical datasets instead of some standard dataset. Choi, Choi [12] claimed 97.5% sensitivity as well as specificity with 6.76 FPs/scan. Ye et al. [15] produced 8.2 FPs/scan with 90.2% sensitivity. Cascio et.al [22] claimed 88% sensitivity and 97% accuracy with 6.1 FPs/scan. Teramoto, Fujita [26] achieved a low FPs/scan of 4.2 but at the cost of accuracy which is as low as 80%. Han et al. [19] claimed a low FPs/scan of 4.14 but their sensitivity was low as 89.2%. Nibali et al. [24] showed 89.9, 91.07 and 88.64% accuracy, sensitivity and specificity, respectively. Tan et al. [9] produced 87.5% sensitivity with 4.0 FPs/scan and Riccardi et al. [14] claimed 6.5 FPs/scan with a low sensitivity of 71.0%. By concluding the whole discussion, it is evident that the proposed method has shown better results i.e., 99.0% accuracy, 98.6% sensitivity and 98.2% specificity with a 3.4 FPs/scan which is a clear proof of the effectiveness of the presented method.

Conclusion

A critical issue with the nodule diagnosis methods is handling small, boundary attached and vessel attached nodules. If a detection method misses any nodule, it leads to inaccurate diagnosis which might be life threatening. To handle such problems, a nodule segmentation method from 3D CT scan is presented. The first step of proposed method is based on hybrid multistep segmentation technique to extract lung ROI and nodule candidates. SVM-ensemble is used to distinguish nodules from the set of selected candidates. The classification is based on a GTFD feature vector encapsulating texture and geometric features. The proposed method is tested on a standard publicly available dataset LIDC–IDRI. The presented method achieved 99.0% accuracy with 98.6% sensitivity and 98.2% specificity. The number of FPs/scan is also low as 3.4/scan. These results show the significance of presented method that it will facilitate radiologists during the diagnosis process of lung cancer.