1 Introduction

Because false mutations radically alter human existence, the genetic issue affects the majority of people in new technological breakthroughs. The DNA functions and structure are totally changed based on the false mutation [15]. External variables such as chemical gas exposure, alcohol consumption, and population air-breathing all contribute to abnormal mutation. Mainly, tumors are created based on the mutation of the abnormal cell (DNA) that arises in different places of the human body like the brain, breast, skin, and lung [6]. Due to the external factors, the most caused disease is lung cancer (LC) among the several tumors that generally affects there spiratory system of human.

In the world, the most popular kinds of malignant tumor are lung cancer, which takes second place based on a study of the United States. The two kinds of LC are non-small cell lung cancer (NSCLC) and small cell lung cancer (SCLC) respectively [27]. Lung cancer diagnoses are performed via various processes like CT, X-ray, isotope, and MRI. Two recognizable anatomic imaging are Computer Tomography (CT) and X-ray chest radiography [17]. The presence of disease is recognized and identified by radiologists and physicians using CT images thereby describing the severity of diseases. The scan time is shortened in which the spiral scans introduce a volumetric CT technique.

With the advancement of CT technology, high-resolution CT examinations have become the imaging modality of choice for the diagnosis and recognition of lung diseases [2]. Improved anatomic resolution with lung imaging based on High-Resolution Computed Tomography (HRCT) [18]. Several computer-aided diagnostic models and detection techniques for lung cancer segmentation have been developed during the last several decades [11]. The existing techniques like deep learning (DL), recurrent 3d-denseness (R3-D), hybrid segmentation network (HSN), deep active learning (DAL), etc. are used for lung cancer segmentation [16]. The development of computer-aided systems promotes the necessity for objective and reliable analysis. Fast and accurate lung cancer segmentation is the major aim of this study.

The early-stage lung cancer patient often possesses no symptoms which increases the complexity of disease diagnosis. Hence early diagnosis plays a main role in lung cancer diagnosis. The CT scans are mainly utilized in this work to overcome the shortcomings (low contrast, similar tissue absorption, and plane overlap) associated with the two-dimensional X-rays since it provides high-density resolution. The lung disease can be identified in an early stage via the CT images since it helps to identify the normal and malignant lung nodules accurately in a non-invasive manner. Because CT scan images contain large scanning layers, the radiologist encounters numerous obstacles while analyzing them. Processing mass CT images is often time-consuming, involves extensive manual intervention, and results in incorrect results. Hence the clinical needs cannot be met by manual segmentation and it needs an artificial intelligence technique which can identify the tumor lesions in a massive dataset [10].

As a result, manual segmentation cannot meet the clinical needs, and an artificial intelligence technique that can identify tumor lesions in a large dataset is required. The Convolutional Neural Network (CNN) architecture has demonstrated its ability to overcome these challenges by dynamically monitoring the liver tumor areas, resulting in accurate diagnosis. Despite the fact that the CNN architecture has many advantages, it has a number of drawbacks, including low generalization and interpretation capabilities, low robustness, and data accessibility constraints. This paper presents a modified optimization algorithm to overcome the complexities associated with the CNN architecture. The major contribution of this study is summarized below:

  • During pre-processing, the histogram equalization (HE) and adaptive bilateral filter (ABF) are used to increase the CT scan image contrast level and enhance CT scan images for de-noising.

  • We combined and used Modified mayfly optimization and modified particle swarm optimization (M2PSO) algorithms thereby obtaining the improved segmentation accuracy.

  • The performance of ensemble deep convolutional neural network (EDNN) for lung cancer segmentation is enhanced by combining M2PSO algorithms.

  • The Shandong Cancer Hospital Affiliated to Shandong University (SHFSU) offers dataset image information as well as measurements such as dice similarity score (DSS), accuracy, sensitivity, dice loss, and generalized dice loss to assess the efficacy of the proposed model.

The rest of the article is shortened as follows: The literature survey is analyzed in Sect. 2. Section 3 describes the proposed methodology. The experimental investigations are examined in Sect. 4 and Sect. 5 concludes the article.

This study proffers solutions to the issues of high computational cost due to the implementation of facial expression recognition by providing a model close to the accuracy of the state-of-the-art model. The study concludes that deep learning-enabled facial expression recognition techniques enhance accuracy, better facial recognition, and interpretation of facial expressions and features that promote efficiency and prediction in the health sector.

2 Related works

Deep learning (DL) methods were suggested by Li et al. [16] for the segmentation of lung cancer. From 200 patients, an annotated dataset of 50 test images and 150 training images were used to focus the cancer tissue segmentation. The metrics like DICE coefficient, specificity, sensitivity, accuracy, and precision were evaluating the performance of DL techniques thereby obtaining better segmentation accuracy with higher computational complexities. Men et al. [20] introduced deep active learning (DAL) for lung cancer segmentation. Depending upon the deep learning activity, the candidate set provides the selected quality images. The Hausdorff distance and Dice are the segmentation metrics to validate the performance of the DAL model. Various sensitivity, balanced, and accuracy outputs were obtained with respect to the right lung, left lung, spinal cord, esophagus, and heart. But, this method had taken a larger execution time for implementation.

The recurrent 3d-denseunet (R3-D) was introduced by Kamal et al. [14] for the segmentation of the lung cancer tumor region. The fine-grained Spatio-temporal information was extracted with the help of multiple Convolutional Long Short-Term Memory (CLSTM). Several 3D-convolutional layers have a decoder and encoder blocks. In practice for this task, various loss functions based extensive ablation study was performed. The segmentation result with a 0.7228% average dice score was achieved than other state-of-art techniques but the segmentation accuracy is lower. A hybrid segmentation network (HSN) was introduced by Chen et al. [9] for small cell lung cancer segmentation. The severe class data imbalance problem was tackled by utilizing the generalized Dice loss function (GDLF). Depending upon the experimental results, the HSN model provided 0.909% precision, 0.872% sensitivity score, and 0.888% dice score values. This method failed to include the healthy people scan in the dataset.

Skourt et al. [26] suggested deep neural networks (DNN) for lung segmentation CT images. The symmetric expanding path was recovered and high-level information was extracted via the contracting path. The dice-coefficient index was 0.9502 and proved accurate segmentation based on the U-net architecture. The computational complexity and time delay were higher. For chest X-ray analysis of lung cancer, Gordienko et al. [12] suggested the deep learning (DL) model. Without rib and clavicle shadows, the original JSRT dataset was used to perform both training and validation. In the simplified configuration, the pre-processing techniques were considered to demonstrate usefulness and high-efficiency results. This method is unsuitable to train larger and more complicated image models.

The K-nearest neighbor (KNN) [23], deep learning (DL) [24], and Convolutional Neural networks (CNN) [8] tend to be promising in cancer tumor segmentation. However, the features learned by these models are often redundant. Even though, these works offer improved performance still they need modifications to improve the accuracy of the lung cancer segmentation and search the smaller homogeneous regions. The summary of the literary works is presented in Table 1.

Table 1 Summarization of the literary works

3 Proposed methodology

This section proposed a novel approach for lung tumor segmentation. Initially, the input dataset image is fed for pre-processing. From the pre-processed image, the lung tumors are segmented through a deep convolutional neural network with a modified mayfly-based modified particle swarm optimization (M2PSO) algorithm. The overall proposed architecture is depicted in Fig. 1.

Fig. 1
figure 1

Overall proposed architecture

3.1 Pre-processing

Histogram equalization (HE) enhances the input CT scan image contrast during pre-processing. Equation (1) shows the contrast enhancement of the CT image (AI) by applying histogram equalization.

$$AI={~}^{\Delta \mathrm P}\!\left/ \!{~}_{\mathrm L}\right.$$
(1)

where, IN and TP are the available intensity m with a number of pixels and the total number of pixels. The ranges from 0 to 256 are the pixel intensities. For available intensity, the normalized histogram image bin (I) is denoted as M. Image is divided into multiple sub-regions and recovering the image contrast. Transform the specified target histogram with the intensity of all sub-regions independently to fulfill [19].

For de-noising, enhanced CT scan images are obtained by applying an adaptive bilateral filter (ABF). Equation (2) shifts the range filter on the histogram by adding a counterbalance to the range [31].

$$AB{F}_{{y}_{0},{z}_{0}}=\sum_{y={y}_{0-M}}^{{y}_{0}+M}\sum_{z={z}_{0}}^{{z}_{0}+M}\mathrm{exp}(-\frac{{(y-{y}_{0})}^{2}+{(z-{z}_{0})}^{2}}{2{\nabla }_{c}^{2}})+{exp}\ (-\frac{(H[{y}_{0}{z}_{0}]-H[{y}_{0}{z}_{0}-\chi {[{y}_{0}{z}_{0}]}^{2}])}{2{\nabla }_{e}^{2}[{y}_{0}{z}_{0}]})$$
(2)

From Eq. (2), the index column of a current image pixel and the neighboring pixel row index are defined as \({z}_{0}\) and y. The neighboring pixel column index and the neighboring pixel column index are z and y. The size of the neighboring window is \({\xi }_{{y}_{0}{z}_{0}}\) and the neighboring window pixel is M.

During the process of image acquisition, the CT images are surrounded bya few of the noises that lead to failure nodules detection. Detected abnormalities were incorrectly assumed to be cancer nodules for a while, and additional abnormalities need to be removed for a better and more accurate cancer diagnosis. With the ABF sharpening the CT scan images without causing overrun, the edge slope rises [7]. While improving textures and edges in the images, ABF is able to smooth the noise. When compared to other median and mean filters, the ABF produced good results. The ABF overcome the causes of halo artifacts and objectionable ringing based on issues in conventional filters like under and overshoots around the renders clean. The entire appearance is enhanced and the slope edge in the image is enhanced.

3.2 Lung cancer segmentation

In this study, the lung cancer segmentation is performed using a Deep CNN models-based ensemble approach with Modified Mayfly and modified particle swarm optimization algorithm. The steps involved for lung cancer segmentation is delineated as follows:

3.2.1 Ensemble deep convolutional neural network (EDCNN):

One of the widely used deep learning models is the convolutional neural network (CNN). In an input image, CNN studies various biases and weights relating to different items, allowing it to distinguish between them. The algorithm uses the number of channels, width, and height of an input image in the appearance of a pixel value matrix [29]. A network is made up of layers, and the first layer is a convolutional neural network, which feeds the input image. The convolutional kernel set may be found in the convolutional layer that covers the small image areas. Equation (3) expresses the convolution operation.

$${Y}_{c}^{d}=({Z}_{z,x}*{A}_{c}^{d})$$
(3)

From this equation, the spatial locality and the input image is expressed as z, x and \({Z}_{z,x}\). At dt hlayer, the cth convolutional kernel is \({A}_{c}^{d}\). The activation and pooling layer along with the convolutional neural network present in a CNN. For analyzing non-linear image properties, the decision function or activation function has a major role in the activation layer [1]. The image downsampling is performed via pooling layers. Equation (4) expresses the pooling operation.

$${G}_{c}={R}_{O}({Y}_{z,x}^{c})$$
(4)

Based on Eq. (4), the pooling operation type is \({R}_{0}(\cdot )\). The cth input and output feature maps are \({Y}_{z,x}^{c}\) and \({G}_{c}\). At the end of a network, attach a fully connected layer in which the M number of output classes related to the M-dimensional vector is produced.

  1. A.

    AlexNet and VGG-FaceNet:

The number of CNN parameter optimization models builds up the AlexNet model, which is the deeper of CNN. Extend the feature extraction from 5 (inLeNet) to 7 (in AlexNet) to make the application of CNN to diverse image classifications. While increasing model depth, the issues like vanishing gradient descents and overfitting arise. In the initial layers, introduce large-sized filters and minimize overfitting.

One of the effective design principles of CNN is VGG-FaceNet. Over the input matrix, place the small-sized filters and the larger-sized filters are induced. For image classification and localization, VGG-FcaeNet possesses homogeneous and feature topology. Due to the approximate 140 million parameter applications, it cause by the limitation of higher computational cost.

  1. B.

    FlowImageNet and ResNet:

There are 3 fully connected layers and 5 convolutional layers present in this model. The vanishing descent and overfitting issues are resolved via ReLU activation function. The computational complexities of CNN, AlexaNet, VGG-FaceNet, and FlowImageNet are overcome by using ResNet.

  1. III.

    Ensemble

The patterns in data are highly recognized via neural networks. These networks are sensitive to a particular training data is the drawback. Each time they are trained, the various weight sets are obtained. GASEN, simple averaging, and weighted voting are a few of the ensemble methods. The simple prediction models are surpassed via these techniques. The structure of EDCNN is depicted in Fig. 2.

Fig. 2
figure 2

Structure of EDCNN [29]

3.2.2 Modified Mayfly based particle swarm optimization (M2PSO):

This section describes the modified mayfly-based particle swarm optimization (M2PSO) algorithm.

  1. A.

    Modified Particle Swarm Optimization (MPSO) algorithm:

While modifying the personal best update, the modified PSO is the enhanced version of the particle swarm optimization (PSO) algorithm. The following section delineates the MPSO algorithm.

  1. (i)

    The number of decision variables (DV) defines the algorithm parameters. The population size (Pmax) with the upper and lower bounds are \({U}_{p}\) and \({L}_{o}\). The velocity is VE with the weight coefficient are \(w{c}_{1}\) and \(w{c}_{2}\).

  2. (ii)

    The initial value with the matrix is similar to the \(DV*{P}_{\mathrm{max}}\). The below equation expresses the vectors based on initial populations [21].

    $${Y}_{j,k}={L}_{o}+({U}_{p}-{L}_{o})*random\ [0,1]$$
    (5)
    $$Pa=\left[\begin{array}{c}{Y}_{\mathrm{1,1}}\ {Y}_{\mathrm{1,2}}\ \cdots \ {Y}_{1,DV}\\ {Y}_{\mathrm{2,1}}\ {Y}_{\mathrm{2,2}}\ \cdots \ {Y}_{2,DV}\\ \ \vdots \ \vdots \ \ddots \ \vdots \\ {Y}_{{P}_{\mathrm{max}},1}\ {Y}_{{P}_{\mathrm{max}},2}\ \cdots \ {Y}_{{P}_{\mathrm{max}},DV}\end{array}\right]$$
    (6)
  3. (iii)

    The following steps generate the new particle. Equation (7) updates the particle velocity.

    $$V{E}_{j,k}^{t+1}=\beta V{E}_{j,k}^{t}+{D}_{1}random \left[0,\ 1\right]\times \left({Bes{t}_{pa}}_{j,k}^{t}-{Y}_{j,k}^{t}\right)+{D}_{2}random\ [0,\ 1]\times ({Bes{t}_{G}}_{j,k}^{t}-{Y}_{j,k}^{t})$$
    (7)
  4. (iv)

    Eq. (8) updates the positions of the particles.

    $${Y}_{j,k}^{t+1}=V{E}_{j,k}^{t+1}+{Y}_{j,k}^{t}$$
    (8)
  5. (v)

    According to the objective functions, evaluate the new improved vectors. Accept \({Y}_{j,DV}^{t+1}\) if the fitness of \({Y}_{j,DV}^{t+1}\) is smaller to \({Bes{t}_{pa}}_{j,DV}^{t+1}\) fitness. For next-generation, the new particle replaces if the new particle is unaccepted in the particle best (\({Bes{t}_{pa}}_{j,DV}^{t+1}\)) updating step.

  1. B

    Modified mayfly Optimization (MMO) algorithm:

Based on the explanations, the Modified Mayfly Optimization (MMO)method is a fresh and new metaheuristic for solving many problems the conventional MMO method has a few issues in obtaining the optimal resolution in some circumstances. A new modification mechanism is provided to overcome this problem in this part. The global best and individual best position is \({Bes{t}_{G}}_{j}\) and \({Bes{t}_{po}}_{j}\) if the current candidate position \({Y}_{j,k}^{t}\) is overlapped. Each candidate stops flying if the preceding candidate velocities are adjacent to zero [25].

In the whale optimization algorithm, derive the updating position from the bubble net searching schemes. Based on the surrounding and moving spiral around the prey, the whale huts its prey based on the whale optimization algorithm. From the preceding time step, jth particle in the dth dimension updates its positions based on the spiral updating policy. Update the candidate position by considering \({Y}_{j}\) as the candidate's current positions.

$${Y}_{j}(t+1)=C{e}^{al}\mathrm{cos}(2\pi l)+{Y}_{jk}(t)$$
(9)
$$C=|{Y}_{jk}(t)-{Y}_{ki}(t)|$$
(10)

The spiral shape defines the constant value a. For the jth individual to the best solution, the random values among 0 and 1. For the jth individual to the best solution, the distance is denoted as C. Where, A is the probability factor in the interval 0 and 1.

  1. III.

    Modified mayfly based particle swarm optimization (M2PSO):

The following steps express the modified mayfly-based particle swarm optimization (M2PSO) algorithm.

  1. (i)

    Initialization of MPSO and MMF

  2. (ii)

    Constraints determine the best solution

  3. (iii)

    Fitness function evaluation

  4. (iv)

    The two stages of the MPSO and MMF algorithm select the best solution.

  5. (v)

    MPSO algorithm selects the BestG

Initialize the population of the MPSO algorithm. Update the particle velocity and position and generate the new particle [4]. Calculate new improved vectors depending upon the objective function. The process is repeated till the end of the termination condition is satisfied.

  1. (vi)

    At the same time, MMF algorithm selects the BestG

Initialize the population of MMF and update the position with spiral movement. Again, update the candidate positions. The procedure is continued at the end of obtaining BestG solution.

  1. (vii)

    The ideal solution is classified based on the priority of BestG

  2. (viii)

    Obtain the optimal best solution

3.2.3 Lung tumor segmentation using (EDCNN-M2PSO)

In general, the accuracy of the EDCNN model's results improves. By combining the distinct outputs of each model, simple averaging yields a single outcome. The average of the results of each model is calculated [13]. But this output has a few limitations with the inaccurate result due to computational complexity, poor segmentation efficiency and etc. Hence, we suggested the M2PSO algorithm to provide better segmentation results because this method has better exploration and exploitation ability with optimal best solution selection and so on [5]. Therefore, the combination of EDCNN- M2PSO algorithm is applied for lung tumor segmentation. The lung tumor segmentation using the proposed EDCNN-M2PSO algorithm is depicted in Fig. 3. From the pre-processed CT scan image, the proposed EDCNN-M2PSO algorithm segments the cancer nodules. The proposed algorithm segments both normal and abnormal lung tumors that mean mild and severe kinds of tumors. In order to diagnosis the disease, distinct areas are analyzed and cancer affected region is segmented from the CT images using the EDCNN-M2PSO algorithm.

Fig. 3
figure 3

Proposed EDCNN-M2PSO algorithm for lung tumor segmentation

4 Result and discussion

This section investigates the performance of the proposed EDCNN with the M2PSO algorithm for lung tumor segmentation. For 50 epochs, train the EDCNN-M2PSO model on 11 GB of RAM with NVIDIA 1080Ti GPU. The working platform of MATLAB software implements the proposed model [5]. Various performance measures and comparative analyses are taken to validate the performance of the proposed method. Table 2 delineates the parameter settings of the proposed model.

Table 2 Parameter settings

4.1 Dataset details

Under the approval of the institutional review board, Shandong Cancer Hospital Affiliated with Shandong University (SHFSU) provides the dataset image details which consists of 134 contrast-enhanced CT images. Based on pulmonary CT investigation, this study uses all CT images by utilizing a Brilliance 128i CT scanner by means of a 5-mm reconstruction interval, 64 × 0:625 mm collimation, 1.0 helical pitch, 220 mA current, and 120 kV voltage of standard clinical protocol. The scan ranges from 0.58 to 0.98 mm associated with the pixel size via imaging matrix pixels of 512 × 512. In thoracic malignancies of CT imaging, more than ten years of experience and two radiologists annotates all scans. By using Itk snap software, primary tumors on transversal plane boundaries are outlined [30]. Figure 4 explains the sample dataset image.

Fig. 4
figure 4

Sample dataset images, (a-d) Input images and (e–h) segmented images

4.2 Performance metrics

The measures like precision, sensitivity and dice similarity score (DSS), generalized loss and dice loss are quantitatively validating the accuracy of lung tumor segmentation. The ground truth and segmentation results similarities are analyzed through DSS [3, 22, 28].

$$DSS=\frac{2\ true\ positive}{false\ positive\ +2\ \ true\ positive\ +\ false\ negative}$$
(11)
$$Sensitivity =\frac{\ true\ positive}{true\ positive\ +\ false\ negative}$$
(12)
$$\mathrm{Pr}ecision=\frac{\ true\ positive}{true\ positive\ +\ false\ positive}$$
(13)

Voxel-wise overlap of each metric between the ground truth and segmentation results are analyzed via precision, recall, and DSS. Better the performance of segmentation with higher values.

The EDCNN model is optimized via the M2PSO algorithm by utilizing generalized dice loss (GDL) to minimize the data imbalance impact in the case of training.

$$Generalized\ dice\ loss\ =1-\frac{2}{K}\sum\limits_{{W}_{k}}\frac{{W}_{k}\cdot \sum\limits_{m}{P}_{mk}{Q}_{mk}}{(\sum\limits_{m}{P}_{mk}+\sum\limits_{m}{Q}_{mk})}$$
(14)

where M and k are the voxels and classes. The ground truth segmentation maps of one hot encoding are Q and the network softmax output is P. For various label set properties, the weight to provide invariance is \({W}_{k}=\frac{1}{{(\sum\limits_{m=1}^{M}{R}_{km})}^{2}}\)  

An Eq. (15) explains the dice loss;

$$Dice\ loss=1-\frac{2}{K}\sum\limits_{{W}_{k}}\frac{\sum\limits_{m}{P}_{mk}{Q}_{mk}}{(\sum\limits_{m}{P}_{mk}+\sum\limits_{m}{Q}_{mk})}$$
(15)

The invariance to the various label set properties is provided with generalized dice loss introducing weight \({W}_{k}\) compared to dice loss.

4.3 Performance analysis

The performance evaluation of both training and validation accuracy is plotted in Fig. 5. Randomly divide the dataset into two subsets namely 70% for training and 30% for validation. The number of epochs varies from 0 to 50 with respect to varying percentages of accuracy. The X-axis represents a varying number of epochs and the Y-axis signifies the varying number of accuracy. The number of epochs increases by increasing both training and validation accuracy respectively.

Fig. 5
figure 5

Training and validation accuracy evaluation

The training progress along with the DSS means on the validation set is represented in Fig. 6. The generalized Dice loss function trains all of these networks to 50 epochs. The lowest dice score similarity is obtained by using the deep learning (DL) technique. The other existing methods like DAL and R3-D are higher than the DL model. But, the proposed method has superior performance among all these other techniques such as DAL, DL, and R3-D methods respectively. The DSS of the proposed model is superior to other models and it yielded inferior performances. Table 3 expresses the result of the confusion matrix.

Fig. 6
figure 6

Training progress along with the DSS mean on the validation set

Table 3 Confusion matrix

The various loss functions in the learning process are plotted in Fig. 7. The dice loss function with the proposed EDCNN-M2PSO algorithm is examined. The number of epochs varies from 50 with respect to varying percentages of dice loss. The dice loss is superior to the generalized loss function. The pictorial representation of various loss functions with the mean DSS on the set of validation is plotted in Fig. 8. At the 10th epoch, the dice loss with respect to DSS decreases with increasing generalized loss is obtained.

Fig. 7
figure 7

Various loss functions with the learning process

Fig. 8
figure 8

Various loss functions with the mean DSS on the set of validation

Figure 9 depicts the pictorial representation of segmentation efficiency analysis. The methods such as DL [16], DAL [20], R3-D [14], HSN [9], and proposed EDCNN-M2PSO algorithms are applied to this investigation thereby obtaining varying segmentation output results. The affected regions are successfully extracted and segmented and their performance is shown by using three measures DSS, precision, and sensitivity results with different measures. From this investigation, the proposed method provided superior segmentation results than other existing methods like DL, DAL, R3-D, and HSN respectively.

Fig. 9
figure 9

Segmentation efficiency analysis

The overall segmentation efficiency based on average precision, sensitivity, and dice score similarity is represented in Fig. 10. The performance measures like dice score similarity, precision, and sensitivity validate the efficiency of lung cancer segmentation using the EDCNN-M2PSO algorithm. The existing method like DL [16], R3-D [14], and HSN [9] are used in this evaluation. From this investigation, the proposed EDCNN-M2PSO algorithm ideally demonstrated superior segmentation performances in the case of average dice score similarity, precision, and sensitivity than other existing techniques.

Fig. 10
figure 10

Overall segmentation efficiency based on average precision, sensitivity, and accuracy

5 Discussion

The digitization of healthcare services increased the usage of healthcare-related applications. The proposed model can be used for early disease diagnosis if it is implemented in healthcare-related applications. An effective lung disease prediction application can help people diagnose the abnormalities in the lung in the early stage when integrated with a conventional healthcare system. It can also help to monitor the patient's condition and give more information regarding their lung. The performance of the proposed model is evaluated in terms of different performance metrics such as accuracy, sensitivity, precision, DSS, and Dice loss. The M2PSO algorithm improves the generalization (segmentation capability) and robustness (anti-interference capability) of the EDNN model. It helps to segment the lung tumor from the CT scan by overcoming the issues associated with CT image consistency. The results show that the proposed model is effective in terms of different aspects but the different challenges faced by us while conducting this study are specified as follows. Unavailability of labeled datasets, lack of diverse datasets due to hospital unwillingness, and country-specific security legislation prevented us from acquiring high-quality data for training the model. These are the challenges that hindered us from testing our framework fully.

6 Conclusion

This study proposed an ensemble deep convolutional neural network (EDNN) based on Modified mayfly optimization and modified particle swarm optimization (M2PSO) algorithm to segment the lung tumor images. The proposed model is implemented using MATLAB software. The dataset images were collected from Shandong Cancer Hospital Affiliated with Shandong University (SHFSU). The precision, sensitivity, and dice similarity score (DSS), generalized loss, and dice loss are the performance metrics for analyzing the performance validation of proposed lung tumor segmentation. The proposed technique demonstrated higher training and validation accuracy, as well as the training progress along with the DSS, mean on the validation set. The DSS, precision, and sensitivity results of the proposed method are superior to other existing methods such as DL, DAL, R3-D, and HSN. The proposed model offers an average accuracy, sensitivity, and precision scores of 97%, 98%, and 98%. At the 10th epoch, the dice loss concerning DSS decreases while generalization loss increases. The DSS value of the proposed model is 98.6%, which is relatively higher than the existing techniques such as DL (95.2%), DAL (90%), R3-D(94%), and HSN (90%). In the future, we plan to detect lung cancer by analyzing the small and ground nodules from low-dose CT scans. We also plan to implement this model in a real-time setting in various parts of the world.