Abstract
Brain tumor segmentation is a critical task in medical image analysis. In recent years, several deep learning-based models have been developed for brain tumor segmentation using magnetic resonance imaging (MRI) data. To address challenges such as high computational time and the requirement of huge storage and resources, we proposed BrainSegNet, which is a lightweight extension of U-Net with progressively expanded neurons that require fewer weights and less memory space. Unlike other DL models, our proposed model has the simplest architecture and is more accurate than other state-of-the-art methods for brain tumor segmentation. The proposed approach was extensively analyzed using the BraTS2020 benchmark dataset for segmenting brain tumors. The experimental findings demonstrate the effectiveness of the proposed system, producing a 96.01% Dice score and 95.89% mean IoU for brain tumor segmentation from brain MRI images.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Hundreds of various forms of brain tumors afflict humans, making it one of the most fatal diseases [6]. Traditional treatments, such as surgery, chemotherapy, and radiotherapy, are mainly employed to address these conditions. Artificial intelligence (AI) [4] is gaining momentum and significance as it progresses, and can play a crucial role in brain tumor diagnosis and surgical pre-assessment procedures [30]. To assist in preoperative decisions segmenting brain tumor from medical brain image is now a cutting edge research area [21, 22, 29]. Researchers around the world are vigorously making efforts to establish different kinds of solutions [34] for early detection for better diagnosis and one of the solutions can be the usage of deep learning approaches [7, 9]. CNN-based segmentation methods have also achieved state-of-the-art performances in various tests [14, 25]. The classical DL model improves the connection weight between the original neurons while analyzing lesion features in large medical images [28]. Consequently, the standard DL model [19], such as the convolutional neural network (CNN), the entire convolution networks (FCN) [18], the U-Net network, and the attention U-Net [24], has considerably improved brain tumor segmentation performance. Inspired by the successful use of CNNs in multiple medical image analysis tasks, we propose a novel lightweight high-performance model for brain tumor segmentation based on deep learning. Our approach aims to design a simple deep learning architecture for brain tumor segmentation. However, all the architectures proposed [5, 14, 17, 20, 23, 25, 33] for brain tumor segmentation have a complex structure that requires high computational time, more effective GPUs, large amounts of storage, and many resources.
Bare in mind these issues, we do straightforward modification of the U-Net developed by merging merits of convolutional blocks, progressive expansion layers, and residual network connections to obtain not only a better segmentation model but also reduce the parameters. The motivation behind this is to improve accuracy and, at the same time, maintain simplicity. The goal of this study is to construct a less complex CNN architecture for brain tumor segmentation, and our main contributions include the following:
-
1.
We propose a simple U-Net architecture incorporating with progressive expansion layer, and residual network connection, referred to as BrainSegNet, that requires fewer weights and less memory space for brain tumor segmentation;
-
2.
An extensive evaluation is done on two challenging datasets, then compared with the state-of-the-art architectures, and the evaluations show that the BrainSegNet results in outstanding performance compared to baselines.
The rest of this paper is organized as follows. A literature review of recent methodology 2. In Sect. 3, a description of the proposed system along with its data collection. Section 4.2 presents the experimental results and a comparative performance analysis are presented in Sect. 4.2.
2 Related Work
Manual tumor segmentation is subjective and time-consuming. In addition, the detection result is laborious and depends on the doctor’s theoretical knowledge and practical ability. Therefore, designing an automatic and robust brain tumor segmentation system is essential to realize tumor diagnosis [1, 8, 10]. Various research studies have been suggested for segmenting brain tumors with deep learning methods. For example, Pius Kwao Gadosey et al. [1] proposed a lightweight model modifying the U-Net model named SD-UNet to obtain higher performance with fewer computational requirements. To reduce the computational complexity, they applied depthwise separable convolutions in their proposed network and achieved a 90.70% dice score for the brain segmentation task. However, their suggested architecture has 3.9M parameters, requiring 15.8 MB of disk space which can be further reduced. In the same year, Sergio Pereira et al. [25] presented a cascade-CNN system where they applied a 3\(\,\times \,\)3 kernel to mitigate overfitting. Their approach could segment the brain MRI image into four areas: normal tissue, necrosis, edema, and enhancing tumor. According to the authors, two CNN architectures were used for better feature extraction, especially for low-grade and high-grade glioma. However, one drawback is that the operator needs manually determine the glioma grade in the early stage, which demands prior medical expertise. Besides these issues, the introduced system showed poor segmentation results for core areas in the BRATS 2015 dataset. Similarly, Havaei et al. [12] proposed another novel Cascading style to segment brain tumors automatically. This research utilized the cascade style of CNN to capture global and local contextual features to deal with imbalanced tumor labels. Moeskops et al. [21] suggested an automatic system incorporating a multi-scale CNN to segment white matter hyperintensities of presumed vascular origin (WMH) from MRI modalities. Chen et al. [5] introduced a model combing prior knowledge with DCNN to improve the captured motifs of DCNN for brain tumor sub-compartment identification.
In 2020, Zhang et al. [31] introduced an attention U-Net for the purpose of brain tumor segmentation. Their suggested architecture improved the local responses to down-sampling and the recovery effects of the up-sampling simultaneously. Furthermore, Wu et al. [29] introduced a CNN-based multi-features refinement and aggregation network (MRANet) for brain tumor segmentation. They adopted the feature fusion strategy to utilize hierarchical features in this system. Recently, Lee et al. [16] proposed an intelligent brain tissue segmentation method utilizing MR images. They partitioned the brain MRI image into small patches through a patch-wise U-Net model, and each patch was predicted separately. However, although the introduced model could overcome the limited disk space problem, the training stage showed higher computation complexity.
3 Research Methodology
3.1 U-Net
U-Net architecture [27] is one of the most popular deep-gaining knowledge of architectures used for medical image segmentation. This architecture is on the whole utilized through the research community due to the fact of its architectural variations with superior outcomes in the image segmentation tasks. In clinical image segmentation U-Net has established itself as well-known for recent proposed ideas including R2YU-Net, attention U-Net, and U-Net++. The place u-shape model is retained along with the two fundamental paths encoder and decoder. By incorporating more than a few strategies such as skip connections, attention modules, residual blocks or aggregate of them those u-net versions alter the encoder or the decoder blocks.
3.2 dPEN: deep Progressively Expanded Network
The dPEN method takes a pixel from a multispectral image as an input vector and returns the types of objects present in the input pixel vector. Figure 1 illustrates the structure of dPEN, which is formulated as follows: four progressive expansion layers (PEL), three convolutional, three maximum-pooling, and then a fully connected layer (FC). Progressively expandable neural networks (PEN nets), developed for identifying hyperspectral images, serve as the basis of dPEN. The core idea of the dPEN paradigm is the addition of PEL, which incrementally expands each node from the input using a McClorin series expansion of a predefined nonlinear function. It is shown that the dPEN architecture is capable of extracting significant discriminative features from multispectral bands. Hence, the dPEN architecture outperforms well-known machine learning models such as random forest, support vector machine, CNN, etc.
3.3 Proposed Architecture: BrainSegNet
In this study, we propose a modified U-Net named BrainSegnet for brain tumor segmentation. BrainSegNet combines U-nets’ modularity with the integration of expanded neurons using nonlinear function expansion. Our proposed architecture takes a 224 \(\times \) 224 input image as input and outputs the segmented points of interest shown in Fig. 2. The U-Net’s U-shape is preserved, with the encoder path on the left side and the decoder path on the right. We incorporate PEL in the encoder path only. A PEL layer, also known as PE2: progressive expansion with the first two terms in Maclaurin series expansion, is included in each residual block along with two batch normalization layers (BN), two ReLU activation layers (ReLU), two 3 * 3 2D convolutional layers (Conv 3 * 3), and a PEL layer. The enlarged area of Fig. 2 demonstrates the arrangement of the Residual PE block.
4 Dataset and Results Evaluations
In the experiments, the training, testing, and validation datasets are all from the BraTS2020 benchmarks [2, 3]. The MICCAI Brain Tumor Segmentation competition uses BraTS, a significant public dataset for multimodal brain tumor segmentation, and it is frequently utilized in research on this subject. Every year, new data are added, removed, or replaced to the dataset to enrich its scale. BraTS2020 have 369 annotated brain tumor samples for model training and 120 samples for testing. Each case includes MRI scans of four different modalities: T1-weighted (T1), T1-enhanced contrast (T1-ce), T2-weighted (T2), and T2 fluid-attenuated inversion recovery (Flair). The datasets are labeled by domain experts and contain four classes: background, necrotic and non-enhancing tumors (NCR/NET), GB-peritumoral edema (ED), and GB-enhancing tumors (ET). The evaluation is based on three different brain tumor regions:
-
Whole Tumor (WT) = NCR/NET + ED + ET
-
Tumor Core (TC) = NCR/NET + ET
-
Enhancing Tumor (ET)
4.1 Evaluation Metrics
To verify the efficiency of the proposed model framework, two most common metrics used in medical image segmentation, mean IoU score and Dice score are used. The corresponding formulas are given:
where \(\tau \rho \), \(f\rho \), \(\tau \eta \), and \(f\eta \) refer to true positive, false positive, true negative, and false negative respectively.
4.2 Evaluation Results
It is essential to note that while all images are generated with 240 \(\times \) 240 resolution for training, all datasets are evaluated at native resolution. No pre-processing or post-processing is done. We also train U-Net, ResU-Net, and Trans-Unet and keep all the settings the same.
The training and validation accuracy and loss curves are shown in Fig. 3 and Fig. 4, respectively. These Figures show that our proposed BrainSegnet and ResUnet showed almost similar accuracy and loss, while TransU-Net depicted poor performances in both cases. In BraTS2020 dataset, BrainSegNet obtained a dice score of 97.47%, 82.33%, and 90.33% for WT, ET, and TC, respectively, and the average dice score achieved 96.01%. For the sake of comparison, we also trained some state-of-the-art models with the same setup and data distribution and the experimental results are shown in Table 1 in terms of Dice score.
Generally, BrainSegNet performs well in ET, TC, and WT categories. Table 1 demonstrates that the Dice score of the BrainSegNet has a significant advantage over the conventional U-Net with less complexity. The findings indicate that the approach can segment tumors more effectively than U-Net and ResU-Net. BrainSegNet is more conducive to learning long-distance artifacts, so the effect of TC and WT category segmentation is more prominent. For ET cases, ResU-Net and Typical U-Net obtain poor results, which indicates that for brain tumor segmentation, these models cannot extract sufficient local features. However, for ET case, compared to U-Net and ResU-NeT, it performs poorly. On the other hand, Trans-Net obtains an 89.83% dice score. BrainSegNet performs exceptionally well in the ET and TC categories, demonstrating the method’s superior capacity to extract both global and local features.
In Table 1, the data in brackets shows the standard deviation of the proposed model segmentation outcomes. According to the table, the proposed model has a minimum classification deviation for each of the three groups. These findings suggest that the model has the best and most consistent segmentation impact and that the segmentation outcomes will not have noticeable differences. Furthermore, we can see from Table 2 that compared with U-Net and ResUet, BrainSegNet can gain 0.69% and 0.36% higher \(\mu IoU\) result and also has a 10.06% improvement compared to the TransU-Net model. Our model also has the minimum standard deviation among all the models mentioned in Table 2, which indicates the stability of the segmentation results of the BrainSegNet model.
Computational and Memory Requirements: The less complex methods suggested in this study are more suited for resource-constrained applications, both regarding training and deployment, for example, in portable devices. Table 3 compares the proposed architecture with three well-known architectures regarding performance vs. the volume of parameters and memory space. We can observe that our proposed BrainSegNet achieves impressive results and has the lowest memory space (18 MB) and almost \(\sim \)4.8 million trainable parameters.
Visualization of Tumor Segmentation Result: In Fig. 5, a visualization of the brain tumor segmentation results is made for the BrainSegNet method. The original flair images, ground truth, all class predicted, necrotic/core, edema, and enhancing tumor are shown in the first to sixth columns, respectively. Figure 5 illustrates that the segmentation result for the WT area is the best for all the models, and the segmentation results for the two complicated edges of ET and TC are pretty different. Our model shows more accurate outcomes compared with Ground Truth for detail segmentation.
4.3 Comparison with Existing State-of-the-Art
Table 4 presents the comparative insights of few recent similar research based on DL methods. Based on this performance metrics in the comparative analysis, BrainSegNet method is the most suitable for a image segmentation of Brain tumor.
5 Conclusion
In this study, we proposed a novel segmentation system, a new variant of the U-Net architecture named BrainSegNet, which can automatically segment diseased tissues in brain MRI images. The model efficiently combines U-Net with PEN to accurately and efficiently segment tumors from brain MRI. Experimental outcomes demonstrate that our proposed system performs better in brain tumor MRI image segmentation than state-of-the-art methods (such as U-Net, ResU-Net, and TransBTS). Furthermore, the visualization results show that the proposed system has a good segmentation performance for all three lesion regions of brain tumors. Comparisons with recent research on brain tumor area segmentation also demonstrate that our method achieves promising results on BraTs2020 datasets, indicating its potential for practical applications in auxiliary diagnostic procedures.
Further research may analyze the efficacy of additional distinctive features for segmenting brain tumors and explore the suggested method for other semantic segmentation issues.
References
Akbar, A.S., Fatichah, C., Suciati, N.: SDA-UNET2. 5D: shallow dilated with attention Unet2. 5D for brain tumor segmentation. Int. J. Intell. Eng. Syst. 15(2) (2022)
Bakas, S., et al.: Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci. Data 4(1), 1–13 (2017)
Bakas, S., et al.: Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the brats challenge. arXiv preprint arXiv:1811.02629 (2018)
Biswas, M., Kaiser, M.S., Mahmud, M., Al Mamun, S., Hossain, M.S., Rahman, M.A.: An XAI based autism detection: the context behind the detection. In: Mahmud, M., Kaiser, M.S., Vassanelli, S., Dai, Q., Zhong, N. (eds.) BI 2021. LNCS (LNAI), vol. 12960, pp. 448–459. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-86993-9_40
Chen, H., Qin, Z., Ding, Y., Tian, L., Qin, Z.: Brain tumor segmentation with deep convolutional symmetric neural network. Neurocomputing 392, 305–313 (2020)
Das, S., Nayak, G.K., Saba, L., Kalra, M., Suri, J.S., Saxena, S.: An artificial intelligence framework and its bias for brain tumor segmentation: a narrative review. Comput. Biol. Med. 143, 105273 (2022)
Ghose, P., et al.: Detecting COVID-19 infection status from chest X-ray and CT scan via single transfer learning-driven approach. Front. Genet. 13, 980338 (2022)
Ghose, P., Sharmin, S., Gaur, L., Zhao, Z.: Grid-search integrated optimized support vector machine model for breast cancer detection. In: 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 2846–2852. IEEE (2022)
Ghose, P., Uddin, M.A., Acharjee, U.K., Sharmin, S.: Deep viewing for the identification of COVID-19 infection status from chest X-ray image using CNN based architecture. Intell. Syst. Appl. 16, 200130 (2022)
Ghose, P., Uddin, M.A., Islam, M.M., Islam, M., Acharjee, U.K.: A breast cancer detection model using a tuned svm classifier. In: 2022 25th International Conference on Computer and Information Technology (ICCIT), pp. 102–107. IEEE (2022)
Guan, X., et al.: 3D AGSE-VNET: an automatic brain tumor MRI data segmentation framework. BMC Med. Imaging 22(1), 1–18 (2022)
Havaei, M., et al.: Brain tumor segmentation with deep neural networks. Med. Image Anal. 35, 18–31 (2017)
Huang, H., et al.: A deep multi-task learning framework for brain tumor segmentation. Front. Oncol. 11, 690244 (2021)
Iqbal, S., Ghani, M.U., Saba, T., Rehman, A.: Brain tumor segmentation in multi-spectral MRI using convolutional neural networks (CNN). Microsc. Res. Tech. 81(4), 419–427 (2018)
Jiang, Y., Ye, M., Wang, P., Huang, D., Lu, X.: MRF-IUNET: a multiresolution fusion brain tumor segmentation network based on improved inception U-net. Comput. Math. Methods Med. 2022 (2022)
Lee, B., Yamanakkanavar, N., Malik, M.A., Choi, J.Y.: Correction: automatic segmentation of brain MRI using a novel patch-wise u-net deep architecture. PLoS ONE 17(2), e0264231 (2022)
Liu, X., Zhang, L., Li, T., Wang, D., Wang, Z.: Dual attention guided multi-scale CNN for fine-grained image classification. Inf. Sci. 573, 37–45 (2021)
Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431–3440 (2015)
Mahbub, M.K., Zamil, M.Z.H., Miah, M.A.M., Ghose, P., Biswas, M., Santosh, K.: Mobapp4infectiousdisease: classify COVID-19, pneumonia, and tuberculosis. In: 2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS), pp. 119–124. IEEE (2022)
Maqsood, S., Damasevicius, R., Shah, F.M.: An efficient approach for the detection of brain tumor using fuzzy logic and U-NET CNN classification. In: Gervasi, O., et al. (eds.) ICCSA 2021. LNCS, vol. 12953, pp. 105–118. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-86976-2_8
Moeskops, P., et al.: Evaluation of a deep learning approach for the segmentation of brain tissues and white matter hyperintensities of presumed vascular origin in MRI. NeuroImage Clin. 17, 251–262 (2018)
Mohammed, Y.M., El Garouani, S., Jellouli, I.: A survey of methods for brain tumor segmentation-based MRI images. J. Comput. Design Eng. 10(1), 266–293 (2023)
Nguyen, D.H., et al.: ASMCNN: an efficient brain extraction using active shape model and convolutional neural networks. Inf. Sci. 591, 25–48 (2022)
Oktay, O., et al.: Attention U-net: learning where to look for the pancreas. arXiv preprint arXiv:1804.03999 (2018)
Pereira, S., Pinto, A., Alves, V., Silva, C.A.: Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Trans. Med. Imaging 35(5), 1240–1251 (2016)
Ranjbarzadeh, R., Bagherian Kasgari, A., Jafarzadeh Ghoushchi, S., Anari, S., Naseri, M., Bendechache, M.: Brain tumor segmentation based on deep learning and an attention mechanism using MRI multi-modalities brain images. Sci. Rep. 11(1), 10930 (2021)
Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015, Part III. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Wei, D., Zhang, L., Wu, Z., Cao, X., Li, G., Shen, D., Wang, Q.: Deep morphological simplification network (MS-net) for guided registration of brain magnetic resonance images. Pattern Recogn. 100, 107171 (2020)
Wu, D., Ding, Y., Zhang, M., Yang, Q., Qin, Z.: Multi-features refinement and aggregation for medical brain segmentation. IEEE Access 8, 57483–57496 (2020)
Zhang, J., Jiang, Z., Dong, J., Hou, Y., Liu, B.: Attention gate resU-Net for automatic MRI brain tumor segmentation. IEEE Access 8, 58533–58545 (2020)
Zhang, J., Lv, X., Zhang, H., Liu, B.: AResU-Net: attention residual u-net for brain tumor segmentation. Symmetry 12(5), 721 (2020)
Zhang, W., et al.: ME-Net: multi-encoder net framework for brain tumor segmentation. Int. J. Imaging Syst. Technol. 31(4), 1834–1848 (2021)
Zhao, X., Wu, Y., Song, G., Li, Z., Zhang, Y., Fan, Y.: A deep learning model integrating FCNNs and CRFs for brain tumor segmentation. Med. Image Anal. 43, 98–111 (2018)
Zhu, Z., He, X., Qi, G., Li, Y., Cong, B., Liu, Y.: Brain tumor segmentation based on the fusion of deep semantics and edge information in multimodal MRI. Inf. Fusion 91, 376–387 (2023)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ghose, P., Biswas, M., Gaur, L. (2023). BrainSegNeT: A Lightweight Brain Tumor Segmentation Model Based on U-Net and Progressive Neuron Expansion. In: Liu, F., Zhang, Y., Kuai, H., Stephen, E.P., Wang, H. (eds) Brain Informatics. BI 2023. Lecture Notes in Computer Science(), vol 13974. Springer, Cham. https://doi.org/10.1007/978-3-031-43075-6_22
Download citation
DOI: https://doi.org/10.1007/978-3-031-43075-6_22
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-43074-9
Online ISBN: 978-3-031-43075-6
eBook Packages: Computer ScienceComputer Science (R0)