1 Introduction

Globally, cancer is the second leading cause of deaths specially in middle and low-income countries. In the year 2020, it was accounted 10 million deathsFootnote 1, which are approximately 1 in 6 of all the global deaths [8]. Cancer is the abnormal growth of cells that can spread rapidly to any body organ [8]. These cancerous cells due to abnormal growth combined at any location in the body organ or collection of abnormal cells in body organ become tumor. These tumors affect body organ very badly and same in the case with brain tumor. It can affect almost anybody organ, which has many anatomic and molecular subtypes. Accurate detection of cancerous cells or tumor is central to effective disease prognosis. Early detection can significantly increase the chances of successful treatment.

Various techniques working for segmenting the tumor from the body organ. On other side medical imaging is playing an important role for cancerous cells diagnosis and tumor segmentation. In the medical analysis the most common modalities [7] nowadays are Magnetic Resonance Images (MRI), Computer Tomography (CT Scan), Ultrasound [9] using on large scale, specifically for brain tumor detection, classification and grading. MRI is preferred over other modalities, because of its non-invasiveness property and using radio waves that didn’t affect the sensitive body organs like brain. MRI has different sequences like T1-w, T2-w, FLAIR, T1c, PD, T1Ce, and DTI [11]. Can visualize the structure of the brain in all possible planes such as axial views, sagittal views, and coronal views [12], where we can examine the brain in 3D view.

Three different types of brain tumor segmentations are performing in the clinical environment, i.e., manual, semi-automated, and fully- automated segmentation. Manual segmentation although more reliable because radiologist examine themselves, however, it requires major human involvement to identify the tumor regions in brain MRI [3]. Manual method takes time to analyze the large volume patients and somehow found human errors during the examination due to negligence or time factors, that’s why it is the best approach to use in practical application only for few patients. Semi-automatic method required initial input to start the process and then system continue to segment out from the MRI images. Whereas, fully automatic require minimum input to perform segmentation, and there is no need for any human efforts. Using this approach, we can easily examine large volume patients in very short time as compared to other two mentioned approaches.

Ramy A. Zeineldin et al. [29] proposed a neural network for brain tumor segmentation, where this study address the crucial problem to distinguish tumor boundaries with healthy cells. It is based on decoupling modular framework, which has two important parts, which consists of the encoder and decoder relationship. To extract the spatial information for the images used convolutional neural network (CNN). To get the complete resolution map, the resulting semantic map pushed towards the decoder. This framework has been introduced on the basis of modified U-net architecture, and other CNN models including ResNet, DensNet, and NASNet. MICCAI BARTS 2019 dataset has been used for evaluation and Dice Coefficient & Hausdorff values are used as standard evaluation measures. The evaluation obtained for this research study in Dice and Hausdorf are 0.81 to 0.84 and 9.8 to 19.7 respectively. LING TAN et al. [27] proposed model for the segmentation of tumor in MRI, which was based on the ACU-Net network. They choose the deep separable convolutional layer for distinguishing the spatial correlation and appearance correlation of the mapped convolutional channel. Further, they replaced the ordinary architecture in the U-Net. The propagation capacity of the features was highlighted in the ACU-Net network by introducing the residual skip connection. To identify the image contour and edge cracks, the active contour model was used for the tumor segmentation.

Early tumor detection and segmentation is the need of time because it helps in early diagnosis, and treatment planning. Tiejun Yang et al. [28] highlight the brain tumor crucial problem such that structural diversity, and low segmentation accuracy. An automatic brain tumor segmentation method proposed with SK-TPCNN (Small Kernels Two Path Convolutional Neural Network) and RF (Random Forest) algorithm for classification. Feature extraction techniques and joint optimization techniques are introduced clearly in this research study. Basically SK-TPCNN joins the small and large convolutional kernels to safe from over fitting and to enhance the non-linear ability. At the end features that are learned pass to the RF classifier which classify between normal and tumorous images. The proposed model evaluated using BRATS 2015 training dataset and obtained 0.86 DSC values in whole tumor or edema.

Kadkhodaei et al. [15] proposed a technique to address the brain tumor segmentation issues in variation of tumor size, shape and appearance properties. They proposed the technique in which they map all MRI images on the same scale. After scaling, they enhanced and normalize the dataset, enhanced images were then segment out using 3D super voxel approach including saliency algorithm and edge-aware filtering technique. Shashwat et al. proposed a segmentation method using the Fuzzy C-Means algorithm, employing extraction of tumor core using area and circulatory as criteria. A method was implemented by Nabizadeh et al. [20] in which they performed the automated segmentation for single spectrum MRI using the region-based segmentation approach. The whole research study covers in these aspects first is to extract the features to highlight the image textures, second proposed mechanism to reduce the less important features, third apply machine learning algorithms to classify images that have tumorous cells in them. To find the tumor contours they used Skippy greedy snake algorithm which helps to obtain the best results in accuracy are 93.8%. Integrating the dual-tree complex wavelet transform, a fully automated unsupervised brain MRI segmentation method was introduced by Jingdan Zhang et al. [30]. The research study used DT-CWT (Dual Tree Complex Wavelet Transform) and K-means algorithm for brain MRI segmentation and they did the qualitative analysis. Clinical analysis for the brain tumor segmentation is still challenging task, due to the ambiguous tumor boundaries, Chaiyanan et al. [26] proposed a novel method in which a gray level co-occurrence matrix (GLCM) based on cellular automata was used for brain tumor segmentation. They introduced improved tumor cut (ITC) algorithm to get the high performance. State-of-the-art ITC and GLCM used for the segmentation and proposed model were evaluated using BRATS 2013 dataset with 0.84 DSC value.

Li, Yuhong et al. [16] proposed a technique based on the probabilistic model for the segmentation of multimodal brain MRI images. The solution for the spatial and structural variability in the segmentation problem in the referred technique combines two algorithms, i.e., Sparse Representation and Markov Random Forest. Nabizadeh et al. [19] proposed a fully automatic brain tumor segmentation method on MRI. The efficacy of the statistical features is evaluating the Gabor wavelet feature. Tom Haeck et al. [13] proposed a fully-automated method and this was the generative brain tumor segmentation method by Expectation-Maximization Approach. Soltaninejad Mohammadreza et al. [25] proposed the method using the superpixel-based extremely randomized trees. Extract texture based and structural based features. Superpixel-based segmentation applied and segment out the image into the number of patches. Support Vector Machine (SVM) and Extremely Randomized Tree (ERT) classifier used for classification between the normal and abnormal superpixels. The proposed model evaluated using BRATS 2012 dataset with SVM and ERT classifier, that obtained DSC 0.83 and 0.88 respectively. Rehman Zaka et al. [22] suggested superpixel based segmentation using statistical features, texton features based histogram, and fractal features. Feature selection and normalization has been applied using principal component analysis and independent component analysis. The proposed model evaluated by BRAST 2012 dataset and obtained 0.91 DSC.

Existing techniques are facing problems due to variation in tumor size, diversity in tumor shape and appearance. Recently, proposed solutions are discussed above which used statistical, textual, and curvature type features in tumor segmentation. Few of them used pixel-based, superpixel-based and region-based segmentation. Recent research methods in deep learning approaches to solve these tumor segmentation problems are also mentioned. The proposed model introduced an automatic technique to find the accurate tumorous area from brain MRI using superpixel-based segmentation with improved parameter values in superpixel segmentation. Different types of features such that statistical features, texton (Gabor) features, fractal features, curvature feature, and SIFT features have been extracted to get tumorous cells information. The reported approach also introduced a ‘class balancing algorithm’ to balance the feature vector obtained after feature extraction. By using this proposed algorithm enabled the model to have more accurate results. An ensemble technique has been introduced to get more accurate tumor area where we used SVM, Decision Tree, KNN and Ensemble classifiers to classify the superpixels based on the extracted feature values. DSC used as evaluation measure on the BRATS 2017 training dataset.

The proposed model has been divided into five main sections: the first section discussed brain MRI dataset with pre-processing techniques, the second section highlights the superpixel-based segmentation, the third section explain the feature extraction approaches, fourth section elaborate class balancing algorithm, and in fifth section classification between the normal and abnormal superpixels has been discussed.

2 The proposed methodology

The proposed model in an automatic detects and segment brain tumor abnormal tissues as shown in (Fig. 1) Proposed methodology has five main sections which are clearly defined below.

Fig. 1
figure 1

The proposed model flow

2.1 Preprocessing

Proposed model acquired open-source dataset known MICCAI BRATS 2017. It has both type of LGG and HGG volumes including T1-w, T2-w, T1c and FLAIR sequences and axial views. This training dataset total slices in one volume are 155 but few slices from start and few slices at the end are not have sufficient information. All these slices, which have lack of information are discarded by taking the mean value for all the slices in the dataset. All 3D slices have been converted into 2D format to feed into the model. Because the proposed model worked for the 2D slices.

2.2 Superpixel-Based segmentation

Simple Linear Iterative Clustering (SLIC) [1] algorithm is used to partition the images into approximately the equal size of patches. SLIC method has some parameters used to tune its algorithm to have more favorable results of segmentation. The boundary of the superpixel is defined by the spatial distance and the intensity distance. Moreover, it has very low computational cost and is an efficient in memory usage perspectives. In the proposed model, we choose the FLAIR sequence for evaluation of the proposed model. Every given slice is gridded into the approximately equal side size ‘R’, defined by the user. We represent grid size ‘R’ for these superpixels, and the geometrically center of the superpixel is also the center of the superpixel itself. On each iteration, these coordinates are updated based on intensity distance and spatial distance. Another parameter ‘r’ is there which known as the regularizer coefficient. The flexibility in all the superpixel boundaries is determined by regularizer r. More compact segments create when the regularizer r value high, and the segments create more flexible boundaries of the superpixels when the regularizer value low. To obtain the normalized intensity values the range is defined [25]. The detailed about the parameters tuned are discussed in the superpixel parameters section.

2.3 Feature extraction

We have extracted possible features in the proposed model using different image processing techniques. All the extracted features are mentioned in below sub-sections.

2.3.1 Intensity-based statistical features

The proposed model used grayscale images [24] and the intensity-based features [25] are extracted. In the grayscale images, the most important is a grey level which is distributed with different values in the selected region and these selected regions are known as superpixels. The details are discussed in the superpixel-based segmentation section. There is total 16 statistical features which are calculated listed: Angular second moment, Contrast, Correlation, Variance, Homogeneity, Sum Average, Sum Variance, Sum Energy, Entropy, Difference Variance, Difference Entropy, Information Measure of Correlation (I&II), Maximal Correlation Coefficient, Standard Deviation and Average.

2.3.2 Fractal features

Segmentation-based Fractal Texture Analysis (SFTA) [6] algorithm used to extract the fractal features, which is segmentation-based on the texture analysis. Otsu algorithm used to decompose the images into a set of binary images. These images (patches) are converted into binary images by multilevel thresholding. The thresholds represent by nt and the value of nt is defined as user required. In the proposed model the threshold values are set to nt =2. All the image boundaries extracted by using edge detection [4] for each binary channel. The fractal feature extracts the area, intensity and fractal dimension from all these binary edge channels. To represent the number of all possible edge pixels in the selected superpixel which are known as area feature. As the mean intensity of the given image pixels which correspond to the edge pixels in the superpixel are known as intensity feature. The complexity in the structure of the image, which is calculated from the boundaries of the image is known as the fractal dimension, and it is calculated as shown in the Eq. (1).

$${D}_{0}=\underset{\epsilon \to 0}{\text{lim}}\frac{\text{log}N \left(\epsilon \right)}{\text{log}{\epsilon }^{-1}}$$
(1)

Where N (\(\epsilon\)) represents the counting of hypercube of dimension E and length \(\epsilon\). By using the box-counting algorithm [23] an approximation of the fractal distance is obtained from the binary images. Figure 2 represents the flowchart for the extraction of fractal analysis only.

Fig. 2
figure 2

Fractal features extraction flowchart

2.3.3 Texton features

The brain tissues are complex and have a complicated structure. Texture features are extracted to increase the accuracy of brain tumor segmentation. A specific filter bank produces texton after the convolution of the image, these textons are small elements of an image. Gabor filters [14] are used and discussed in Eq. (2).

$$G\left(x,y;\theta ,\sigma ,\lambda , \psi ,\gamma \right)=exp\left(\frac{{\stackrel{\prime}{x}}^{2}+{\gamma }^{2}{\stackrel{\prime}{y}}^{2}}{{2\delta }^{2}}\right)exp\left(i\left(2\pi \frac{\stackrel{\prime}{x}}{\lambda }+\psi \right)\right)$$
(2)

Where \(\sigma\) is represented as the filter size, while the wavelength of the sinusoid is represented by \(\lambda\), phase as the shift is denoted as \(\psi\) and spatial aspect ratio denoted as \(\gamma\). The flow for texton features shown in Fig. 3.

Fig. 3
figure 3

Gabor filters applied to extract the texton features

2.3.4 Curvature feature

The image shape-based features are the curvature features and these features are calculated by taking the gradients along with the given directions such that x and y-direction are nominated by fx and fy. the most important is that image normal also calculated at the pixel (x, y) represent in the Eq. (3)Footnote 2.

$$\widehat N\left(x,y\right)=\frac1{\left({f_x^2+f}_y^2\right)^{1/2}}\begin{pmatrix}f_x\\ f_y\end{pmatrix}$$
(3)

The divergence of the normal is the curvature of the 2D images and it is computed as shown in the Eq. (4).

$$Curv=\frac{{f}_{xx}{f}_{y}+{f}_{yy}{f}_{x}-2{f}_{xx}{{f}_{x}f}_{y}}{{\left({{f}_{x}^{2}+f}_{y}^{2}\right)}^{3/2}}$$
(4)

Where fxx and fyy are represented as the second derivatives and I (x, y) is denoted as the intensity of the image. When we take an average of all the curvature values for all those pixels in the superpixel is known as the curvature of each superpixel.

2.3.5 SIFT features

Scale-Invariant Feature Transform (SIFT) has two main parts one is to detect key points and another one is the keypoint descriptor. The detailed description of these paradigms is discussed.

For the SIFT features the first step is keypoint detection. To compute the SIFT features it’s necessary to detect the key points which are calculated by the Degree of Gradient (DoG) and the image D (x, y, σ) after that L (x, y, σ) which represents the difference of smoothed images. Convolution applied of the given variable scale Gaussian with the given input image I (x, y) and resulted in L (x, y, σ) as represent in the Eqs. (5) and (6).

$$D\left(x,y,\sigma\right)=L\left(x,y,k\;\sigma\right)-L(x,y,\sigma)$$
(5)
$$L\left(x,y,\sigma\right)=\frac1{2\pi\sigma^2}exp\left(-\frac{x^2+y^2}{2\sigma^2}\right)\ast I\left(x,y\right)$$
(6)

There are many DoG images obtained and it’s performed between the different scales. Another aspect is that the local extremes which are detected from these DoG images by comparing 26 neighborhoods of a pixel within a set of three DoG images, and the pixel is represented as the key point if it is extremum [17].

The second point in the SIFT features is key point descriptor which is very important and gradient magnitude denoted as m (x, y) and θ (x, y) represents the orientation at each pixel in the smoothed image and the calculation shown in the Eqs. (7, 8 and 9) for key points:

$$m\left(x,y\right)=\sqrt{{f}_{x}{\left(x,y\right)}^{2}+{f}_{y}{\left(x,y\right)}^{2}}$$
(7)
$$\theta \left(x,y\right)={\text{tan}}^{-1}\frac{{f}_{y}\left(x,y\right)}{{f}_{x}\left(x,y\right)}$$
(8)
$$\left\{\begin{array}{*{20}c}{f}_{x}\left(x,y\right)=L\left(x+1,y\right)-L\left(x-1,y\right)\\ {f}_{y}\left(x,y\right)=L\left(x,y+1\right)-L\left(x,y-1\right)\end{array}\right.$$
(9)

After that by the gradient magnitude, the weighted histogram of 36 directions is made and around the keypoint orientation in the region around, and the peak which assumed to be 80% and sometime it may be more for the maximum value of the histogram is assumed to be the orientation of the key point. And after giving some rotation to the given region around the key point to the orientation, at that point, the descriptor will be created.

2.4 Class balancing algorithm

A feature vector obtained after the extraction of features for each volume. These feature vectors are unbalanced like a large number of normal superpixels and very fewer amounts of abnormal. It needs to balance the feature vector. The proposed model introduced an algorithm to solve this problem, as shown below. The proposed model applies this algorithm and gets better results for classification. The proposed algorithm as shown as Algorithm 1, taking unbalanced feature vector ‘F’ as input, after applying the random function on them and give output vector Ω. The output vector is used for the classification between normal and abnormal superpixels.

figure a

2.5 Superpixel classification

In this phase we train the model using different classifiers such that SVM, KNN, Decision Tree and Ensemble classifier. After getting the trained model we classify the superpixels for each slice and these based on the superpixel feature values which obtained after feature extraction. Based on the feature values of each superpixel classified either normal or abnormal superpixel. The abnormal superpixel is basically tumorous area in the slice.

2.6 Evaluation measures

It is necessary to evaluate the developed model how well it’s performed. The evaluation measures which we used to evaluate the proposed model performance is Dice Coefficient (DSC), Precision, Sensitivity and BER. The DSC evaluation parameter used for the evaluation of medical images, DSC is an overlap ratio between the segmented images from the proposed model and ground truths, the equation for DSC is shown in (10).

$$\text{Dice Coefficient}=\frac{2|\text{X}\cap \text{Y}|}{\left|\text{X}\right|+\left|\text{Y}\right|}$$
(10)

Sensitivity is defined as the number of all true positives, is the percentage of correctly classified positive examples among all examples, the truly signify the tumor pixels, sensitivity explained in Eq. (11).

$$\text{Sensitivity}=\frac{{N}_{TP}}{{N}_{TP+}{N}_{TN}}\times 100\%$$
(11)

Precision is the most common evaluation measure used, precision is the ratio of the brain pixels and the sum of truly represent the pixels of the brain and the falsely identified pixels of the brain. The formula to compute precision is shown in Eq. (12).

$$\text{Precision}=\frac{{N}_{TP}}{{N}_{TP+}{N}_{FP}}\times 100\%$$
(12)

Balanced Error Rate (BER) is an evaluation measure. BER used to check an average of the errors on each class, the formula for BER is shown in Eq. (13).

$$\text{BER}=0.5\left[\frac{{N}_{FP}}{{(N}_{TN+}{N}_{FP})}+\frac{{N}_{FN}}{{(N}_{FN+}{N}_{TP})}\right]$$
(13)

We used all these evaluation measures mentioned above to check the proposed model performance against the research problem statement.

3 Results and discussion

Experimental results of the proposed model are discussed in this section. The proposed model used MICCAI BRATS-2017 training datasetFootnote 3 [18] for testing and training purposes. There are some superpixel based segmentation parameters which we tuned for the experimental analysis. In superpixels the region side size ‘R’ controls the region of any superpixel or patch as shown in Fig. 4. Whereas the regularizer ‘r’ controls the compactness of the patches as shown in Fig. 5. These parameters play an important role for the segmentation of images and if we pick large size superpixels then there is possibility for more than one classes in the single patch or superpixel, but if we set the size or take smaller size superpixel then there are less chances for more than one class in the superpixel. For this experiment we took different values for region side size ‘R’ as shown in the Fig. 4, where we took R = 5 value in case (b) but we took region side size R = 10 in the case (c), this experiment clearly shows there are chance of more than one class in the case (c) where we took region side size ‘R = 10’, because patches are bigger.

Fig. 4
figure 4

Superpixel-based segmentation of slices with different region sizes ‘R ‘ but same value for regularizer ‘r’: (a) Original image of FLAIR with High Grade Glioma (HGG) tumor (b) Region side size R = 5 and regularizer r = 0.2 (c) Region side size R = 10 and regularizer r = 0.2

To check the effect of regularizer, ‘r’ which is varied but the region side size is kept constant (R = 10) as shown in Fig. 5. Regularizer is the parameter which used to check the optimum compactness of the superpixel or patch. We can see in Fig. 5 where the regularizer ‘r’ values impact on the image patches such that in the case (a) patches are not clear due to low ‘r’ value, case (b) due to increase in the ‘r’ value patches are more clear than previous case but these patches more visible in case (c) due to optimum value of the regularizer ‘r’.

Fig. 5
figure 5

Superpixel-based segmentation where R=10 (a) r=0, (b) r=0.2 and (c) r=0.5 are different compactness parameter

The effect of region side size ‘R’ by varying the superpixel size is shown in Fig. 6. The optimum superpixel size, i.e., 400, which produces better results than others are finalized after a number of experiments performed using MATLAB, has been selected.

Fig. 6
figure 6

Variation in the region side size ‘R’ and regularizer ‘r’ in superpixel segmentation

Figure 6 clearly shows the optimum size value for the superpixel with their tuned parameter values, at the very low values and very high values gives very bad dice coefficient values. The optimum value is 400 which gives best dice values results shown in the Fig. 6 graph. After various experiments and analysis finally we pick the superpixel value 400 for whole dataset of training and testing purposes.

3.1 Fractal features parameters

To extract the fractal features, there are several threshold levels applied to have optimum features. These features are used to check the threshold level for accuracy in the classification. Examined by increasing the threshold level nt=2, these values create six binary channels and do not increase significantly by overlap measure. As we expand more levels, there is an increase of six more features (every binary channel creates a total three fractal features) in each feature vector. So, the threshold level for fractal feature extraction is chosen nt=2, which gives optimum results for superpixel classification and feature extraction [25].

3.2 Texton features parameters

NFB filter represents the total number of filters in the filter bank, has convolved all sequences. For each pixel, the response vector is created and the length of the vector are NFB. As the number of pixels in the image is same, therefore, number of response vectors are clustered into the k-clusters and are calculated using k-means clustering. The response vector from the filter response, which is corresponding to each cluster considered as the texton of the texture class. In the end, the texton map is created by assigning the cluster number for each pixel. After that, for each superpixel, the texton features are calculated taking the histogram of the texton map.

Gabor filters are used to extract texton features in six directions, i.e., [00, 300, 450, 600, 900, 1200], while size and wavelength coefficient of Gabor filter, are selected empirically, which are 0.3, 1.5 respectively. Applying the k-means clustering algorithm on the different responses produced by these filters create texton maps [25].

3.3 Results analysis

The proposed model has been evaluated on MICCAI BRATS 2017 dataset [18], which consists of both High Grade Glioma (HGG) and Low Grade Glioma (LGG) volumes. HGG has 210 volumes, while LGG contains 75 volumesFootnote 4. Although, each volume has four types of sequences, i.e., T1, T1Ce, T2, and FLAIR with its ground truth files. However, in this study, we utilized only FLAIR sequence. As FLAIR is the sequence having more contributing patterns for tumor region analysis [10]. The dataset is in the nifty files with extension “. nii”, and has 3D data volumes. Each sequence is with 155 slices. Sample FLAIR sequence and its ground truth slice is shown in Fig. 7.

Fig. 7
figure 7

MICCAI BRATS’17 training dataset (a) FLAIR slice (b) Ground Truth for the FLAIR slice

Normal and abnormal superpixels are classified using SVM, Decision Tree, KNN and an Ensemble classifier. Standard evaluation measures, i.e., Dice Coefficient (DSC), Precision, Sensitivity, and Balance Error Rates (BER) are used to evaluate the proposed model.

First of all, we evaluated the proposed model against each feature separately, the detailed average value for DSC is computed for each feature. DSC average value against statistical features, fractal features, texton features, curvature feature, and SIFT features are 0.52, 0.43, 0.68, 0.35, and 0.66 respectively. But this individual calculation of features couldn’t play best role due to low DSC values.

After experimental analysis, we examined these individual features are not performing best, therefore, to check, which features combination performed the best, we employed the “leave one out” strategy. We did make all different possible combinations for the feature extraction. All the possible combination sets show with their optimum results, but when we combine all these features and used for the experiments it gives better results as compare to other sets, so we decide to use overall features for model training. We did experiment for LGG and HGG volumes both and the DSC values for all the combinations are shown in the Table 1. Statistical analysis on these feature combinations shows, when we combine all the features extracted and dice values greater than the other combination sets.

Table 1 Results of features leave-one-out strategy

We used some machine learning classifiers, and check the model dice values by using these classifiers differently. We took SVM with their different kernels, Decision Tree, KNN and ensemble classifier individual to check which classifier performing better than others. But in all of them SVM with linear kernel has been used in this study, which obtained results closer to the ground truths as compare to other used classifiers. The results achieved on different classifiers are presented in Table 2. At the end we also did an experiment with ensemble technique where we combine all these classifiers and get the results based on maximum voting for the classifier.

Table 2 Results of different classifiers

We took some volumes for the experimental analysis as shown in the Table 2, and used SVM, Decision Tree, KNN and Ensemble classifier for the classification of normal and abnormal superpixels. After experimental analysis shown in Table 2, the SVM classifier and Ensemble got maximum DSC values as compare to the other classifiers. So, we choose this classifier for our model training and testing purposes.

3.4 Results for LGG segmentation

The visual results of LGG volumes are shown in Fig. 8. Figure 8(a) represents original image of LGG, Fig. 8(b) represents ground truth, Fig. 8(c) represents segmented image obtained through the proposed model, Fig. 8(d) represents original FLAIR image, Fig. 8(e) represents ground truth of Fig. 8(d) and (f) represent segmented image obtained through the proposed model.

Fig. 8
figure 8

Results of the proposed model for LGG volume, (a) Original Image (b) Ground Truth of (a), (c) Segmented Image of (a), (d) Original image, (e) ground truth of (d), (f) Segmented image of (d)

Volume wise results obtained through the proposed model for LGG are shown in Table 3. Although, we evaluated the proposed model against each volume of the dataset, however, for ease of its representation, we represented results of every fifth volume in Table 3. 105 slices have been used for training, while 10 are used for testing of the proposed model. The maximum value obtained against all LGG volumes (average) as per DSC is 0.8593. To evaluate the proposed model in respect of its validation, precision, sensitivity, and BER achieved is 87.63%, 92.40% and 0.08 respectively.

Table 3 Comparison evaluation on superpixel classification using SVM classifier for LGG volumes

Table 3 clearly shows the statistical results only for the LGG volumes, and these results shows our proposed model performance in the supervised learning approaches. Where we obtained these results using SVM and Ensemble classifier.

3.5 Results for HGG segmentation

The visual results of HGG volumes are shown in Fig. 9. Figure 9(a) represents original image of HGG, Fig. 9(b) represents ground truth of Fig. 9(a), (c) represents segmented image obtained through the proposed model, Fig. 9(d) represents original FLAIR image, Fig. 9(e) represents ground truth of Fig. 9(d), and (f) represent segmented image obtained using the proposed model.

Fig. 9
figure 9

Results of the proposed model for HGG volume representing (a) original Image (b) Tagged Image (c) Segmented Image (d) original image, (e) ground truth and (f) segmented image by model

Volume wise results obtained through the proposed model for HGG are shown in Table 4. Although, we evaluated the proposed model against each volume of the dataset, however, for ease of its representation, we represented results of every fifth volume in Table 3. 105 slices have been used for training, while 10 are used for testing of the proposed model. The maximum value obtained against all HGG volumes (average) as per DSC is 0.8528. To evaluate the proposed model in respect of its validation, precision, sensitivity, and BER achieved is 87.03%, 97.74% and 0.08 respectively.

Table 4 Comparison evaluation on superpixel classification using SVM-based classifier on BRATS 2017 dataset for HGG volumes

Table 4 statistical results against the HGG volumes, and all these results evaluated using SVM classifier with the linear kernel. Few volumes results are mentioned above on the five difference volumes only. The average DSC values for the HGG volumes are proved our proposed model performance.

3.6 Comparison with state-of-the-art studies

It is an important to compare the proposed model with state-of-the-art studies. Visual and statistical results obtained from proposed model is compared with state-of-the-art studies as shown in Tables 5 and 6. First, we have compared the reference paper DSC values for the same volumes that have used for their model evaluation. The reference paper used MICCAI BRATS-2012 and 2013 dataset volumes for model evaluation, whereas these same volumes are also available in the MICCAI BRATS-2107 training dataset that we used for our proposed model evaluation. Average DSC values for the proposed model on the same volumes is 0.82 while in reference paper [25], average DSC value is 0.80, which depicts that the proposed model outperformed as compared to the reference paper. The proposed model set the superpixel segmentation parameters with different experimental analysis, also introduce the class balancing algorithm for feature vectors and extract more features like SIFT which gives better results as compared to the reference paper [25].

Table 5 Case by case comparison w.r.t DSC values with study [25]

Moreover, the results obtained through the proposed model have also been compared with the recent more state-of-the-art studies, which used MICCAI BRATS dataset, as shown in Table 6. It is evident that the average DSC value achieved in case of the results obtained using the proposed model is high as compare to the state-of-the-art studies.

Table 6 Comparison w.r.t dice co-efficient with state-of-the-art studies

Various techniques are introduced for this complex problem like tumor segmentation in brain MRI but there are need to address the spatial and intensity variation in the edema tumor, also in the tumor appearance. Hao Chen et al. [5] proposed solution to overcome the tumor segmentation problem in MRI using novel approach named Deep Convolutional Symmetric Neural Network (DCSNN), where they add more symmetry masks in the DCNN architecture layers, and they used BRATS 2015 dataset to evaluate the model where they obtained 0.852 DSC value and address the spatial variability of tumor in MRI. Linmin et al. [21] address the segmentation problem in brain MRI using two different approaches, feature fusion and joint label fusion (JLF). The purpose of this study was to find the accurate tumor segments in the BRATS 2015 MRI dataset, and this technique obtained average 0.85 DSC value. Accurately detection and segmentation for the brain is the need of time and there are various problems by using any manual method, Salma et al. [2] discussed about the novel approach using SegNet algorithm with the decision tree (DT) algorithm for classification of tumor. BRATS 2017 dataset used to evaluate the model and obtained 0.85 DSC/F-measure value. This fully automatic segmentation approach deal to overcome the various segmentation problems such that to give the solution for accurate segmentation of tumorous region.

Table 6 DSC results for the whole tumor/edema shows our proposed model outperformed as compare to the other state-of-the-art studies. The reason is that we used novel approach as discussed in the research methodology and used large volume dataset for the better performance. There are some key points which helps in this study to improve the proposed model such that we used superpixel-based segmentation and tune the best parameters which perform best, extract various possible features like SIFT features, proposed class balancing algorithm for the unbalanced feature vector and used MICCAI BRATS 2017 whole dataset for the evaluation of proposed model.

4 Conclusions

This paper presented an automated segmentation of the tumor from brain MRI images using superpixel-based segmentation technique. To segment out the tumors from FALIR MRI pre-processing techniques applied to remove the outliers from slices, because it affects in the overall process. Superpixel-segmentation applied on the slices with the improved parameter values. In superpixel-based segmentation, the most important is the optimum size of the superpixel, in which the compactness parameter plays a vital role in the selection of the size of the superpixel. If the size of the superpixel is large, then more than one class will be available in the superpixel and result in false classification. Similarly, if the size of superpixel is small, then possible unique class is found. The optimum size of the superpixel has also been proposed in this study for better classification of tumors in brain MRI images. Several features including statistical, fractal, texton, curvature, and SIFT have been extracted. Due to large volume dataset and multiple features the obtained feature vector was unbalanced as it has larger number of normal class and small values of the abnormal class. The proposed model introduced class balancing algorithm to overcome this issue and different machine learning classifiers such that SVM, KNN, Decision Tree and ensemble have been used for training purpose. The proposed model has been evaluated on MICCAI BRATS-2017 training dataset using FLAIR modality. Evaluation measures, i.e., dice coefficient (DSC), precision, sensitivity, and BER results are 0.85, 87%, 97%, and 0.08 respectively evaluated both for segmentation and validation outcomes. Various existing techniques compared with the proposed model and our model outperformed due to the novel approach in superpixel segmentation parameters, class balancing algorithm and various new features. In future we can use large volume data with some deep learning techniques, and can give the practical application in real-time problem.