Keywords

1 Introduction

Breast Cancer is one of the most common cancers in women. As per official reports, in India, 25%–30% of all female related deaths were resultant of this cancer [1]. A study showed that in 2018, 1.6 million new cases were registered, and 87,090 deaths were reported [2]. A major reason for this is less public awareness along with none or very fewer screenings with high testing prices. Figure 1. shows, number of cases when compared to 25 years ago shows an increase in breast cancer in the age group between 20–50 [3]. Although the exact reason for the development of breast cancer is still unknown, several lifestyle guidelines are stated, which decreases the chances of development of breast cancer. Maintaining balanced BMI with regular physical exercise and breastfeeding are several suggestions [4]. But not all reasons cannot be controlled, menstruation in younger age, menopause in the older age, late marriage, contraceptive drug are namely a few, which increases the chances of breast cancer.

Fig. 1.
figure 1

Brest cancer % change in India of different age group

With the advancement of technology in both medicine and computer science, the detection, diagnosis, and treatment of diseases have improved drastically. New methods and techniques are being discovered which aids in the medical process. For breast cancer detection, many imaging modalities exist. In hospitals, various breast imaging methods are used in early breast cancer detection and screening, including MRI [5], computed tomography (CT) [6], magnetic resonance imaging (MRI) [7, 8]. But a mammogram is gaining popularity for its low complexity and better availability. Figure 2 [9] shows all the imaging techniques which are used for breast cancer detection.

However, this paper focuses mainly on Mammograms and used MIAS dataset for experiments. Mammography is achieved through an X-ray exposure of the breast. The breast tissue can absorb X-ray radiation as it is exposed to X-ray. There are various signal levels of the breast tissue and cancer cells. But the question of classifying lumps in mammogram lands on the radiologist, whose prediction is based on experience [10] and, quality of mammogram [11]. Adding to this, breast anomalies are also hidden by the breast tissue structure which makes it more difficult to detect [12]. A major problem in mammogram images is its low contrast [13], which makes it difficult to detect lumps, and have shown a high rate of false-positive cases (regular change as cancerous) and false negative (actual abnormality not detected) [14,15,16].

Fig. 2.
figure 2

Different imaging modalities for the diagnosis of breast cancer

Image fusion is a process which applies different methods and techniques for combining several images information either from the same platform or from different spectroscopic platforms to create a single output image. The resultant image (known as fused image) has more detailed and useful or predictable information for machine perception or human understanding [17]. Each input image might have a different focus area, and the complete information might not be presented in a single image. Image fusion process combines these images, which is more detailed than a single image. To combine multiple images, all images should be of the same area, and different angled photos result in difficulty in the fusion process.

Different medical conditions require a separate process to be followed for its treatment, image fusion can combine MRI, CT, PET, SPECT data together for better results. Better achievements have been achieved in improving clinical accuracy by using Multi-modal medical image fusion algorithms [18].

Pixel level [19], feature level and decision level fusion [20] are types of fusion techniques and is explained with help of diagram in Fig. 3(a). In this paper, we used decision level fusion for our experiment. In this method, we use the fusion process after the classification step is completed. This process is generally a combination of multiple algorithms to obtain the resultant image. When confidences are used instead of decision, it is known as soft fusion. Otherwise, it is called hard fusion. Methods of decision fusion are shown in Fig. 3(b).

Fig. 3.
figure 3

(a) Pixel, feature and decision level fusion. (b) decision level methods

With advancements in algorithms and better computational power, Machine learning has helped to solve many real-life problems. These algorithms help in managing huge amount of data along with finding correlation among data which is not possible to find manually. Deep learning is also a subfield of AI which is gaining a lot of popularity. In this paper, we have used deep learning for feature extraction and used these features for classification using SVM and Decision tree. Finally, we have used voting classifier for decision level fusion to combine the results from these two classifiers and predicted our output.

2 Literature Review

This section reviews the related work done by researchers using fusion techniques. Pixel level, feature level and decision level fusion techniques are used by researchers for different modalities like MRI, CT and mammogram [21] as well as other modalities are also used and reviewed. Multiple authors have also purposed CADx solution [22, 23], pipeline structures and frameworks for fusion and classification of breast cancer.

While using MIAS dataset for their experiment, authors implemented pixel level fusion in their experiment. They conducted their experiment on three different mammograms from dataset of Normal, Benign and Microcalcification X-rays. They tested simple average and weighted average method and documented the results based on Signal to Noise Ratio (SNR), Peak Signal to Noise Ratio (PSNR), Root Mean Square Error (RMSE), Mutual Information (MI), etc. They concluded that, using image fusion provides better results than original image [24].

While using the same MIAS dataset, authors presented their image fusion method using Particle swarm optimization (PSO). The PSO used to calculate the optimum weighted weights for fusion and compared the results with conventional DWT and genetic algorithms. They compared the results on same fusion parameters as of author [24] and concluded that genetic method based DWT provides better results than Weighted average and traditional DWT [25].

Authors of paper [26], presented local entropy maximization based image fusion technique to improve the contrast of mammograms. They used MIAS and TMCH dataset for their experiment. Using Haar wavelet they decomposed the original and CLAHE image into 3 levels. Using sliding window of 5 × 5 window size, they fused the coefficients while choosing the maximum entropy. Finally, the fused image is reconstructed using these coefficients and validated their outcomes based of edge contents (EC), edge-based contrast measure (EBCM), feature similarity index measure (FSIM) and absolute mean brightness error (AMBE). They achieved 1.87 EC, 120.1 EBCM, 0.97 FSIM and 2.01 AMBE. They compared their results with HE, BBHE and CLAHE and their method showed better results.

Using 400 mammogram images from hospitals, authors [27], have purposed a CAD system using feature fusion techniques. First, they suggested a method of mass detection based on CNN deep features and clustering with Unsupervised Extreme Learning Machine (US-ELM). Second, they establish a collection of features that incorporate deep features, morphological features, texture and density features. Third, using the merged function collection to distinguish benign and malignant breast masses, an ELM classifier is established.

Authors of paper [28], purposed a wavelet fusion along with CLAHE enhancement for their experiment. They used multi-modalities images. In first step, they enhanced the contrast of image using CLAHE and second, they used 2D wavelet transformation fusion to generate the fused image. They compared their results on parameters like SNR and found their method performs better for different medical images with low contrast.

[29] presented a CAD system in which they used DDSM dataset for their experiment. Their experiment includes merging features of MIO and CC views of mammograms for better results. While using five features namely GLRLM, GLCM and others. While using SVM as classifier and using RBF kernel as performance booster they were able to achieve 97.5% accuracy, 100% sensitivity, 97.2% specificity, 97.1% precision, 96.23% F1 score, 0.952% Mathews Correlation Coefficient and 98.74% Balanced Classification Rate.

While using DDSM dataset for their experiment, authors of paper [30], used ensemble of CNN for classification of mammograms. The implemented data cleaning by contrast fading and removed white strips in input images of dataset and in pre-processing padding, dilation and cropping is applied. Since they used CNN for their experiment, they used data augmentation to solve overfitting issues in their model. Finally, they used GoogleNet for their classification step. Their decision fusion is based on max ensemble technique. After training their model for 50000 iterations they were able to achieve 91.3% recall value in stand-alone setup and were able to increase this to 97.3% with ensemble. 94.5% F1 score and 95% precision value is achieved.

[31] while also using decision level fusion used 65 thermography images gathered from [32, 33] and [34]. They purposed a novel texture feature extraction based on Markov Random Field (MRF) model and another texture based on LBP are extracted from images. While implementing decision fusion based on HMM, they were able to achieve 8.3% false negative and 5% false positive rate.

Whereas authors of paper [35] implemented a deep feature fusion of 3 different imaging modalities together. They used mammogram dataset FFDM containing 245 unique images, Ultrasound dataset containing 1125 images and DCE-MRI dataset containing 690 images. While using publicly available VGG19 model they implemented CNN model and were able to achieve AUC = 0.89 for DCE-MRI, AUC = 0.86 for FFDM and AUC = 0.9 for ultrasound (Table 1).

Table 1. Fusion techineques overview.

3 Materials and Methods

3.1 Datasets

In this experiment, we have used MIAS dataset. MIAS dataset consists of 161 pair of films of abnormalities and normal cases. It consists of 322 mammograms which are selected from United Kingdom National Breast Screening Program. A major factor for selection MIAS dataset is, it consists of mammograms which are cheap, low complexity and easily available in countries. MIAS dataset is available in two sizes (50 μ and 200 μ). There are other mammography datasets publicly like DDSM, TMCH,B-SCREEN [36] etc.

3.2 Pre-processing

A major problem in mammogram images is its low contrast, which makes it difficult to detect lumps, and have shown a high rate of false-positive cases (regular change as cancerous) and false negative (actual abnormality not detected). To solve this problem, we have implemented the CLAHE enhancement technique. CLAHE is a version of AHE in which we define a threshold level at which the intensities are clipped. Clip limit of 0.2 was used for this experiment and was coded in python.

3.3 Segmentation

Another issue with mammograms is the non-important area in the film. For our algorithms to achieve better results and to reduce the processing time we must trim the images to select the Region of Interest (ROI). Generally, according to the view of the breast (left or right), a majority of the portion is pixels with 0 value which should be removed. In this experiment, we have implemented a sliding vertical line from left or right depending on the image to trim it until pixels with non-zero pixel value is encountered.

3.4 Feature Extraction and Selection

This is one of the main important steps in our workflow. The output of the process heavily depends on the pre-processing techniques to enhance the images and features used for classification. Features more associated to the output class contribute better than non-associated features. In this step, we have used the power of deep learning algorithms for patterns in our input image. A CNN model is created which extracts the features and these features are used by classification algorithm (SVN and Decision Tree). A model of CNN architecture is shown in Fig. 4. A total of 1754 high-level features and 288 low-level features were used by the classification algorithm.

Fig. 4.
figure 4

CNN model architecture for feature extraction

3.5 Classification and Decision Fusion

In this step, we have implemented SVM and Decision classifier. SVM is a machine learning algorithm which can be used for both classification and regression. Decision tree is also implemented for the classification of breast cancer. Finally, a voting classifier is used for making a decision based on these two input classifiers and final output is generated. Figure 5, shows the workflow of the process which we have used for our experiment.

Fig. 5.
figure 5

Workflow of the process for breast cancer detection

4 Results

In this experiment, we have used MIAS dataset consisting of 161 pair of mammograms (322 total). CLAHE enhancement technique is used for improving the contrast of the images and the CNN model is used for the extraction of features. A total of 288 low-level and 1754 high-level features were extracted and were used for classification. Standalone SVM was able to achieve 90.3% accuracy, 87.8% sensitivity and 93% sensitivity while Decision tree was able to achieve 92.03% accuracy. After combining both the classification techniques using a voting classifier, we were able to achieve 93.4% accuracy.

5 Conclusion

Breast cancer is major affected diseases in women. After reviewing many techniques and methods in this paper we found out that a CAD system seems to be a good solution for real-life use by radiologist. With radiologist own expertise and second and helping opinion from CAD system will help to address the accuracy of diagnosis by improving the image, selecting the ROI. Further, a feature fusion along with decision fusion can be implemented to improve the results.