Abstract
In modern health-care, for evidence-based diagnosis, there is a requirement for an efficient image retrieval approach to retrieve the cases of interest that have similar characteristics from the large image databases. This paper presents a feature extraction approach that aims at extracting texture features present in the medical images using Local Pattern Descriptor (LPD) and Gray-level Co-occurrence Matrix (GLCM). As a main contribution, a novel local pattern named Local Mesh Vector Co-occurrence Pattern (LMVCoP) has been proposed by concatenating the Local Mesh Co-occurrence Pattern (LMCoP) and the Local Vector Co-occurrence Pattern (LVCoP). The fusion of GLCM with the Local Mesh Pattern (LMeP) and the Local Vector Pattern (LVP) produces LMCoP and LVCoP respectively. The LMVCoP method has been investigated on the Open Access Series of Imaging Studies (OASIS): a Magnetic Resonance Imaging (MRI) brain image database. LMVCoP descriptor achieves 87.57% of ARP and 53.21% of ARR which are higher than the existing methods of LTCoP, PVEP, LBDP, LMeP and LVP. The LMVCoP method enhances the retrieval results of LMeP/LVP from 81.36%/83.52% to 87.57% in terms of ARP on OASIS MRI brain database.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Introduction
Every day massive amount of medical images is acquired and preserved as databases in health care centers for future references, disease diagnostics and for proper treatment. Physician uses these large medical image data as the modern frontier of the medical field and improves the quality of medical treatment. Many referring physicians also show a great interest for direct access to medical image data and to discuss with their hospital colleagues by retrieving the required image immediately and have the chance to confer the case collaboratively with other specialists for successful decision making. Healthy people also show interest to have a master medical check up periodically which increases the accumulation of medical images in the hospitals. Hence, there is a need to find the most clinically relevant images in response to specific information required. Content-Based Image Retrieval (CBIR) focused to follow a line of investigation to fulfill this requirement. Where CBIR is a well-known technology used to organize an image database by their features (contents) and to search it according to the requirement. Ultimately, the Content-Based Medical Image Retrieval (CBMIR) attracts extensive attention in the research community by its significant clinical benefits and the pace of its research is still rising.
Duncan and Ayache listed out four important challenges should be faced while analysing medical images [1]. They are: (1) addressing the biological problems, (2) Characterizing different imaging modalities, (3) identifying proper techniques for geometrical analysis of both normal and abnormal information, and (4) integrating with other communities’ research activities to provide new opportunities. Lakovidis et al. [2] extracted multi-scale statistical features using a 2D discrete wavelet transform and by clustering those features into small patterns, by comprising the sets of these smaller patterns and representing as complex patterns. Rahman et al. [3] proposed a multimodal query expansion method by integrating textual keywords and visual data. Bag-of-features [4] such as Scale Invariant Feature Transform (SIFT) and image patches have been modelled using the quadratic programming assignment and analyzed by both the strong and weak classifiers on the ImageCLEF, CT and the basal-cell carcinoma datasets by Wang et al. in 2011. Followed by, Quellec et al. [5] introduced a content-based heterogeneous information retrieval framework using a Bayesian network. Then Cao et al. [6] proposed a multimodal medical image retrieval approach using the statistical graphic model and deep learning. To overcome the limitation of semantic gap, the authors introduced an extended probabilistic latent semantic analysis model by combining both the visual and textual information from medical images for cancer care. And to derive the missing modality, they developed a Boltzmann machine-based multimodal learning model. Next, saliency based folding method has been developed by employing Local Binary Patterns (LBP) and Support Vector Machines (SVM) in 2015 [7]. Albarqouni et al. [8] used Convolutional Neural Networks (CNN) through additional crowd sourcing layer (AggNet) for deep learning for Mitosis detection from the breast cancer histology images. Later, fuzzy aggregation operators based dental diagnosis has been performed dental X-ray images [9]. Shortly, Speeded Up Robust Feature (SURF) descriptor has been developed on the lung images for implementing an efficient CBMIR [10]. On the Image Retrieval in Medical Applications (IRMA) dataset, Radon Bar-Code (RBC), Convolutional Neural Network Code (CNNC) and Regions of Interest (ROI) based matching has been investigated along with the cross-correlation method [11]. Recently, Expectation Maximization (EM) and Fuzzy C-Means (FCM) were implemented on MRI brain images [12].
Feature extraction takes a vital role of extracting the fine features of images for better image retrieval. The efficiency of the retrieval can be augmented by choosing the perfect feature extraction technique. Furthermore, the features present in the medical images are complex; anatomical structures are highly irregular and multimodality image’ structures are misaligned. To face this challenge, a novel texture based local pattern descriptor has been proposed in this paper to extract features from MRI brain images and to retrieve them automatically whenever needed.
The evolution of local patterns is as follows: the Local Binary Pattern (LBP) [13] has been initially intended for texture description. Subsequently, Tan and Triggs [14] provided the Local Ternary Pattern (LTP) as an expansion of the LBP. Later, Murala et al. proposed the Local Tetra Pattern (LTrP) [15] for facial image recognition and the Directional Binary Wavelet Pattern (DBWP) [16] for MRI, CT and facial image retrieval. After that, Murala and Wu proposed the Peak Valley Edge Patterns [17] and the Local Mesh Pattern (LMeP) [18] on biomedical images. Later, the Local Vector Pattern (LVP) [19] has been developed for facial feature extraction. Then Yesubai and Ravi proposed the Local Mesh Co-occurrence Pattern (LMCoP) on natural scenery images [20]. Thus, the local patterns have gained much attention in the facial recognition area than the medical field. It stimulates the idea of introducing a new hybrid local pattern LMVCoP for medical image retrieval application.
The proposed method
This section describes the proposed CBMIR based on LMVCoP technique in more detail. Figure 1 depicts the flow of the proposed system. The steps followed in this work are listed below:
Step 1: Pre-processing
Given image is first pre-processed to synchronize the content in a database and thus create effective feature extraction from the images based on the same basis. The image is initially converted into a gray-scale image and resized into 256 × 256 pixels. The image is then normalized using contrast enhancement of adaptive histogram equalization. Since adaptive histogram equalization distributes the most repeated intensity values and transforms the lower local contrast into higher contrast effectively, this method is used here. Followed by, the process of smoothing that reduces the noise was performed to produce a more accurate intensity surface. Since Gaussian filter is very effective for removing Gaussian noise, reducing edge blurring, rotationally symmetric and computationally efficient for MRI brain images, Gaussian smoothing is used in this work for a removal of small unconnected objects and to adjust for residual between-subject neuro-anatomic differences.
Step 2: Extracting features using local vector co-occurrence pattern
Given a pre-processed local sub-region I, α is the index angle, g c is the center pixel in I, D is the distance between the centre pixel and its neighboring pixels in the α direction. First, a vector that denoted as V α , D (g c ) is calculated by
Then, the LVP N , R , α (g c ), in the α direction is encoded as
Where N = 1 , 2 , … , M ; R = 1,
T (·, ·) is defined as
Next, the LVP N , R (g c ), is computed by combining four 8-bit binary patterns. For every center pixel of I, a 32-bit binary pattern is generated.
Finally, the resultant LVP map is computed using GLCM, where, M is a matrix whose size M x × M y is equal to the number of gray-levels 0 to M-1. At an angle θ and d distance, the variation between gray level i and j, has been calculated and represented as the matrix element C(i,j|d,θ). C x (i) and C y (j) are computed by
Once the Co-occurrence matrix has been formed, the following Haralick texture features can be computed by
where n = |i − j| , i ≠ j
where mean μ x and μ y take the average level of intensity of given input such as
whereas the variance σ x and σ y describe the variation of intensity around the mean.
Step 3: Extracting features using local mesh co-occurrence pattern
Given a pre-processed local sub-region I, the relationship among the surrounding neighbours of the centre pixel I 1(g n ) is computed by
where n = 1 , 2 , … , 8. Then the LMeP can be calculated by
After computing the LMeP, apply GLCM on the resultant LMeP map using Eqs. (5–12).
Step 4: Computing local mesh vector co-occurrence pattern
The local pattern results are combined using the sum rule. For Example, LMCoP denotes GLCM of LMeP, similarly LVCoP denotes GLCM of LVP, and then the combined approach of these two patterns is abbreviated as LMVCoP, which can be written as:
where, w is a weighted parameter that ranges from 0.1 to 0.9. The relationship between LMCoP and LVCoP can be investigated by changing w, and the performance of LMVCoP can also be evaluated. In this work, initially, all the tests were performed for w = 0.2
Step 5: Retrieval
Once the proposed model LMVCoP is trained for all the medical images of the given dataset, the feature vectors are stored as a feature dataset with reduced dimension. Then, the step, retrieval is performed. The similarity measurement is computed between the feature vectors of feature dataset and the feature vector of query image based on Squared Euclidean distance and sorted the values in ascending order. The image of the database which has very minimum difference or shorter distance attains a higher position and considered as the image having higher similarity. As per request from the user the retrieved images have been displayed in the main Graphical User Interface (GUI).
Step 6: Performance evaluation
The effectiveness of the LMCoP and LVCoP techniques is evaluated using the following error sensitivity measures:
Mean Squared Error:
Peak Signal-to-Noise Ratio:
Signal-to-Noise Ratio:
Average Difference:
Normalized Absolute Error:
Normalized Cross-Correlation:
where, I is a given input image. E is the estimated or output image. M × N is the size of the image. The efficiency of the proposed image retrieval system is evaluated using the following performance evaluation metrics: Precision P j (M) is calculated by
Whereas Recall, R j (M) is computed by
Consider j th dataset image is used as a query image. I M is the number of relevant images retrieved, M is the total number of images retrieved and S is the total number of relevant images. The average precision \( {P}_a^i(M) \) and average recall \( {R}_a^i(M) \)for each class in the database are computed by
where i is the class number and a is the average. The Average Retrieval Precision (ARP) and Average Retrieval Rate (ARR) respectively were calculated using Eqs. (27 and 28).
where C is the total number of classes in the database. In this work, since the proposed LMVCoP is investigated only in MRI brain images, C = 1, Hence \( {P}_a^i(M)={P}_a^T(M) \) and \( {R}_a^i(M)={R}_a^T(M). \)
Results and discussion
The proposed LMVCoP is implemented in MATLAB software by building GUI in an Intel (R) Core i3 processor, 3.20 GHz, 4 GB RAM computer with the Microsoft Window7 Professional operation system platform. To test the efficiency of the proposed methodology a large data set of OASIS: MRI images [21] have been used in this work. There are four groups of dataset contain 106, 89, 102, and 124 MRI brain images were used for the test. For a data set, the evaluation is performed independently.
The proposed system interacts with the user through a succession of GUIs. First, in the pre-processing step, the given query image is converted into gray-scale image. After that, the scaling algorithm has been applied to resize the gray-scale image. Then, the image features have been enhanced with the help of contrast enhancement algorithm. Next, Gaussian smoothing is used to reduce the noise. Second, the features are extracted using the LVP and GLCM. Finally, LVCoP map is obtained. The results are shown as a snapshot in Fig. 2. The validation has been done using error sensitivity measures and values are displayed. After computing LVCoP, the function calls back to the main menu and the computation of LMCoP carries out as per the above-mentioned procedure. The results of LMCoP are shown in Fig. 3. The fusion of LVCoP and LMCoP are done and the LMVCoP features are generated and stored as feature vector database. The similarity between the features of database images and the query image are performed using Squared Euclidean distance and the retrieval images are displayed to the user. A screen shot of the results of top 10 retrieved images are shown in Fig. 4. Further, more number of (20 and 30) images are retrieved and displayed. A screen shot of the results of the various number of images retrieved are shown in Figs. 5 and 6 respectively.
The performance of LMCoP and LVCoP has been compared for a given input brain MRI image in terms of error sensitivity measures such as Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR), Signal-to-Noise Ratio (SNR), Average Difference (AD), Normalized Absolute Error (NAE) and Normalized Cross Correlation (NXC) and the results are listed out in Table 1. When MSE is low, the PSNR value is high, then the quality rate turn into high. Here, PSNR is in the range of 50-60db. Hence, it is understood that the quality of processed image is high. Generally, SNR is defined as the power ratio and expressed in db. But, here it is expressed as the signal ratio. Hence, it is given in number. In general, there are 17 Haralick features. Out of 17 features, the Contrast, Correlation, Energy and Homogeneity features are considered as very important features that are obtained by Eqs, (7, 8, 11 and 12) and recorded in the Table 2. The contrast values reflect the sharp structural variations of medical images. The correlation imply that fine linear dependent features are present in the given images. The energy is used to measure texture crudeness; if it is equal for all the images, then it is concluded that the co-occurrence matrix has same values. But, from the values in Table 2, it is understood that the values of co-occurrence matrix are unequal. In view of the fact that the values of homogeneity of images #1, #5, #10 are nearer to 1 illustrate that they have diagonal GLCM. The Table 3 listed out the time spent for performing feature extraction and retrieval processes using the LMVCoP and the existing methods of LTCoP, PVEP, LMeP LBDP and LVP on bio-medical images. From Table 3, it is understood that the speed of feature extraction using LMVCoP is faster than the existing methods. Table 4 depicted the group-wise ARP based retrieval performance of all the state-of-the-arts of the LPD on the OASIS: MRI brain dataset. The comparison Table 4 depicts that the the LMVCoP outperforms in groups 1, 2 and 3. In group 4, GLTCoP, GLTCoPu2, LDMaMEP and LMeCH perform well than the proposed LMVCoP descriptor.
Figure 7 illustrates the performance of LMVCoP on OASIS: MRI brain images in terms of the ARP based on Squared Euclidean distance. It is evident that the average retrieval precision of LMVCoP, depicted in Fig. 7 is higher than all other existing methods. The high precision values obtained indicate that the LMVCoP is able to retrieve more relevant images. The ARR result exposed in Fig. 8 illustrates that as precision increases, the recall value decreases. Hence, it is known that the LMVCoP has given better performance. The proposed LMVCoP descriptor achieved the ARP of 87.57% and the ARR of 53.21% for the first 10 number of retrieval images.
Conclusions
In this paper, the LMVCoP descriptor has been proposed to obtain better retrieval performance on medical images for the fast decision making in clinical application. The LMVCoP extracted the texture features of the MRI brain images by concatenating the LPD and GLCM. The performance of LMVCoP has been compared in terms of performance evaluation metrics with the existing descriptors that were used for biomedical indexing and retrieval such as LTCoP, PVEP, LMeP, LBDP and LVP. LMVCoP descriptor achieves 87.57% of ARP and 53.21% of ARR. The time spent by LMVCoP descriptors for feature extraction is 04.29 s and retrieval is 0.21 s that were also less. The simulation results and investigations prove that the LMVCoP outperforms the LTCoP by 31.61%, the PVEP by 17.43%, the LMeP by 6.21%, the LBDP by 4.6%, and the LVP by 4.05% in terms of ARP. Implementing higher-order LMVCoP is an immediate challenge. In future, it is planned to implement higher order LMVCoP and to combine LPDs and neural network methods for retrieving large-scale high dimentional image datasets.
References
Duncan, J.S., and Ayache, N., Medical image analysis: Progress over two decades and the challenges ahead. IEEE Trans. Pattern Anal. Mach. Intell. 22(1):85–106, 2000.
Iakovidis, D., Pelekis, N., Kotsifakos, E., Kopanakis, I., Karanikas, H., and Theodoridis, Y., A pattern similarity scheme for medical image retrieval. IEEE Trans. Inf. Technol. Biomed. 13(4):442–450, 2009.
Rahman, M. M., Antani, S. K., Long, R. L., Demner, D-F., and Thoma, G. R., Multi-Modal Query Expansion Based on Local Analysis for Medical Image Retrieval. Medical Content-Based Retrieval for Clinical Decision Support. Berlin, Springer; 110–119, 2010.
Wang, J., Li, Y., Zhang, Y., Wang, C., Xie, H., Chen, G., and Gao, X., Bag-of-features based medical image retrieval via multiple assignment and visual words weighting. IEEE Trans. Med. Imag. 30(11):1996–2011, 2011.
Quellec, G., Lamard, M., Cazuguel, G., Roux, C., and Cochener, B., Case retrieval in medical databases by fusing heterogeneous information. IEEE Trans. Med. Imag. 30(1):108–118, 2011.
Cao, Y., Steffey, S., He, J., Xiao, D., Tao, C., Chen, P., and Müller, H., Medical Image Retrieval: A Multimodal Approach. Cancer Informat. 13(S3):125–136, 2014.
Zehra, C., Amlica, H., Tizhoosh, R., and Khalvati, F., Medical Image Classification via SVM using LBP Features from Saliency-Based Folded Data, Proceedings of the 14th International Conference on Machine Learning and Applications (IEEE ICMLA’15), Miami, Florida, USA, (pp 1–5), 2015.
Albarqouni, S., Baur, C., Achilles, F., Belagiannis, V., Demirci, S., and Navab, N., AggNet: Deep learning from crowds for mitosis detection in breast cancer histology images. IEEE Trans. Med. Imag. Special Issue on Deep Learning. 35(5):1313–1321, 2016.
Ngan, T.T., Tuan, T.M., Son, L.H., Minh, N.H., and Dey, N., Decision making based on fuzzy aggregation operators for medical diagnosis from dental X-ray images. J Med Syst. 40:280, 2016.
Govindaraju, S., and Ramesh Kumar, G.P., A novel content based medical image retrieval using SURF features. Indian J. Sci. Technol. 9(20):1–8, 2016.
Liu, X., Tizhoosh, H. R., and Kofman, J., Generating Binary Tags for Fast Medical Image Retrieval Based on Convolutional Nets and Radon Transform. Proceedings of the 2016 I.E. International Joint Conference on Neural Networks (IJCNN 2016), Vancouver, Canada, (pp. 1–7), 2016.
Meena, P.R., and Shantha, S.R., Spatial fuzzy C means and expectation maximization algorithms with bias correction for segmentation of MR brain images. J. Med. Syst. 41:15, 2017.
Ojala, T., Pietikäinen, M., and Harwood, D., A comparative study of texture measures with classification based on featured distribution. Pattern Recognit. 29(1):51–59, 1996.
Tan, X., and Triggs, B., Enhanced local texture feature sets for face recognition under difficult lighting conditions. IEEE Trans. Image Process. 19(6):1635–1650, 2010.
Murala, S., Maheshwari, R.P., and Balasubramanian, R., Local tetra patterns: A new descriptor for content-based image retrieval. IEEE Trans Image Process. 21(5):2874–2886, 2012.
Murala, S., Maheshwari, R.P., and Balasubramanian, R., Directional binary wavelet patterns for biomedical image indexing and retrieval. J. Med. Syst. 36(5):2865–2879, 2012.
Murala, S., and Wu, Q. M. J., Peak Valley Edge Patterns: A New Descriptor for Biomedical Image Indexing and Retrieval. Proc. IEEE CVPR Workshops (pp. 444-449), 2013.
Murala, S., and Wu, Q.M.J., Local mesh patterns versus local binary patterns: Biomedical image indexing and retrieval. IEEE J. Biomed. Health Inform. 18(3):929–938, 2014.
Fan, K.-C., and Hung, T.-Y., A novel local pattern descriptor-local vector pattern in high-order derivative space for face recognition. IEEE Trans. on Image Process. 23(7):2877–2891, 2014.
Yesubai, R. C., and Ravi, R., Local mesh co-occurrence pattern for content based image retrieval. Inte. J. Comput. Electr. Autom. Control Info. Eng. 9(6): 1510–1515, 2015.
Marcus, D.S., Wang, T.H., Parker, J., Csernansky, J.G., Morris, J.C., and Buckner, R.L., Open access series of imaging studies (OASIS): Cross-sectional MRI data in young, middle aged, non-demented, and demented older adults. J. Cogn. Neurosci. 19(9):1498–1507, 2007.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Human and animal rights and informed consent
This article does not contain any studies with human participants or animals performed by any of the authors.
Additional information
This article is part of the Topical Collection on Image & Signal Processing
Rights and permissions
About this article
Cite this article
Jenitta, A., Samson Ravindran, R. Image Retrieval Based on Local Mesh Vector Co-occurrence Pattern for Medical Diagnosis from MRI Brain Images. J Med Syst 41, 157 (2017). https://doi.org/10.1007/s10916-017-0799-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10916-017-0799-z