Abstract
Artificial neural network (ANN) introduced in the 1950s, is a machine learning framework inspired by the functioning of human neurons. However, for a long time the ANN remained inadequate in solving real problems, because of - the problems of overfitting and vanishing gradient while training a deep architecture, dearth of computation power, and non-availability of enough data for training the framework. This concept has lately re-emerged, in the form of Deep Learning (DL) which initially developed for computer vision and became immensely popular in several other domains. It gained traction in late 2012, when a DL approach i.e. convolutional neural network won in the ImageNet Classification – an acclaimed worldwide computer vision competition. Thereafter, researchers in practically every domain, including medical imaging, started vigorously contributing in the massively progressing field of DL. The success of DL methods can be owed to the availability of data, boosted computation power provided by the existing graphics processing units (GPUs), and ground-breaking training algorithms. In this paper, we have overviewed the area of DL in medical imaging, including (1) machine learning and DL basics, (2) cause of power of DL, (3) common DL models, (4) their applications to medical imaging and (5) challenges and future work in this field.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Machine learning (ML), a subset of “artificial intelligence” (AI), can be described as a class of algorithms which identify patterns within the data automatically and use these for prediction or decision making on future data [1] with minimum interventions by a human. The usage of ML has seen a rapid increase in medical imaging [2,3,4,5,6,7,8,9], and medical image analysis, since the objects in medical images can be quite intricate to be represented correctly by some basic function. Representing such objects requires complex modelling with several parameters. Determining numerous parameters manually from the data is nearly impossible. Therefore, ML plays an essential role in the detection and diagnostics in medical imaging field. Computer-aided diagnosis (CAD) - one of the most prevalent usages of ML [2, 10] - involves classifying objects into pre-defined classes (e.g., lesions or non-lesions, and benign or malignant) depending upon input data.
Towards the end of the 1990s, ML techniques, both supervised (that use training data labels to understand patterns in data) and unsupervised (that do not use labelled training data to understand the data patterns), became very popular in medical image analysis. Examples include segmentation of brain regions, extraction of representative features and classification for CAD. ML is immensely popular and happens to be the foundation of several successful commercial medical image analysis products. ML has brought a paradigm shift from manually-defined systems to systems that are computer-trained with the help of historical data by extracting representative features from the data. ML algorithms then derive the optimal decision function. Thus, extracting discriminative features from the input medical images becomes a critical step in designing an ML system. In traditional ML frameworks, feature extraction is defined by human experts. Therefore, such features are called handcrafted features and thus the resultant system performance is subject to the developers’ expertise.
Deep learning (DL) is a specific branch of ML. It is based on artificial neural network (ANN) which mimics the multilayered human perception system. DL has further brought a shift from conventional ML systems to a class of self-taught systems that learn the features which capture optimal representation of the data for the given problem thereby eliminating the need for handcrafted features. For decades even after introduction, ANNs found less acceptance as there were severe hinderances in the training of deep architecture to solve real problems. This was mainly because of vanishing gradient and overfitting, lack of powerful computing machines, and the scarcity of enough data for training the system. However, most of these constraints have now been resolved, due to improvements in - the obtainability of data, training algorithms and power of computing using graphics processing units (GPU). DL frameworks have shown promising results in duplicating humans in several arenas, such as medical imaging. The idea of applying DL to medical imaging is a getting increasingly popular however, there are several limitations that impede its progress.
We opine that ultimately, the implementation of ML/DL-based medical image analysis tools in radiology practice at a commercial scale will take place. However, we also believe that this will not completely replace radiologists, although some specific subset of manual tasks will get replaced with complementary modules which will provide an overall augmentation of the entire medical imaging system. In this review, we discuss various state-of-the-art applications of ML and DL in medical image analysis. We also discuss some common DL frameworks, strengths of DL and the challenges to be overcome in future.
2 Machine Learning
2.1 Types of Machine Learning Algorithms
Based on system training, ML algorithms are categorized as supervised and unsupervised.
Supervised Training.
Supervised learning determines a function which reconstructs output with inference from the input which is constructed with representative numerical or nominal features vectors, comprising of independent variables and the corresponding output variable also called dependent/target variable. If the output is a numerical variable, the training method is known as regression. When the output is a categorical variable, the method is called classification.
Unsupervised Training.
Unsupervised learning performs data processing independent of labels and is trained to describe hidden patterns from unlabeled data and group the data into segments wherein each segment consists of samples which are similar as per some attributes. Dimensionality reduction methods like Principal component analysis and clustering methods like k-means are some examples of unsupervised learning algorithms.
2.2 Applications of Machine Learning in Medical Image Analysis
A large number of research papers have described applications of ML in medical imaging, such as CAD of Parkinson’s disease using brain MRI using graph‐theory‐based spectral features [4] and 3D local binary pattern features [7], CAD of Schizophrenia using functional MRI using non-linear fuzzy kernel-based feature extraction [8] and fuzzy-rough-set feature selection [9]. Some more applications have been - detection of lung nodule in CXR [11,12,13,14] and thoracic CT [13,14,15,16,17,18] diagnosis of lung nodules into malignant or benign in CXR [19], microcalcifications detection in mammography [20,21,22,23], masses detection [24] and classification of masses into benign or malignant [25,26,27] in mammography. Recently, Aggarwal et al. [28] employed surfacelet transform, a 3D multi-resolution transform for capturing complex direction details while extracting features from brain MRI volumes for CAD of Alzheimer’s disease. Further, modified and enhanced intuitionistic fuzzy-c-means clustering methods have been suggested for segmenting of human brain MRI scans [29]. A survey of works on ML in CAD has been presented [30]. Some works on ML in CAD and medical image analysis are available in the literature [10, 31].
2.3 Machine Learning Algorithms – Towards Deep Learning
A very prevalent method for classification, support vector machine (SVM), generally shows superior performance in most of the classification problems, owing to its merits of convex optimization and regularization [32, 33]. Lately, ensemble learning, including algorithms leveraging boosting and bagging, is being commonly used for enhanced classification performance [34].
Numerical or nominal representations of input data are called features in ML. Thus, the performance of the ML algorithm largely depends on the quality of features fed to it as input. Determining informative features is an important aspect in ML systems. Many domain experts attempt to learn and create handcrafted features by the help of various estimation techniques, such as performance tests and statistical analysis. In this direction, various feature extraction and selection algorithms have been established for obtaining high-performing features.
Artificial Neural network (ANN) is one of the popular algorithms for ML. ANN models the computational units of human brain neurons and their synapses. ANN comprises of layers of interconnected artificial neurons. Every neuron is an implementation of a basic classifier which generates a decision depending upon the weighted sum of inputs. The weights of the edges in the network are trained by a back-propagation algorithm, wherein tuples of input and expected output are provided, imitating the situation wherein the brain depends on external stimuli for learning about achieving certain goals [35]. ANNs had notable performance in several areas but suffered limits like trapping in the local minima while learning optimized weights of network, and overfitting on the given dataset.
3 Overview of Deep Learning
ANN is a ML framework that derives inspiration from the functioning of human neurons. However, for a long time these did not gain much success. This field of research has recently re-emerged, in the form of DL which was developed in the field of computer vision and is becoming immensely useful in several other fields. It gained immense popularity in 2012, when a convolutional neural network (CNN) won ImageNet Classification [36]. Thereafter, researchers in most domains, including medical imaging, started implementing DL. With fast and increasing improvement in computational power and the availability of massive amounts of data, DL is becoming the most preferred ML technique for implementation since it is able to learn much more high-level, abstract and complex patterns from given data than any other conventional ML techniques. Unlike conventional ML algorithms, DL methods make the feature engineering process significantly simple. Moreover, DL can even be applied directly to raw input data. Hence, this lets more researchers to explore more novel ideas. For instance, DL has been successfully applied to CAD of Alzheimer’s [37], Parkinson’s [38] and Schizophrenia [39] diseases using MRI, as well as segmentation of brain regions in MRI and ultrasound [40].
Some of the common DL frameworks are - Convolutional Neural Network (CNN), Recurrent Neural Networks (RNNs), Auto-encoders (AEs) and Stacked Auto-encoders (SAEs), among others. CNNs and RNNs are supervised frameworks while AEs and SAEs are unsupervised frameworks. These are explained as follows.
3.1 Convolutional Neural Network
A CNN is a neural network proposed by Fukushima [41] for simulating the human visual system in 1980. A CNN comprises of several layers of connected computation units (neuron-like) with step-wise processing. It has attained substantial advancements in the field of computer vision. A CNN can derive hierarchical information by pre-processing, going from edge-level features to object-representations in images.
The CNN comprises of inter-connected convolutional, pooling and fully connected layers. The input is an image, and the outputs are class categories. The main task of a convolutional layer is detecting characteristic edges, lines, objects and various visual elements. Data propagation in the forward direction is equivalent to a performing shift-invariant convolution. The outcome of convolution is composed into the respective unit in the next layer which further processes it with an activation function and computes subsequent output. The activation function is a non-linear unit like sigmoid or rectified linear unit. For capturing an increasing field of view as we move to deeper layers, feature maps computed at every layer are reduced gradually and spatially by the pooling layer wherein a maximum operation is performed over the pixels within the local region. The convolutional and pooling layers are iterated a few times. After this, fully connected layers are used for integrating the features from the whole of the feature maps. Finally, a softmax layer is generally used as the output layer. Figure 1 depicts the architecture of a CNN.
CNNs are different from simple Deep Neural Networks [42]. In CNNs the network weights are shared so that the network carries out convolution on the input images. Therefore, the network is not required to separately learn detectors for an object that occurs at multiple positions within the image. This makes the network invariant to input translations and considerably decreases the number of parameters to be learnt.
3.2 Recurrent Neural Networks
RNN is a special category of ANN well-suited for temporal data i.e. speech and text for sequence modelling. A conventional ANN assumes that all inputs (and outputs) are independent of each other. This assumption, however, is not applicable for many types of problems. For examples, in machine translation task of natural language processing –prediction of the next word in a sentence is dependent on the words preceding it. RNNs are called “recurrent” since the same computation is performed for each element of the given sequence, with the outcome being dependent on the prior computations. An alternate way to visualize RNNs is that they have a “memory” which saves information of the computations done so far and creates an internal state of the network. This way RNNs yield substantial enhancement in performance in natural language processing, speech recognition and generation, machine translation, handwriting recognition tasks [43, 44].
Although primarily introduced for 1-D input, RNNs are increasingly being used for images as well. In medical applications, RNNs have been employed successfully for segmentation [45] in the MRBrainS challenge.
3.3 Auto-encoders and Stacked Auto-encoders
AEs are networks trained with one hidden layer for reconstructing the input on the output layer. Key aspects of AEs are – (1) bias and weight matrix from input to hidden layer and (2) corresponding bias and weight from the hidden layer to the reconstructed output. A non-linear activation function is employed at the hidden layer. Further, the dimension of the hidden layer is kept smaller than the input layer dimension. This projection of input data onto a lower dimension space captures a dominant latent structure of the data.
The denoising auto-encoder (DAE) [46] is useful in preventing the network from learning a trivial solution. DAEs are trained for reconstructing the input by providing corrupted variant of the input data by adding noise (for example salt-and-pepper-noise). Initially, deep neural networks (DNNs) were difficult to efficiently train. These did not gain popularity until 2006 [47,48,49] when it was demonstrated that unsupervised training of DNNs in a layer-wise ‘greedy’ fashion (pre-training), followed by fine-tuning of the whole network in a supervised manner, could achieve better results. Stacked auto-encoder (SAE), also called deep AE, is one such popular architecture. SAEs are constructed by placing layers of AEs on top of each other.
4 Deep Learning in Medical Image Analysis
DL is particularly important for medical imaging analysis because it can take years of manual training to achieve acceptable level of domain expertise for appropriate hand-crafted feature determination required to implement traditional ML algorithm. DL has been extensively implemented in various medical image analysis tasks. The results demonstrate the immense potential of DL in achieving higher accuracies of automated systems in future. Some of these studies, carried out in majorly two application areas, have been discussed as follows:
4.1 Classification
Image classification was initially the area wherein DL made significant contribution to medical imaging. Early reports of DL-based CAD systems for breast cancer [50], lung cancer [51, 52, 58] and Alzheimer’s disease (AD) [53,54,55] demonstrate good performance in detection of the diseases. DL has been implemented in detection, diagnosis and analysis of breast cancer risk [50, 56, 57]. Some works have used DL for AD diagnosis, in which the diseases are detected using multi-modal brain data owing to effective feature computation by DL [53,54,55]. Numerous studies have surveyed the usage of DL in medicine [59,60,61].
In such settings, dataset sizes are comparatively smaller (hundreds/thousands) than those in computer vision (millions of samples). Transfer learning, therefore, becomes a natural choice for such applications. Fundamentally, transfer learning is the usage of networks pre-trained on a large dataset of natural images to attempt to work for smaller medical data. Two main transfer learning strategies are: (1) use of pre-trained network for extracting features, (2) fine-tuning a pre-trained network using medical data. Both strategies are prevalent and have been implemented extensively. The first strategy allows plugging the computed features in to image analysis system without the need to train a deep network. However, fine-tuning a pre-trained Google’s Inception v3 network on medical data attained a performance close to human expert [62]. CNNs pre-trained using natural images have astoundingly demonstrated promising results in some tasks, matching up to the performance of human experts.
Initially, the medical imaging researchers focused on network architectures like SAEs. Some studies [63,64,65] have applied SAEs for CAD of Alzheimer’s disease using brain MRI. Lately, a move towards CNNs is evident with application such as retinal imaging, brain MRI and lung computed tomography (CT) to name some.
Some researchers also train their networks from scratch rather than utilizing pre-trained networks. Menegola et al. [66] compared training from scratch to fine-tuning of pre-trained networks in their experiments and demonstrated that fine-tuning performed better on a small dataset of about a 1000 skin lesions images.
In a nutshell, DL, along with transfer learning, is capable in improving the performance of existing CAD to the level of commercial use.
4.2 Segmentation
Medical image segmentation, identifying the objects from background medical images to capturing meaningful information about their volumes and shapes, is a major challenge in medical image analysis. The potential of DL approaches has put them as a primary choice for medical image segmentation [67].
Application of CNN in segmentation of medical image has been incorporated in various studies. The overall perception is to carry out segmentation using 2D image as input and to apply 2D filters on it [68]. In an experiment, Bar et al. [69], low-level features are extracted from a model pre-trained on Imagenet. Thereafter, high-level features are extracted from PiCoDes [70] and fused together for segmentation. 2.5D approaches [71,72,73] are inspired by the fact that these have more spatial information than 2D but less computation costs than 3D. Usually, they encompass extraction of three orthogonal 2D patches in the three planes.
The benefit of a 3D CNN is to compute a powerful representation along all 3 dimensions. The 3D network is trained for predicting the label of a volume element based on the surrounding 3D patch content. The accessibility of 3D medical imaging and vast development in hardware has enabled full utilization of spatial 3D information.
5 Limitations of Deep Learning in Medical Image Analysis
Deep learning applied to medical imaging has the potential to become the most disruptive technology medical imaging has witnessed since the introduction of digital imaging. Researchers recommend that DL systems will take over humans in not only diagnosis but also in prediction of disease, prescribing medicine and guiding in treatments. Even though there are promising results from several research studies, there are a few limitations/challenges yet to be overcome before DL can become a part of mainstream radiological practice. Some of these challenges are discussed below.
5.1 Training Data
DL requires enormous training data as performance of DL classifier depends on the size and quality of the dataset to a large extent. However, scarcity of data is one the main barriers in the success of DL in medical imaging. Further, construction of large medical imaging data is a challenge since annotating the data requires much time and effort from not just single but multiple experts to rule out human error. Furthermore, annotations may not always be feasible due to non-availability of experts. Class imbalance in data is also a major challenge that is common in health sector especially in case of rare diseases wherein the number of samples in disease class are highly insufficient as compared to controls. In addition, due to the variations in prevalence of diseases and imaging machines, protocols and standards used in hospitals across the worlds, it is difficult to build generalized DL systems that can work for varied datasets from different sources.
5.2 Interpretability
The present DL methods are black-box in nature. That is, even when the DL method demonstrates excellent performance, mostly, it is impossible to derive a logical or technical explanation of the functioning of the system. This brings us to a question - whether it is acceptable to use a system which is unable to provide a reasoning for the decision it takes in a critical domain like healthcare?
5.3 Legal and Ethical Aspects
Issues may arise concerning the usage of clinical imaging data for development of DL systems at a commercial scale as subjects’ private and sensitive information is captured in their medical image data. With the growth of healthcare data, medical image analysis practitioners also face a challenge of how to anonymize patient information to avert its disclosure. Additional legal issues would arise if we implement a DL system in clinical practice, independent from intervention of a clinician. This raises a question - who would be accountable for any error that causes harm to patients?
6 Discussion and Conclusion
Currently, clinicians are experiencing a growing number of complex imaging standards in radiology. This adds to their difficulty to complete analysis on time and compose accurate reports. However, Deep Learning is showing impressive performance outcomes in a wide range of medical imaging applications. In many of the applications DL methods have mostly outperformed traditional ML methods. In future, DL is expected to support clinicians and radiologists in providing diagnosis with more accuracy at a commercial scale. However, penetration of DL in medicine is not as fast as in other real-world problems such as automated speech recognition, machine translation, object recognition etc.
Innovation and development in DL has led to disruption in ubiquitous technology space including virtual digital assistants and autonomous driving, to name some. With these ground-breaking technological advancements, it is not unreasonable to foresee that there will be critical changes in radiology in the coming years owing to DL.
In this review, we have discussed DL from the perspective of medical image analysis. The usage of DL in medical image analysis is currently at very initial stage. With the recent technological innovations, larger fully annotated databases are required to progress with DL-based progress in medical imaging. It will be essential to train the DL network, and to evaluate its performance. There are several other problems and challenges to resolve and overcome. Such as, legal and ethical issues arising due to the presence of identification related patient information present in the patient data used for developing commercial DL systems.
As we contemplate the use of DL in medical imaging, we see this revolution to be more of a collective medium in reducing the load from several repetitive tasks and increasing objective computations in presently subjective processes, rather than replacing clinical experts. The active involvement of experts is indispensable in this critical and sensitive area of medical imaging.
References
Murphy, K.P.: Machine Learning: A Probabilistic Perspective, 1st edn., p. 25. The MIT Press, Cambridge (2012)
Wang, F., Shen, D., Yan, P., Suzuki, K. (eds.): MLMI 2012. LNCS, vol. 7588. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-35428-1
Shen, D., Wu, G., Zhang, D., Suzuki, K., Wang, F., Yan, P.: Machine learning in medical imaging. Comput. Med. Imaging Graph. 41, 1–2 (2015)
Rana, B., et al.: Graph-theory-based spectral feature selection for computer aided diagnosis of Parkinson’s disease using T 1-weighted MRI. Int. J. Imaging Syst. Technol. 25(3), 245–255 (2015)
Suzuki, K., Zhou, L., Wang, Q.: Machine learning in medical imaging. Pattern Recognit. 63, 465–467 (2017)
El-Baz, A., Gimel’farb, G., Suzuki, K.: Machine learning applications in medical image analysis. Comput. Math. Methods Med. 2017, 2361061 (2017)
Rana, B., et al.: Relevant 3D local binary pattern based features from fused feature descriptor for differential diagnosis of Parkinson’s disease using structural MRI. Biomed. Signal Process. Control 34, 134–143 (2017)
Juneja, A., Rana, B., Agrawal, R.K.: fMRI based computer aided diagnosis of schizophrenia using fuzzy kernel feature extraction and hybrid feature selection. Multimed. Tools Appl. 77(3), 3963–3989 (2018)
Juneja, A., Rana, B., Agrawal, R.K.: A novel fuzzy rough selection of non-linearly extracted features for schizophrenia diagnosis using fMRI. Comput. Methods Programs Biomed. 155, 139–152 (2018)
Suzuki, K.: Computational Intelligence in Biomedical Imaging. Springer, New York (2014). https://doi.org/10.1007/978-1-4614-7245-2
Shiraishi, J., Li, Q., Suzuki, K., Engelmann, R., Doi, K.: Computer aided diagnostic scheme for the detection of lung nodules on chest radiographs: localized search method based on anatomical classification. Med. Phys. 33(7), 2642–2653 (2006)
Coppini, G., Diciotti, S., Falchini, M., Villari, N., Valli, G.: Neural networks for computer-aided diagnosis: detection of lung nodules in chest radiograms. IEEE Trans. Inf. Technol. Biomed. 7(4), 344–357 (2003)
Hardie, R.C., Rogers, S.K., Wilson, T., Rogers, A.: Performance analysis of a new computer aided detection system for identifying lung nodules on chest radiographs. Med. Image Anal. 12(3), 240–258 (2008)
Chen, S., Suzuki, K., MacMahon, H.: A computer-aided diagnostic scheme for lung nodule detection in chest radiographs by means of two-stage nodule-enhancement with support vector classification. Med. Phys. 38, 1844–1858 (2011)
Arimura, H., Katsuragawa, S., Suzuki, K., et al.: Computerized scheme for automated detection of lung nodules in low-dose computed tomography images for lung cancer screening. Acad. Radiol. 11(6), 617–629 (2004)
Armato III, S.G., Giger, M.L., MacMahon, H.: Automated detection of lung nodules in CT scans: preliminary results. Med. Phys. 28(8), 1552–1561 (2001)
Ye, X., Lin, X., Dehmeshki, J., Slabaugh, G., Beddoe, G.: Shape-based computer-aided detection of lung nodules in thoracic CT images. IEEE Trans. Biomed. Eng. 56(7), 1810–1820 (2009)
Way, T.W., Sahiner, B., Chan, H.P., et al.: Computer-aided diagnosis of pulmonary nodules on CT scans: improvement of classification performance with nodule surface features. Med. Phys. 36(7), 3086–3098 (2009)
Aoyama, M., Li, Q., Katsuragawa, S., MacMahon, H., Doi, K.: Automated computerized scheme for distinction between benign and malignant solitary pulmonary nodules on chest images. Med. Phys. 29(5), 701–708 (2002)
Wu, Y., Doi, K., Giger, M.L., Nishikawa, R.M.: Computerized detection of clustered microcalcifications in digital mammograms: applications of artificial neural networks. Med. Phys. 19(3), 555–560 (1992)
El-Naqa, I., Yang, Y., Wernick, M.N., Galatsanos, N.P., Nishikawa, R.M.: A support vector machine approach for detection of microcalcifications. IEEE Trans. Med. Imaging 21(12), 1552–1563 (2002)
Yu, S.N., Li, K.Y., Huang, Y.K.: Detection of microcalcifications in digital mammograms using wavelet filter and Markov random field model. Comput. Med. Imaging Graph. 30(3), 163–173 (2006)
Ge, J., Sahiner, B., Hadjiiski, L.M., et al.: Computer aided detection of clusters of microcalcifications on full field digital mammograms. Med. Phys. 33(8), 2975–2988 (2006)
Wu, Y.T., Wei, J., Hadjiiski, L.M., et al.: Bilateral analysis based false positive reduction for computer-aided mass detection. Med. Phys. 34(8), 3334–3344 (2007)
Huo, Z., Giger, M.L., Vyborny, C.J., Wolverton, D.E., Schmidt, R.A., Doi, K.: Automated computerized classification of malignant and benign masses on digitized mammograms. Acad. Radiol. 5(3), 155–168 (1998)
Delogu, P., Evelina Fantacci, M., Kasae, P., Retico, A.: Characterization of mammographic masses using a gradient-based segmentation algorithm and a neural classifier. Comput. Biol. Med. 37(10), 1479–1491 (2007)
Shi, J., Sahiner, B., Chan, H.P., et al.: Characterization of mammographic masses based on level set segmentation with new image features and patient information. Med. Phys. 35(1), 280–290 (2008)
Aggarwal, N., Rana, B., Agrawal, R.K.: Role of surfacelet transform in diagnosing Alzheimer’s disease. Multidimens. Syst. Signal Process. 30(4), 1839–1858 (2019)
Verma, H., Agrawal, R.K., Sharan, A.: An improved intuitionistic fuzzy c-means clustering algorithm incorporating local information for brain image segmentation. Appl. Soft Comput. 46, 543–557 (2016)
Suzuki, K.: Machine learning in computer-aided diagnosis of the thorax and colon in CT: a survey. IEICE Trans. Inf. Syst. E96-D(4), 772–783 (2013)
Suzuki, K.: Machine Learning in Computer-Aided Diagnosis: Medical Imaging Intelligence and Analysis. IGI Global, Hershey (2012)
Byvatov, E., Fechner, U., Sadowski, J., Schneider, G.: Comparison of support vector machine and artificial neural network systems for drug/nondrug classification. J. Chem. Inf. Comput. Sci. 43, 1882–1889 (2003)
Tong, S., Chang, E.: Support vector machine active learning for image retrieval. In: Proceedings of the 9th ACM International Conference on Multimedia, Ottawa, Canada, 5 October–30 September 2001, p. 107. ACM, New York (2001)
Arbib, M.A.: The Handbook of Brain Theory and Neural Networks, 2nd edn. The MIT Press, Boston (2003)
Haykin, S.S.: Neural Networks: A Comprehensive Foundation, pp. 107–116. Macmillan College Publishing, New York (1994)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097–1105 (2012)
Basaia, S., et al.: Automated classification of Alzheimer’s disease and mild cognitive impairment using a single MRI and deep neural networks. NeuroImage Clin. 21, 101645 (2019). Alzheimer’s disease neuroimaging initiative
Chen, L., Shi, J., Peng, B., Dai, Y.: Computer-aided diagnosis of Parkinson’s disease based on the stacked deep polynomial networks ensemble learning framework. Sheng wu yi xue gong cheng xue za zhi = J. Biomed. Eng. = Shengwu yixue gongchengxue zazhi 35(6), 928–934 (2018)
Zeng, L.L., et al.: Multi-site diagnostic classification of schizophrenia using discriminant deep learning with functional connectivity MRI. EBioMedicine 30, 74–85 (2018)
Milletari, F., et al.: Hough-CNN: deep learning for segmentation of deep brain regions in MRI and ultrasound. Comput. Vis. Image Understand. 1(164), 92–102 (2017)
Fukushima, K.: Neocognitron: a self organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol. Cybern. 36(4), 193–202 (1980)
Suzuki, K.: Overview of deep learning in medical imaging. Radiol. Phys. Technol. 10(3), 257–273 (2017)
Mikolov, T., Karafiát, M., Burget, L., Cernocký, J., Khudanpur, S.: Recurrent neural network based language model. In: Proceedings of the 11th Annual Conference of the International Speech Communication Association (INTERSPEECH 2010), Makuhari, Japan, 26–30 September 2010. International Speech Communication Association, pp. 1045–1048 (2010)
Gregor, K., Danihelka, I., Graves, A., Rezende, D.J., Wierstra, D.: DRAW: a recurrent neural network for image generation. In: Proceedings of the 32nd International Conference on Machine Learning (ICML 2015), Lille, France, 6–11 July 2015, pp. 1462–1471. JMLR (2015)
Stollenga, M.F., Byeon, W., Liwicki, M., Schmidhuber, J.: Parallel multi-dimensional LSTM, with application to fast biomedical volumetric image segmentation. In: Advances in Neural Information Processing Systems, pp. 2998–3006 (2015)
Vincent, P., Larochelle, H., Lajoie, I., Bengio, Y., Manzagol, P.A.: Stacked denoising autoencoders: learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res. 11, 3371–3408 (2010)
Bengio, Y., Lamblin, P., Popovici, D., Larochelle, H.: Greedy layer-wise training of deep networks. In: Advances in Neural Information Processing Systems, pp. 153–160 (2007)
Hinton, G.E., Osindero, S., Teh, Y.W.: A fast learning algorithm for deep belief nets. Neural Comput. 18(7), 1527–1554 (2006)
Hinton, G.E., Salakhutdinov, R.R.: Reducing the dimensionality of data with neural networks. Science 313(5786), 504–507 (2006)
Wang, D., Khosla, A., Gargeya, R., Irshad, H., Beck, A.H.: Deep learning for identifying metastatic breast cancer. arXiv preprint: arXiv:1606.05718, 18 June 2016
Hua, K.L., Hsu, C.H., Hidayati, S.C., Cheng, W.H., Chen, Y.J.: Computer-aided classification of lung nodules on computed tomography images via deep learning technique. OncoTargets Ther. 8, 2015–2022 (2015)
Kumar, D., Wong, A., Clausi, D.A.: Lung nodule classification using deep features in CT images. In: Proceedings of the 2015 12th Conference on Computer and Robot Vision, Halifax, Canada, 3–5 June 2015, pp. 133–138. IEEE (2015)
Suk, H.I., Lee, S.W., Shen, D.: Hierarchical feature representation and multimodal fusion with deep learning for AD/MCI diagnosis. Neuroimage 101, 569–582 (2014). Alzheimer’s disease neuroimaging initiative
Suk, H.I., Shen, D.: Deep learning-based feature representation for AD/MCI classification. Med. Image Comput. Comput. Assist. Interv. 16(Pt 2), 583–590 (2013)
Liu, S., Lis, S., Cai, W., Pujol, S., Kikinis, R., Feng, D.: Early diagnosis of Alzheimer’s disease with deep learning. In: Proceedings of the IEEE 11th International Symposium on Biodmedical Imaging, Beijing, China, 29 April–2 May 2014, pp. 1015–1018. IEEE (2014)
Cheng, J.Z., Ni, D., Chou, Y.H., Qin, J., Tiu, C.M., Chang, Y.C., et al.: Computer-aided diagnosis with deep learning architecture: applications to breast lesions in US images and pulmonary nodules in CT scans. Sci. Rep. 6, 24454 (2016)
Kallenberg, M., et al.: Unsupervised deep learning applied to breast density segmentation and mammographic risk scoring. IEEE Trans. Med. Imaging 35, 1322–1331 (2016)
Chen, J., Chen, J., Ding, H.Y., Pan, Q.S., Hong, W.D., Xu, G., et al.: Use of an artificial neural network to construct a model of predicting deep fungal infection in lung cancer patients. Asian Pac. J. Cancer Prev. 16, 5095–5099 (2015)
Liu, Y., Wang, J.: PACS and Digital Medicine: Essential Principles and Modern Practice, 1st edn. CRC Press, Boca Raton (2010)
Collins, F.S., Varmus, H.: A new initiative on precision medicine. N. Engl. J. Med. 372, 793–795 (2015)
Aerts, H.J., Velazquez, E.R., Leijenaar, R.T., Parmar, C., Grossmann, P., Carvalho, S., et al.: Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat. Commun. 5, 4006 (2014)
Esteva, A., et al.: Dermatologist-level classification of skin cancer with deep neural networks. Nature 542(7639), 115 (2017)
Suk, H.-I., Shen, D.: Deep learning-based feature representation for AD/MCI classification. In: Mori, K., Sakuma, I., Sato, Y., Barillot, C., Navab, N. (eds.) MICCAI 2013. LNCS, vol. 8150, pp. 583–590. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-40763-5_72
Suk, H.I., Lee, S.W., Shen, D.: Latent feature representation with stacked auto-encoder for AD/MCI diagnosis. Brain Struct. Funct. 220(2), 841–859 (2015). Alzheimer’s disease neuroimaging initiative
Suk, H.I., Wee, C.Y., Lee, S.W., Shen, D.: State-space model with deep learning for functional dynamics estimation in resting-state fMRI. NeuroImage 129, 292–307 (2016)
Menegola, A., Fornaciali, M., Pires, R., Avila, S., Valle, E.: Towards automated melanoma screening: exploring transfer learning schemes. arXiv preprint: arXiv:1609.01228, 5 September 2016
Hesamian, M.H., Jia, W., He, X., Kennedy, P.: Deep learning techniques for medical image segmentation: achievements and challenges. J. Digit. Imaging 32(4), 582–596 (2019)
Zhang, W., et al.: Deep convolutional neural networks for multi-modality isointense infant brain image segmentation. NeuroImage 108, 214–224 (2015)
Bar, Y., Diamant, I., Wolf, L., Greenspan, H.: Deep learning with non-medical training used for chest pathology identification. In: Medical Imaging 2015: Computer-Aided Diagnosis, vol. 9414, p. 94140V. International Society for Optics and Photonics, 20 March 2015
Bergamo, A., Torresani, L., Fitzgibbon, A.W.: PiCoDes: learning a compact code for novel-category recognition. In: Advances in Neural Information Processing Systems, pp. 2088–2096 (2011)
Moeskops, P., et al.: Deep learning for multi-task medical image segmentation in multiple modalities. In: Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W. (eds.) MICCAI 2016, Part II. LNCS, vol. 9901, pp. 478–486. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46723-8_55
Prasoon, A., Petersen, K., Igel, C., Lauze, F., Dam, E., Nielsen, M.: Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network. In: Mori, K., Sakuma, I., Sato, Y., Barillot, C., Navab, N. (eds.) MICCAI 2013, Part II. LNCS, vol. 8150, pp. 246–253. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-40763-5_31
Roth, H.R., Lu, L., Farag, A., Sohn, A., Summers, R.M.: Spatial aggregation of holistically-nested networks for automated pancreas segmentation. In: Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W. (eds.) MICCAI 2016, Part II. LNCS, vol. 9901, pp. 451–459. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46723-8_52
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Agrawal, R.K., Juneja, A. (2019). Deep Learning Models for Medical Image Analysis: Challenges and Future Directions. In: Madria, S., Fournier-Viger, P., Chaudhary, S., Reddy, P. (eds) Big Data Analytics. BDA 2019. Lecture Notes in Computer Science(), vol 11932. Springer, Cham. https://doi.org/10.1007/978-3-030-37188-3_2
Download citation
DOI: https://doi.org/10.1007/978-3-030-37188-3_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-37187-6
Online ISBN: 978-3-030-37188-3
eBook Packages: Computer ScienceComputer Science (R0)