Abstract
Diabetes has become one of the significant reasons for public sickness and death in worldwide. By 2019, diabetes had affected more than 463 million people worldwide. According to the International Diabetes Federation report, this figure is expected to rise to more than 700 million in 2040, so early screening and diagnosis of diabetes patients have great significance in detecting and treating diabetes on time. Diabetes is a multi factorial metabolic disease, its diagnostic criteria are difficult to cover all the ethology, damage degree, pathogenesis and other factors, so there is a situation for uncertainty and imprecision under various aspects of the medical diagnosis process. With the development of Data mining, researchers find that machine learning and deep learning, playing an important role in diabetes prediction research. This paper is an in-depth study on the application of machine learning and deep learning techniques in the prediction of diabetics. In addition, this paper also discusses the different methodology used in machine and deep learning for prediction of diabetics since last two decades and examines the methods used, to explore their successes and failure. This review would help researchers and practitioners understand the current state-of-the-art methods and identify gaps in the literature.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
According to the International Diabetes Federation (IDF) [1] statistics, there were 415 million people suffering from diabetes around the world in 2015. By 2040 this number is expected to rise to over 642 million, as a consequence, diabetes has become the main cause of national disease and death in most countries. Diabetes is a group of metabolic diseases in which a person has high blood glucose, either because the body does not produce enough insulin, or because the cells do not respond to the insulin that is produced [2, 3]. Most diabetes can be categorized into 3 subgroups: type 1 diabetes (T1D), type 2 diabetes (T2D), and gestational diabetes (GDM). Over the long term, T2D patients become resistant to the normal effects of insulin and gradually lose their capacity to produce enough of this hormone. A wide range of therapeutic options is available for patients with T2D. At the early stages of disease, they commonly receive medications that improve insulin secretion or insulin absorption, but eventually they must receive external doses of insulin. On the other hand, T1D patients have severe impairments in insulin production, and must use external insulin exclusively to manage their blood glucose (BG). Treatment of T1D requires consistent doses of insulin through multiple daily injections (MDIs) or continuous subcutaneous insulin infusion (CSII) using a pump. GDM is treated similarly to T2D, but only occurs during pregnancy due to the interaction between insulin and hormones released by the placenta. Figure 1 represents the statistical data of diabetics’ patients from the year 2000 onwards.
In 2000, the global estimate of adults living with diabetes was 151 million. By 2009 it had grown by 88% to 285 million. Today, 9.3% of adults aged 20–79 years – a staggering 463 million people – are living with diabetes. A further 1.1 million children and adolescents under the age of 20, live with type 1 diabetes. A decade ago, in 2010, the global projection for diabetes in 2025 was 438 million. With over five years still to go, that prediction has already been surpassed by 25 million. IDF (International Diabetes Federation) estimates that there will be 578 million adults with diabetes by 2030, and 700 million by 2045. Diabetes is one of the deadliest diseases that claim millions of lives each year. According to the WHO (World Health Organization), it was estimated that 3.4 million deaths are caused due to high blood sugar. It has been found that the over diagnosis of diabetes may lead to comorbidity like cognitive impairment, stroke, cancer, kidney problem etc. Therefore, it should be diagnosed at the earliest. In year 2000, India topped the world with 31.7 million people suffered from diabetes followed by China with second place and United States with third place [4]. It is predicted that by the year 2030 diabetes mellitus may affect up to 79.40 million people in India [5]. In last 40 years, a fourfold rise has been witnessed for this contagious disease [6]. According to International Diabetes Federation, in 2017, there are around 425 million populations suffering from diabetes across the world. It is also estimated that by 2045 the rise in the diabetic population will be increased by 32% [7]. Currently, China, India, USA, Brazil, and Russia are the top five countries with the highest rate of diabetic population. Figure 2 [8] shows the percentage of people affected by diabetes.
Data Mining and Artificial Intelligence (AI) plays an important role in the prediction of diabetes. With the continuous development of artificial intelligence and data mining technology, researchers begin to consider using machine learning and deep learning techniques to search for the characteristics of diabetes. Machine learning techniques can find implied pathogenic factors in virtue of analyzing and using diabetic data, with a high stability and accuracy in diabetes diagnosis. Therefore, machine learning techniques which can find out the reasonable threshold risk factors and physiological parameters provide new ideas for screening and diagnosis of diabetes [9]. Diabetes is a very serious disease that, if not treated properly and on time, can lead to very serious complications, including death. This makes diabetes, one of the main priorities in medical science research, which in turn generates huge amounts of data. Constantly increasing volumes of data are very well suited to be processed using data mining that can readily handle them. Using data-mining methods in diabetes research is one of the best ways to utilize large volumes of available diabetes-related data for extracting knowledge. Both descriptive (association and clustering) and predictive (classification) data-mining methods are used in the process. These data-mining methods are different from traditional statistic approaches in many ways [10].
1.1 Machine learning/deep learning and its application for diabetic prediction
Machine learning and deep learning techniques hold great promise in improving the early detection and management of diabetes, potentially leading to better patient outcomes and reduced healthcare costs.
-
Data Collection and Preprocessing: The first step involves collecting relevant data from patients, which may include demographic information (age, gender), medical history (family history of diabetes, past diagnoses), lifestyle factors (diet, exercise), and clinical measurements (blood glucose levels, blood pressure, cholesterol levels). This data needs to be preprocessed to handle missing values, normalize features, and remove noise.
-
Feature Selection and Engineering: Feature selection involves identifying the most relevant variables that contribute to the prediction of diabetes. Feature engineering may involve creating new features from existing ones or transforming the data to improve model performance.
-
Model Selection: Various machine learning algorithms can be employed for diabetic prediction, including logistic regression, decision trees, random forests, support vector machines (SVM), and neural networks. Deep learning techniques, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), can also be used to capture complex patterns in the data.
-
Model Training and Evaluation: The selected model is trained using labeled data, where the outcome (diabetes status) is known. The model's performance is evaluated using metrics such as accuracy, precision, recall, F1-score, and area under the receiver operating characteristic curve (AUC-ROC) on a separate validation dataset.
-
Deployment and Integration: Once the model is trained and evaluated, it can be deployed in clinical settings to assist healthcare providers in identifying individuals at risk of diabetes. Integration with electronic health record (EHR) systems can facilitate real-time prediction and decision-making.
-
Continuous Monitoring and Updating: As new data becomes available, the model should be periodically retrained and updated to ensure its accuracy and relevance over time. Continuous monitoring of model performance and outcomes can help improve its effectiveness in predicting diabetes and related complications.
Some challenges and considerations in diabetic prediction using machine learning and deep learning include the need for large and diverse datasets, addressing class imbalance (since the number of diabetic patients may be much smaller than non-diabetic patients), interpretability of models, and ensuring privacy and security of patient data. Healthcare data, including patient information and medical records, are sensitive and subject to strict privacy regulations such as HIPAA (Health Insurance Portability and Accountability Act) in the United States. Ensuring the privacy and security of patient data is essential when developing and deploying diabetic prediction models. Integrating ML and DL models into clinical practice requires rigorous validation to demonstrate their effectiveness and safety. The recent proliferation of data mining techniques has given rise to disease prediction systems. Specifically, with the vast amount of medical data generated every day [440].
The remainder of this paper is structured as follows. Section 2 highlights the details of different data mining technique utilized in prediction of diabetes; Section 3 mainly focuses on a detail review of diabetes prediction based on machine learning; Section 4 mainly focuses on a detail review of diabetes prediction based on deep learning; Section 5 mainly for discussion and comparission; and finally concludes the papers in section 6.
2 Data mining techniques
Both Data Mining and Machine learning are areas which have been inspired by each other, though they have many things in common, yet they have different ends. Figure 3 represents the relationship between the machine learning and data mining. Data mining, machine learning, artificial intelligence, and statistics are closely related fields that share common goals and methodologies for analyzing and extracting insights from data. They complement each other and are often used together in various applications, such as predictive modeling, pattern recognition, data visualization, and decision support.
Artificial Intelligence can enable the computer to think. The computer is made much more intelligent by AI. Machine learning is the subfield of AI study. Various researchers think that without learning, intelligence cannot be developed. There are many types of Machine Learning Techniques that are shown in Fig. 4. Supervised, Unsupervised, Semi Supervised, Reinforcement, Evolutionary Learning and Deep Learning are the types of machine learning techniques. These techniques are used to classify the data set [11]. Both supervised and unsupervised learning techniques are used depending on the nature of the data and the specific problem being addressed. Additionally, semi-supervised learning techniques combine elements of both paradigms by leveraging a small amount of labeled data along with a larger pool of unlabeled data. Reinforcement learning is a powerful paradigm in machine learning that enables agents to learn optimal behavior through trial and error, interaction with the environment, and feedback in the form of rewards.
2.1 Supervised learning
A supervised learning technique is used when the historical data is available for a certain problem. The system is trained with the inputs and respective responses and then used for the prediction of the response of new data [12]. Classification and regression are the types of Supervised Learning.
2.1.1 Classification
It gives the prediction of Yes or No, for example, “Is this tumour cancerous?”, and “Does this cookie meet our quality standards?” Common classification approaches include artificial neural network, back propagation, decision tree, support vector machines, Naive Bayes classifier, K-Nearest Neighbors (K-NN), Random forest [12]. Classification is used to classify data into predefined categorical class labels. “Class” in classification, is the attribute or feature in a data set, in which users are most interested. It is defined as the dependent variable in statistics. To classify data (or records), a classification algorithm creates a classification model consisting of classification rules. For example, banks have constructed classification models to categorize the bank loan and mortgage applications into risky or safe. In the medical field, classification can be used to help define medical diagnosis and prognosis based on symptoms and health conditions.
Support Vector Machine (SVM)
Support vector machine (SVM) is used in both classification and regression. An SVM classifier, a concept by Vladimir Vapnik, finds the optimal separating hyperplane between positive and negative classes of data. The optimal hyperplane is the one that gives maximum margin between the training examples that lie closest to the hyperplane and the data points on the two sides belong to different classes. In linear SVM the given data set is considered as p-dimensional vector that can be separated by maximum of p-1 planes called hyper-planes. These planes separate the data space or set the boundaries among the data groups for classification or regression problems. The best hyper-plane can be selected among the number of hyper-planes on the basis of distance between the two classes it separates. The plane that has the maximum margin between the two classes is called the maximum-margin hyper-plane. It can handle nonlinear classification tasks efficiently by mapping the samples into a higher dimensional feature space by using a nonlinear kernel function. Since the SVM approach is data-driven and model free, it has important discriminating power for classification. SVM algorithms use a set of mathematical functions that are defined as the kernel. The function of the kernel is to take data as input and transform it into the required form. Different SVM algorithms use different types of kernel functions. These functions can be different types. For example linear, nonlinear, polynomial, radial basis function (RBF), and sigmoid. The most used type of kernel function is RBF [442]. Because it has localized and finite response along the entire x-axis. The kernel functions return the inner product between two points in a suitable feature space. Both SVM and RF are widely used for classification tasks, including the classification of diabetic patients based on various features such as medical history, clinical measurements, and demographic information. Both algorithms can provide insights into feature importance. SVM determines the support vectors, which are the data points closest to the decision boundary, while RF calculates feature importance based on how much each feature contributes to the model's predictive performance. In diabetes research, this can help identify the most relevant features for predicting diabetes or assessing disease progression.
Decision Tree (DT)
Decision tree (DT) is a supervised learning that can be used as a regression tree while the response variable is continuous and as a classifcation tree while the response variable is categorical. Whereas the input variables are any types, as like graph, text, discrete, continuous, and so on in the case of both regression and classification. The finding of a solution with the help of decision trees starts by preparing a set of solved cases. The whole set is then divided into 1) a training set, which is used for the induction of a decision tree, and 2) a testing set, which is used to check the accuracy of an obtained solution [443]. A decision tree is a tree structure based model which describes the classification process based on input features.
The steps of DT as follows:
-
Construct a tree with its nodes as input features.
-
Select the feature to predict the output from the input features whose gives the highest information gain.
-
Repeat the above steps to form sub trees based on features which was not used in the above nodes.
The decision tree is the most powerful and popular tool for classification and prediction. A Decision tree is a flowchart like tree structure, where each internal node denotes a test on an attribute, each branch represents an outcome of the test, and each leaf node (terminal node) holds a class label. Decision trees classify instances by sorting them down the tree from the root to any leaf node, which provides the classification of the instance.
Naive Bayes Classifier (NB)
Naive Bayes classifier is a well-known type of classifiers, i.e., of programs that assign a class from a predefined set to an object or case under consideration based on the values of descriptive attributes. A well-known Bayesian network classifier is the Naïve Bayes’ classifier is a probabilistic classifier based on the Bayes’ theorem, considering Naïve (Strong) independence assumption [444]. the stage of the calculation Naive Bayes as follows:
-
Find the value of prior probability for each class calculate the average of each class.
-
Find the value of the likelihood that is a process of calculating the probability of each attribute against the class, the possibility of the emergence of a class when an attribute is selected.
-
Find the value of the posterior that is result of calculation likelihood in the form of the probability of the attribute class, calculated to divert the possibility of the attribute of the input with the class, in the process of this can be the probability of the end.
It is a classification technique based on Bayes’ Theorem with an assumption of independence among predictors. In simple terms, a Naive Bayes classifier assumes that the presence of a particular feature in a class is unrelated to the presence of any other feature. For example, a fruit may be considered to be an apple if it is red, round, and about 3 inches in diameter. Even if these features depend on each other or upon the existence of the other features, all of these properties independently contribute to the probability that this fruit is an apple and that is why it is known as ‘Naive’. The naive Bayes model is easy to build and particularly useful for very large data sets. Along with simplicity, Naive Bayes is known to outperform even highly sophisticated classification methods.
K-Nearest Neighbors (KNN)
KNN algorithms are supervised non-parametric learning algorithms that learn the relationship between input and output observations. It is simply based on the idea that “objects that are ‘near’ each other will also have similar characteristics. Thus if you know the characteristic features of one of the objects, you can also predict it for its nearest neighbour.” k-NN is an improvisation over the nearest neighbour technique. It is based on the idea that any new instance can be classified by the majority vote of its ‘k’ neighbours, - where k is a positive integer, usually a small number [445].
The criterion is defined by Euclidean distance; and if the two locations contain O1 = {x11, x12, x13, ………. ., x1n} and O2 = {x21, x22, x23, ………. ., x2n} then the Euclidean distance between them is defined according to Eq. (1).
The algorithm is based on the distance between two instances, which represents their similarity. KNN identifies k instances in the training set and then classifies a new instance based on how similar (i.e., near) it is to its neighbors. Generally speaking, a new instance is classified by a majority vote of its neighbors. Thus, when the algorithm is used for classification purposes, the output is the class membership of the new instance.
The hyperparameter k is a user-defined positive odd integer, typically small. If k = 1, the algorithm considers the neighbor nearest to the unclassified instance. If k = 3, then KNN compares the distances of the three neighbors nearest to the unclassified instance.
Random Forest (RF)
A random forest classifier is the assembly of tree-structured classifiers. This algorithm supplements the objects from array of input to every tree of the forest. The elements of the unit vector are individually voted for classification by every single tree. The forest filters the most voted classifications out of the forest. The simplest random forest with random features is formed by selecting at random, at each node, a small group of input variables to split on. Grow the tree using CART methodology to maximum size and do not prune. Random forest is a machine learning methodology for classification, which is commonly used in computational biology fields. Independently trained decision trees are merged in a random forest, which is done by subsets randomly sampled with replacement from the training data. Every branch of decision tree discovers a best feature in the training time. The best feature randomly chosen from a subset of feature space. Because trees are trained in subset of feature space and training data, they should not be produced with post-pruning [446] .The prediction of RF is the average or a majority vote of all tree predictions that have been trained. Random forest algorithm has three parameters that should be set in training time. These parameters are the number of growing trees, the minimum node size to split and the number of features to select randomly for each split. RF reduces the degree of over-fitting by combining multiple overfit evaluators (ie, decision trees) to form an ensemble learning algorithm. Each decision tree can get the corresponding classifcation decision result. By using the voting results of each decision tree in the forest, the category of the sample to be tested is determined according to the principle of minority obeying the majority, and the category with higher votes in all decision trees was determined to be the final result.
One of the biggest advantages of random forest is its versatility. It can be used for both regression and classification tasks, and it’s also easy to view the relative importance it assigns to the input features. Random forest is also a very handy algorithm because the default hyperparameter it uses often produce a good prediction result. Understanding the hyperparameter is pretty straightforward, and there's also not that many of them. One of the biggest problems in machine learning is overfitting, but most of the time this won’t happen thanks to the random forest classifier. If there are enough trees in the forest, the classifier won’t overfit the model. The main limitation of random forest is that a large number of trees can make the algorithm too slow and ineffective for real-time predictions [447]. In general, these algorithms are fast to train, but quite slow to create predictions once they are trained. A more accurate prediction requires more trees, which results in a slower model. In most real-world applications, the random forest algorithm is fast enough but there can certainly be situations where run-time performance is important and other approaches would be preferred. And, of course, random forest is a predictive modeling tool and not a descriptive tool, meaning if you're looking for a description of the relationships in your data, other approaches would be better.
2.1.2 Regression
Regression analysis consists of a set of machine learning methods that allow us to predict a continuous outcome variable (y) based on the value of one or multiple predictor variables (x). Briefly, the goal of regression models is to build a mathematical equation that defines y as a function of the x variables. Next, this equation can be used to predict the outcome (y) on the basis of new values of the predictor variables (x). Linear regression is the most simple and popular technique for predicting a continuous variable. It assumes a linear relationship between the outcome and the predictor variables.
The linear regression equation can be written as y = b0 + b*x + e, where b0 is the intercept, b is the regression weight or coefficient associated with the predictor variable x and e is the residual error. When it has multiple predictor variables, say x1 and x2, the regression equation can be written as y = b0 + b1*x1 + b2*x2 +e. In some situations, there might be an interaction effect between some predictors that is for example, increasing the value of a predictor variable x1 may increase the effectiveness of the predictor x2 in explaining the variation in the outcome variable [448].
In some cases, the relationship between the outcome and the predictor variables is not linear. In these situations, it needs to build a non-linear regression, such as polynomial and spline regression. Regression is a supervised learning technique which helps in finding the correlation between variables and enables us to predict the continuous output variable based on the one or more predictor variables. It is mainly used for prediction, forecasting, time series modeling, and determining the causal-effect relationship between variables.
In Regression, plot a graph between the variables which best fits the given data points, using this plot, the machine learning model can make predictions about the data. In simple words, "Regression shows a line or curve that passes through all the data points on target-predictor graph in such a way that the vertical distance between the data points and the regression line is minimized." The distance between data points and the line tells whether a model has captured a strong relationship or not. The most common regression techniques are: Linear regression (LR), Logistic regression, Polynomial regression, support vector regression, Decision tree regression and Random forest regression.
2.2 Unsupervised learning
Unsupervised Learning is a machine learning technique in which the users do not need to supervise the model. Instead, it allows the model to work on its own to discover patterns and information that was previously undetected. It mainly deals with the unlabeled data. It allows users to perform more complex processing tasks compared to supervised learning, whereas unsupervised learning can be more unpredictable compared with other natural learning methods. Unsupervised learning algorithms include clustering, association rule, neural networks, etc.
2.2.1 Clustering
Clustering is the task of dividing the population or data points into a number of groups such that data points in the same groups are more similar to other data points in the same group and dissimilar to the data points in other groups. It is basically a collection of objects on the basis of similarity and dissimilarity between them. Clustering is very much important as it determines the intrinsic grouping among the unlabeled data present. There are no criteria for a good clustering. It depends on the user, what are the criteria they may use which satisfy their needs [449] . There are different forms of clustering, which is explained below.
-
Density-Based Methods: These methods are considered the clusters as the dense region having some similarity and different from the lower dense region of the space. These methods have good accuracy and the ability to merge two clusters. Example DBSCAN (Density-Based Spatial Clustering of Applications with Noise), OPTICS (Ordering Points to Identify Clustering Structure) etc.
-
Hierarchical Based Methods: The clusters formed in this method form a tree-type structure based on the hierarchy. New clusters are formed using the previously formed one. It is divided into two categories Agglomerative (bottom up approach) and Divisive (top down approach).
2.2.2 Association rule
Association rule mining finds interesting associations and relationships among large sets of data items. This rule shows how frequently an item set occurs in a transaction. A typical example is Market Based Analysis. Association rule learning is a type of unsupervised learning technique that checks for the dependency of one data item on another data item and maps accordingly so that it can be more profitable. It tries to find some interesting relations or associations among the variables of the dataset. It is based on different rules to discover the interesting relations between variables in the database [450]. To measure the associations between thousands of data items, there are several metrics. These metrics are given below:
-
Support: Support is the frequency of A or how frequently an item appears in the dataset. It is defined as the fraction of the transaction T that contains the item set X. If there are X datasets, then for transactions T, it can be written as\(Supp\ (x)=\frac{freq(x)}{T}\).
-
Confidence: Confidence indicates how often the rule has been found to be true. Or how often the items X and Y occur together in the dataset when the occurrence of X is already given. It is the ratio of the transaction that contains X and Y to the number of records that contain X.
-
Lift: It is the strength of any rule. It is the ratio of the observed support measure and expected support if X and Y are independent of each other. It has three possible values: If Lift= 1: The probability of occurrence of antecedent and consequent is independent of each other. If lift>1: It determines the degree to which the two item sets are dependent on each other. If lift<1: It tells us that one item is a substitute for other items, which means one item has a negative effect on another. Apriori Algorithm is the most common association technique used in machine learning application.
2.3 Semi-supervised learning
Semi-supervised learning is the type of machine learning that uses a combination of a small amount of labeled data and a large amount of unlabelled data to train models. This approach to machine learning is a combination of supervised machine learning, which uses labeled training data, and unsupervised learning, which uses unlabelled training data. Semi-supervised machine learning is a combination of supervised and unsupervised learning. The basic procedure involved is that first, the developer will cluster similar data using an unsupervised learning algorithm and then use the existing labeled data to label the rest of the unlabelled data [451].
2.4 Reinforcement learning
Reinforcement learning differs from the supervised learning in a way that in supervised learning the training data has the answer key with it so the model is trained with the correct answer itself whereas in reinforcement learning, there is no answer but the reinforcement agent decides what to do to perform the given task. In the absence of a training dataset, it is bound to learn from its experience [452].
2.5 Evolutionary learning
Evolutionary algorithms are a heuristic-based approach to solving problems that cannot be easily solved in polynomial time, such as classically NP-Hard problems, and anything else that would take far too long to exhaustively process. When used on their own, they are typically applied to combinatorial problems; however, genetic algorithms are often used in tandem with other methods, acting as a quick way to find a somewhat optimal starting place for another algorithm to work off of [453].
2.6 Deep learning
Deep learning is an artificial intelligence (AI) function that imitates the workings of the human brain in processing data and creating patterns for use in decision making. Deep learning is a subset of machine learning in artificial intelligence that has networks capable of learning unsupervised from data that is unstructured or unlabelled also known as deep neural learning or deep neural network [454]. It has the following characteristics.
-
Deep learning is an AI function that mimics the workings of the human brain in processing data for use in detecting objects, recognizing speech, translating languages, and making decisions.
-
Deep learning AI is able to learn without human supervision, drawing from data that is both unstructured and unlabelled.
-
Deep learning, a form of machine learning, can be used to help detect fraud or money laundering, among other functions.
The different types of neural networks in deep learning are convolutional neural networks (CNN), recurrent neural networks (RNN), artificial neural networks (ANN), etc. The neural network has been widely used to train predictive models for applications such as image processing, disease prediction, and face recognition [439].
2.6.1 ANN
Artificial Neural Network, or ANN, is a group of multiple perceptrons/ neurons at each layer. ANN is also known as a Feed-Forward Neural network because inputs are processed only in the forward direction. ANN consists of 3 layers – Input, Hidden and Output. The input layer accepts the inputs, the hidden layer processes the inputs, and the output layers produce the result. Essentially, each layer tries to learn certain weights. In a neural network, one neuron to the other neuron connection exists with some strength known as weight or synaptic weight. The neural network consists of feedback, and information can flow from the input layer to the output layer via one or more hidden layers, and vice versa known as a feedback neural network [455].
2.6.2 CNN
A Convolutional Neural Network (CNN) is a Deep Learning algorithm which can take in an input image, assign importance (learnable weights and biases) to various aspects/objects in the image and be able to differentiate one from the other. The pre-processing required in a CNN is much lower as compared to other classification algorithms. While in primitive methods, filters are hand-engineered, with enough training, CNN has the ability to learn these filters/characteristics. Convolutional neural network (CNN) is one of the most popular and used of DL networks . Because of CNN, DL is very popular nowadays. The main advantage of CNN compared to its predecessors is that it automatically detects the significant features without any human supervision which made it the most used [456].
2.6.3 RNN
In a feed-forward neural network, the information only moves in one direction — from the input layer, through the hidden layers, to the output layer. The information moves straight through the network and never touches a node twice.
Feed-forward neural networks have no memory of the input they receive and are bad at predicting what’s coming next. Because a feed-forward network only considers the current input, it has no notion of order in time. It simply can’t remember anything about what happened in the past except its training. In a RNN the information cycles through a loop. When it makes a decision, it considers the current input and also what it has learned from the inputs it received previously [457].
A usual RNN has a short-term memory. In combination with an LSTM they also have a long-term memory. A recurrent neural network, however, is able to remember those characters because of its internal memory. It produces output, copies that output and loops it back into the network. Therefore, a RNN has two inputs: the present and the recent past. This is important because the sequence of data contains crucial information about what is coming next, which is why a RNN can do things other algorithms can’t. A feed-forward neural network assigns, like all other deep learning algorithms, a weight matrix to its inputs and then produces the output. RNNs apply weights to the current and also to the previous input. Furthermore, a recurrent neural network will also tweak the weights for both through gradient descent and back propagation through time (BPTT).
3 Review based on machine learning technique in diabetes prediction
3.1 Supervised learning (Classification)
3.1.1 Support vector machine (SVM)
Diabetic retinopathy has become a common eye disease in most developed countries. It occurs in 80% of all diabetic cases and is the leading cause of blindness [13]. Regular screening is the most efficient way of reducing the preventable eye damages. There are two kinds of symptoms in the diabetic retinopathy. One is dark lesion that includes hemorrhages, and microaneurysms. The other is bright lesion such as exudates and cotton wool spots. Microaneurysms are commonly detected in the retinal fluorescein angiography [14, 15].
In 2005, Zhang and Chutatape [16] introduced an SVM approach for detection of hemorrhages in background diabetic retinopathy. This paper focuses on the detection of hemorrhages which have "dot" and "blot" configurations in the background diabetic retinopathy with their color similar to the blood vessels. In this paper, a top-down strategy is applied to detect hemorrhages. The SVM classifier uses features extracted by combined 2DPCA (Two-Dimensional Principal Component Analysis) instead of explicit image features as the input vector. After locating the hemorrhages in the ROI (Region of Interest), the boundaries of the hemorrhages can be accurately segmented by the post-processing stage. The paper demonstrates a new implementation of various techniques on the problem and shows the improvement it offers over the others. Combined 2DPCA is proposed and virtual SVM is applied to achieve higher accuracy of classification. The test result demonstrates that the TP (True Positive) rate of SVM is 89.1%, while that of ANN is 84.6% at FP rate of two FP per image. The Gaussian kernel is used in SVM. The SVM based on SRM (Structural Risk Minimization) appears to be superior to ANN that employs ERM (Empirical Risk Minimization). It also compared the performance of SVM with VSVM (Virtual SVM) and found that classification accuracy of VSVM that uses the rotation invariance and illuminance invariance is better than SVM. When number of FP remains 2 per image, the TP rate of VSVM is 94% while the TP of SVM is 93.2%.
In 2006, Stoean et al. [17] proposed an ESVM (Evolutionary support vector machine) technique for diagnosis of diabetes mellitus. The main aim of this paper is to validate the new paradigm of evolutionary support vector machines (ESVMs) for binary classification also through an application to a real world problem, i.e. the diagnosis of diabetes mellitus. Different algorithms like (CPLEX (COmmercial Solvers for Integer Programming and Mathematical Programming by Linear Programming Extensions), SVM light, Active SVM, and Critical SVM) have been utilized for experimental evaluation and compare the performance with ESVM. The test result depicts that proposed technique offers a good enough accuracy in comparison to the state-of-the-art classical approaches and to the standard SVM formulation. Possibly, application of parameter tuning methods like SPO on ESVMs with a polynomial kernel would lead to better values for the evolutionary parameters that would improve the proportion of self-determined training errors. The proposed method achieves training accuracy of 77.95%, whereas test accuracy is 80.22%.
In 2008, Balakrishnan et al. [18] introduced a feature selection approach for finding an optimum feature subset that enhances the classification accuracy of the Naive Bayes classifier. Experiments were conducted on the Pima Indian Diabetes Dataset to assess the effectiveness of our approach. The results confirm that SVM Ranking with Backward Search approach leads to promising improvement in feature selection and enhances classification accuracy. Polat et al. [19] proposed a new cascade learning system based on Generalized Discriminant Analysis and Least Square Support Vector Machine. The proposed system consists of two stages. The first stage, used Generalized Discriminant Analysis to discriminant feature variables between healthy and patient (diabetes) data as a pre-processing process. The second stage used LS-SVM in order to classification of diabetes dataset. While LS-SVM obtained 78.21% classification accuracy, using 10-fold cross validation, the proposed system called GDA–LS-SVM obtained 82.05% classification accuracy using 10-fold cross validation.
In 2009, WU et al. [20] developed a semi-supervised based learning method (LapSVM) for diabetes disease diagnosis. Firstly, LapSVM was trained as a fully-supervised learning classifier to predict diabetes dataset and 79.17% accuracy was obtained. Then, it was trained as a semi-supervised learning classifier and got the prediction accuracy 82.29%. The obtained accuracy 82.29% is higher than other previous reports. The experiments led to the finding that LapSVM offers a very promising application, i.e., LapSVM can be used to solve a fully-supervised learning problem by solving a semi-supervised learning problem. The result suggests that LapSVM can be of great help to physicians in the process of diagnosing diabetes disease and it could be a very promising method in the situations where a lot of data are not class-labelled.
In 2010, Yu et al. [21] develop and validate SVM models for two classification schemes: Classification Scheme I (diagnosed or undiagnosed diabetes vs. pre-diabetes or no diabetes) and Classification Scheme II (undiagnosed diabetes or prediabetes vs. no diabetes). The SVM models were used to select sets of variables that would yield the best classification of individuals into these diabetes categories. The overall discriminative ability of classification Schemes I and II are represented by their AUC values (83.47% and 73.18%, respectively). Barakat et al. [22] proposed a support vector machine (SVM) for the diagnosis of diabetes. In particular, use an additional explanation module, which turns the “black box” model of an SVM into an intelligible representation of the SVM’s diagnostic (classification) decision. Result in a real-life diabetes dataset shows that intelligible SVMs provide a promising tool for the prediction of diabetes, where a comprehensible ruleset have been generated, with prediction accuracy of 94%, sensitivity of 93%, and specificity of 94%.
In 2011, Calisir et al. [23] proposed an automatic diagnosis system for diabetes on Linear Discriminant Analysis (LDA) and Morlet Wavelet Support Vector Machine Classifier: LDA–MWSVM is introduced. The Linear Discriminant Analysis (LDA) is used to separate features variables between healthy and patient (diabetes) data in the first stage. The healthy and patient (diabetes) features obtained in the first stage are given to inputs of the MWSVM classifier in the second stage. Finally, in the third stage, the correct diagnosis performance of this automatic system based on LDA–MWSVM for the diagnosis of diabetes is calculated by using sensitivity and specificity analysis, classification accuracy, and confusion matrix, respectively. The classification accuracy of this system was obtained at about 89.74%.Gupta et al. in [24] present a study aimed to do the performance analysis of several data mining classification techniques using three different machine learning tools over the healthcare datasets. In this study, different data mining classification techniques have been tested on four different healthcare datasets. The standards used are percentage of accuracy and error rate of every applied classification technique. The experiments are done using the 10 fold cross validation method. A suitable technique for a particular dataset is chosen based on highest classification accuracy and least error rate. The test result based on PIMA Indian Diabetes dataset show that an accuracy rate of 96.74% and 3.18% of the error rate is achieved using SVM technique which is superior when contrasted with another technique.
In 2020, Xue et al. [88] proposed an automatic diagnosis system for diabetes using supervised machine-learning algorithms like Support Vector Machine (SVM), Naive Bayes classifier and LightGBM to train on the actual data of 520 diabetic patients and potential diabetic patients aged 16 to 90. Although the naive Bayes classifier is the most popular classification algorithm, the final accuracy rate on the given data set is only 93.27%. SVM has the highest accuracy rate, with an accuracy rate of 96.54%. The accuracy of LightGBM is only 88.46%. It is found that SVM has the highest accuracy through the confusion matrix evaluation test.
In 2021, Chaves et al. [91] introduce a comparative study of data mining techniques for early diagnosis of diabetes. We use a publicly accessible data set containing 520 instances, each with 17 attributes. Naive Bayes, Neural Network, AdaBoost, k-Nearest Neighbors, Random Forest and Support Vector Machine methods have been tested. The results suggest that Neural Networks should be used for diabetes prediction. The proposed model presents an AUC of 98.3% and 98.1% accuracy, an F1-Score, Precision and Sensitivity of 98.4% and a Specificity of 97.5%. In the first experiment, author applied the Naive Bayes classifier, which correctly predicted 452 instances out of 520, a success rate of 86.92%. In the second experiment, author applied the Neural Network classifier, which correctly predicted 510 instances out of 520, a success rate of 98.08%. In the third experiment, author applied the AdaBoost classifier, which correctly predicted 506 instances out of 520, a success rate of 97.31%. In the fourth experiment, author applied the kNN classifier, which correctly predicted 506 instances out of 520, a success rate of 97.31%.In the fifth experiment, author applied the Random Forest classifier, which correctly predicted 504 instances out of 520, a success rate of 96.92%.In the last experiment, author applied the SVM classifier, which correctly predicted 505 instances out of 520, a success rate of 97.12%.
In 2022, Li et al. [408] proposed an effective biomarkers for an efficient diagnosis of type 2 diabetes. The sensitivity and specificity of the SVM model for identifying patients with type 2 diabetes were 100%, with an area under the curve of 1 in the training as well as the validation dataset. In 2023, Lei et al. [441] propose a publicly verifiable and secure SVM classification scheme (PVSSVM) for cloud-based health monitoring services. It utilize homomorphic encryption and secret sharing to protect the model and data confidentiality in the cloud server, respectively. Based on a multi-server verifiable computation framework, PVSSVM achieves public verification of predicted results. The proposed scheme achieves a reduction of approximately 83.71% in computation overhead through batch verification, as compared to one-by-one verification. Table 1 represent the related work on the diagnosis of diabetes based on SVM algorithm.
3.1.2 Decision Tree (DT)
In 2002, Breault et al. [102] introduce a classification tree approach in Classification and Regression Trees (CART) with a binary target variable of HgbA1c >9.5 and 10 predictors: age, sex, emergency department visits, office visits, comorbidity index, dyslipidemia, hypertension, cardiovascular disease, retinopathy, end-stage renal disease. The first level of the tree shows that just dividing people using an age cut-point of 65.581 years of age, 19.4% of younger people (n ¼ 3987) have a bad HgbA1c. This is 2.8 times the rate of bad HgbA1c values in those who are older (7.0%, n ¼ 3966). The dataset contains the data from 442 bed tertiary care hospital, a 500 physician multi-specialty clinic in 25 locations.
In 2004, Haung [105] investigated the potential for data mining in order to spot trends in the data and attempt to predict outcome. Feature selection has been utilized to enhance the efficiency of the data mining algorithm. Decision Tree (C4.5) is utilized in this work and results show that before feature selection, discretized C4.5 had the best performance of classification. And after feature selection C4.5 obtained the best result. Diabetic data has been collected from the Ulster community and trust hospital for the year 2000 to 2004. The dataset contained 2017 type-2 diabetic patients’ clinical information having 1124 males and 893 females.
In 2008, Liou et al. [109] proposed to detect fraudulent or abusing the reporting by health care providers using their invoice for diabetic outpatient services. The proposed work is validated in Taiwan’s National health insurance system and three kinds of a data mining algorithm like decision tree, logistic regression and neural network have been applied for this proposed work. The experimental result shows that the correct identification rate of decision tree based algorithm (99%) outperforms than the logistic regression model (92%) and neural network model (96%).
In 2010, Patil et al. [112] developed a Hybrid Prediction Model (HPM) model which uses Simple K-means clustering algorithm aimed at validating chosen class label of the given data (incorrectly classified instances are removed, i.e. pattern extracted from original data) and subsequently applying the classification algorithm to the result set. C4.5 algorithm is used to build the final classifier model by using the k-fold cross-validation method. The Pima Indians diabetes data were obtained from the University of California at Irvine (UCI) machine learning repository datasets. The proposed HPM obtained a classification accuracy of 92.38%. In order to evaluate the performance of the proposed method, sensitivity and specificity performance measures that are used commonly in medical classification studies were used.
In 2012, Kelarev et al. [116] introduced detection and monitoring of cardiovascular autonomic neuropathy, CAN, in diabetes patients. Using a small set of features identified previously, this work consists of empirical investigation and comparison of several ensemble methods based on decision trees for a novel application of the processing of sensor data from diabetes patients for pervasive health monitoring of CAN. The experiments relied on an extensive database collected by the Diabetes Complications Screening Research Initiative at Charles Sturt University and concentrated on the particular task of the detection and monitoring of cardiovascular autonomic neuropathy. The best outcomes have been obtained by the novel combined ensemble of AdaBoost (accuracy=94%) and Bagging (accuracy=92.99%) based on J48.
In 2014, Kaur et al. [129] proposed an improved J48 algorithm for the prediction of diabetics. In this proposed work, the modified J48 classifier is used to increase the accuracy rate of the data mining procedure. The data mining tool WEKA has been used as an API of MATLAB for generating the J-48 classifiers. Experimental results showed a significant improvement over the existing J-48 algorithm. Proposed algorithm has large accuracy difference than other algorithms. It has accuracy rate of 99.87% rather than others that show maximum of 77.21%accuracy. The experiment is carried out at Pima Indians diabetes dataset.
In 2018, Zou et al. [155] introduced a decision tree, random forest and neural network to predict diabetes mellitus. The dataset is the hospital physical examination data in Luzhou, China. It contains 14 attributes. In this study, five-fold cross validation was used to examine the models. In order to verity the universal applicability of the methods, chose some methods that have the better performance to conduct independent test experiments. It randomly selected 68994 healthy people and diabetic patients’ data, respectively as training set. Due to the data unbalance, it randomly extracted 5 times data. And the result is the average of these five experiments. In this study, it used principal component analysis (PCA) and minimum redundancy maximum relevance (mRMR) to reduce the dimensionality. The results showed that prediction with random forest could reach the highest accuracy (ACC = 0.8084) when all the attributes were used.
In 2020, Pei et al. [167] proposed a J48 decision tree based diabetic prediction system for chinesh people. A total of 10,436 participants who had a health check-up from January 2017 to July 2017 were recruited. With appropriate data mining approaches, 3454 participants remained in the final dataset for further analysis. Seventy percent of these participants (2420 cases) were then randomly allocated to either the training dataset for the construction of the decision tree or the testing dataset (30%, 1034 cases) for evaluation of the performance of the decision tree. The proposed approach achieved an accuracy of classification of 90.3% with a precision of 89.7% and a recall of 90.3%. Table 2 represent the related work on diagnosis of diabetes based on Decision Tree (DT) algorithm.
3.1.3 K nearest neighbour (KNN)
In 2010, Lee et al. [175] proposed a monitoring and advisory system for diabetes patient management using a Rule-Based method and KNN. This paper proposes a system that can provide appropriate management for diabetes patients, according to their blood sugar level. The system is designed to send the information about the blood sugar levels, blood pressure, food consumption, exercise, etc., of diabetes patients, and manage the treatment by recommending and monitoring food consumption, physical activity, insulin dosage, etc., so that the patient can better manage their condition. The system is based on rules and the K Nearest Neighbor (KNN) classifier algorithm, to obtain the optimum treatment recommendation. Also, a monitoring system for diabetes patients is implemented using Web Services and Personal Digital Assistant (PDA) programming.
In 2015, Farahmandian et al. [181] introduced a case study on data mining algorithms which are crucial in the diagnosis and prediction of diabetes. In this work Support Vector Machine (SVM), K Nearest Neighbors (KNN), Naïve Bayes, ID3, C4.5, C5.0, and CART algorithms are used. Evaluation and conclusion of data mining algorithms which contain 768 records of different patients have been carried out on Pima dataset. Results have shown that the degree of Accuracy in SVM algorithm equals to 81.77.
In 2018, Dey et al. [186] implement a Web Application to Predict Diabetes Disease using Machine Learning Algorithm. This work consists of development of an architecture which has the capability to predict where the patient has diabetes or not. The main aim of this exploration is to build a web application based on the higher prediction accuracy of some powerful machine learning algorithm. It used a benchmark dataset namely Pima Indian which is capable of predicting the onset of diabetes based on diagnostics manner. With an accuracy of 82.35% prediction rate Artificial Neural Network (ANN) shows a significant improvement of accuracy. The proposed model achieved an accuracy of 66.5% using KNN.
In 2020, Gupta et al. [195] introduced a Performance enhancement of diabetes prediction by finding optimum K for KNN classifier with feature selection method. The proposed work KNN and machine learning methods are used in the prediction model to classify whether the patient is diabetic or non-diabetic. The PIMA diabetes dataset is used for research purpose in the python implemented model. A research study has been performed to improve the performance of the KNN classifier by using a feature selection method, normalization and considering the different number of neighbors. The performance of classifier is measured based on different metrics such as accuracy, precision, sensitivity, specificity, f1 score and error rate. The best performance of KNN is achieved when no of neighbors (K) is 33, 40 or 45. The accuracy and error rate is same on these K and it is 87.01% and 12.99 % respectively, while a little variation is shown in other metric’s values.
In 2018, Sarkar et al. [197] proposed a K-Nearest Neighbor Learning based Diabetes Mellitus Prediction and Analysis for eHealth Services. The proposed work consists of optimal K Nearest Neighbor (Opt-KNN) learning based prediction model based on the patient’s habitual attributes in various dimensions. This approach determines the optimal number of neighbors with low error rate for providing better prediction outcome in the resultant model. The effectiveness of this machine learning eHealth model is examined by conducting experiments on the real-world diabetes mellitus data collected from medical hospitals. The setting of this K-value should be optimized according to the patterns in the dataset. It achieved lowest error rate when K=3. Thus, proposed Opt-KNN based prediction model dynamically selects K=3 as an optimal value to build an effective disease risk prediction model.
In 2021, Mohanty et al. [199] developed a KNN based prediction model for diabetic patients. The proposed work machine learning based algorithm is used to figure out various patterns in our dataset and to calculate the accuracy of this data, with hope that this serves as a stepping stone towards developing tools that can help in medical diagnosis/treatment in future. Creating an efficient diagnostic tool will help improve healthcare to a great extent. The fundamental factors considered in this dataset are age, gender, region of stay and Blood groups. The data should be 98 % accurate for it to be acceptable in real time diagnostic tool development.
In 2021, Patra et al. [200] introduced an Analysis and Prediction of Pima Indian Diabetes Dataset using the SDKNN Classifier Technique. The proposed technique is based on a new distance calculation formula to find nearest neighbors in KNN. It utilized standard deviation of attributes as a powerful tool to measure the distance between train dataset and test dataset. This concept is applied on Pima Indian Diabetes Dataset (PIDD). The analysis is carried out on data set by splitting 90% of training data and 10% of testing data. The proposed approach achieved an accuracy rate of 83.2%, which shows better improvement as compared to the other technique. Table 3 represents the related work on diabetes based on KNN.
3.1.4 Naive Bayes (NB)
In 2010, Sopharak et al. [202] proposed a Machine learning approach for automatic exudate detection in retinal images from diabetic patients. The proposed work consists of a series of experiments on feature selection and exudates classification using naive Bayes and support vector machine (SVM) classifiers. First fit the naive Bayes model to a training set consisting of 15 features extracted from each of 115,867 positive examples of exudate pixels and an equal number of negative examples. Perform feature selection on the naive Bayes model, repeatedly removing features from the classifier, one by one, until classification performance stops improving. To find the best SVM, begin with the best feature set from the naive Bayes classifier, and repeatedly add the previously-removed features to the classifier. The result reveals that the naive Bayes and SVM classifiers perform better than the NN classifier. The overall best sensitivity, specificity, precision, and accuracy are 92.28%, 98.52%, 53.05%, and 98.41%, respectively.
In 2013, Lee [206] introduced a prediction of fasting plasma glucose status using anthropometric measures for diagnosing of Type 2 Diabetes. This study aims to predict the fasting plasma glucose (FPG) status that is used in the diagnosis of type 2 diabetes by a combination of various measures among Korean adults. A total of 4870 subjects (2955 females and 1915 males) participated in this study. Based on 37 anthropometric measures, we compared predictions of FPG status using individual versus combined measures using two machine-learning algorithms. The values of the area under the receiver operating characteristic curve in the predictions by logistic regression and naive Bayes classifier based on the combination of measures were 0.741 and 0.739 in females, respectively, and were 0.687 and 0.686 in males, respectively.
In 2016, Songthung et al. [148] proposed a novel approach to enhance the Type 2 Diabetes Mellitus Risk Prediction using different machine learning technique. The proposed work consists of an extensive dataset gathered from 12 hospitals in Thailand during 2011-2012 with 22,094 records of screened population who are females age 15 years or older. This study used RapidMiner Studio 7.0 with Naive Bayes and CHAID (Chi-squared Automatic Interaction Detector) Decision Tree classifiers to predict high risk individuals and compared the results with existing hand-computed diabetes risk scoring mechanisms. The result shows that Naive Bayes has good coverage and good high-risk percentages compared to both risk scoring and Decision Tree.
In 2018, Das et al. [209] introduced an approach for classification of diabetes mellitus disease using data mining technique. The aim of this research is to predict diabetes based on some of the DM techniques like classification and clustering. Out of which, classification is one of the most suitable methods for predicting diabetes. In this study, J48 and Naïve Bayesian techniques are used for the early detection of diabetes. The experimental results based on data from 200 patient records reveal that Naive Bayes algorithm is better than the J48 as the time to build the model is less.
In 2019, Khan et al. [213] introduced a machine learning based intelligent system for predicting diabetes. The objective of this research is to propose an intelligent system based on a machine learning algorithm to improve the accuracy of predicting diabetes. To attain this objective firstly, an algorithm was proposed based on Naïve Bayes with prior clustering. Secondly, the performance of the proposed algorithm was evaluated using 532 data related to diabetic patients. Finally, the performance of the existing Naïve Bayes algorithm was compared with the proposed algorithm. The results of the comparative study showed that the improvement in the accuracy has been made apparent for the proposed algorithm.
In 2021, Jackins et al. [216] proposed an AI based smart prediction of clinical disease using random forest classifier and Naive Bayes. For diabetes data, the Naive Bayes algorithm gives 76.72 and 74.46 accuracies for training and test data, respectively. Random forest algorithm gives 98.88 and 74.03 for training and test data, respectively. Performance analysis of the disease data for both algorithms is calculated and compared. The results of the simulations show the effectiveness of the classification techniques on a dataset, as well as the nature and complexity of the dataset used.
In 2020, Rghioui et al. [218] introduced a smart glucose monitoring system for diabetic patient. The proposed work presents an intelligent architecture for the surveillance of diabetic disease that will allow physicians to remotely monitor the health of their patients through sensors integrated into smartphones and smart portable devices. The proposed architecture includes an intelligent algorithm developed to intelligently detect whether a parameter has exceeded a threshold, which may or may not involve urgency. To verify the proper functioning of this system developed a small portable device capable of measuring the level of glucose in the blood for diabetics and body temperature. The evaluation result showed that the system using the J48 algorithm exhibited excellent classification with the highest accuracy of 99.17%, a sensitivity of 99.47% and a precision of 99.32%. Table 4 represents the related work on diabetes based on NB.
3.1.5 Random Forest (RF)
In 2020, Wang et al. [230] introduced a Prediction of medical expenditures of diagnosed diabetics and the assessment of its related factors using a random forest model. In this work data were collected from the US household component of the medical expenditure panel survey, 2000–2015. Random forest (RF) model was performed with the programs of the random Forest in R software. Spearman correlation coefficients (rs), mean absolute error (MAE) and mean-related error (MRE) was computed to assess the prediction of all the models. The experimental result indicated that the RF model was little superior to traditional regression model. RF model could be used in prediction of medical expenditure of diabetics and an assessment of its related factors well.
In 2021, Wang et al. [231] present an exploratory study on classification of diabetes mellitus through a combined Random Forest Classifier. This study explored different supervised classifiers, combined with SVM-SMOTE and two feature dimensionality reduction methods (Logistic stepwise regression and LAASO) to classify the diabetes survey sample data by unbalanced categories and complex related factors. Analysis and discussion of the classification results of 4 supervised classifiers based on 4 data processing methods. Five indicators, including Accuracy, Precision, Recall, F1-Score and AUC are selected as the key indicators to evaluate the performance of the classification model. According to the result, Random Forest Classifier is combining SVM-SMOTE resampling technology and LASSO feature screening method (Accuracy= 0.890, Precision = 0.869, Recall = 0.919, F1-Score = 0.893, AUC= 0.948) proved the best way to tell those at high risk of DM. Besides, the combined algorithm helps enhance the classification performance for prediction of high-risk people of DM.
In 2021, Ooka et al. [232] present a Random forest approach for determining risk prediction and predictive factors of type 2 diabetes for the peoples of Japan. This study included a cumulative total of 42 908 subjects not receiving treatment for diabetes with an HbA1c <6.5%. It used two analytical methods to compare the predictive powers: RF as a new model and multivariate logistic regression (MLR) as a conventional model. The RF model showed a higher predictive power for the change in HbA1c than MLR in all models. The RF model, including change values showed the highest predictive power. Table 5 represents the related work on diabetes based on RF.
3.2 Supervised learning (Regression)
In 2009, Gani et al. [261] introduced a data-driven model based on glucose data from one diabetic subject, and subsequently applied to predict subcutaneous glucose concentrations of other subjects, even of those with different types of diabetes. This work employed three separate studies, each utilizing a different continuous glucose monitoring (CGM) device, to verify the model’s universality. The predictive capability of the models was found not to be affected by diabetes type, subject age, CGM device, and inter individual differences.
In 2012, Georga et al. [264] present a predictive modeling of subcutaneous (s.c.) glucose concentration in type 1 diabetes. In this work support vector regression (SVR) technique is utilized. The proposed method is evaluated using a dataset of 27 patients in free-living conditions. Tenfold cross validation is applied to each dataset individually to both optimize and test the SVR model. In the case, where all the input variables are considered, the average prediction errors are 5.21, 6.03, 7.14, and 7.62 mg/dl for 15-, 30-, 60-, and 120-min prediction horizons, respectively. The results clearly indicate that the availability of multivariable data and their effective combination can significantly increase the accuracy of both short-term and long-term predictions.
In 2015, Paul et al. [268] proposed a technique of linear auto-regressive (AR) and state space, time series models to analyze the glucose profiles for predicting upcoming glucose levels. However, these modelling approaches have not adequately addressed the inherent dependencies and volatility aspects in the glucose profiles. The prediction performances of GARCH approach were compared with other contemporary modelling approaches such as lower and higher order AR, and the state space models. The GARCH approach appears to be successful in both realizing the volatility in glucose profiles and offering potentially more reliable forecasting of upcoming glucose levels.
In 2018, Wu et al. [279] introduced a novel model based on data mining techniques for predicting type 2 diabetes mellitus (T2DM). The model is comprised of two parts, the improved K-means algorithm and the logistic regression algorithm. The Pima Indians Diabetes Dataset and the Waikato Environment for Knowledge Analysis toolkit were utilized to compare the results with the results from other researchers. The conclusion shows that the model attained a 3.04% higher accuracy of prediction than those of other researchers. Moreover, our model ensures that the dataset quality is sufficient.
In 2019, Qiu et al. [280] present an improved prediction method for diabetes based on a feature-based least angle regression algorithm. This work consists of a method based on feature weights to improve diabetes prediction that combines the advantages of traditional least angle regression (LARS) algorithms and principal component analysis (PCA) algorithms. First of all, a principal component analysis algorithm is used to obtain the characteristic independent variables found in typical diabetes prediction regression models. The experimental results show that the algorithm improved the approximation speed for the dependent variables and the accuracy of the regression coefficients.
In 2019, Yao et al. [281] proposed a multivariable logistic regression and back propagation artificial neural network based model to predict diabetic retinopathy. A total of 530 Chinese residents, including 423 with type 2 diabetes (T2D) aged 18 years or older participated in this study. In this work a back propagation artificial neural network (BP-ANN) model is utilized by selecting tan-sigmoid as the transfer function of the hidden layers nodes, and pure-line of the output layer nodes, with training goal of 0.5×10−5. Based on these parameters, the area under the receiver operating characteristic (ROC) curve for the BP-ANN model was significantly higher than that by MLR (0.84 vs. 0.77, P < 0.001).
In 2020, Alshamlan et al. [282] introduced a gene prediction function for type 2 diabetes mellitus using logistic regression. In this study the process of feature selection is performed using the Fisher score and chi-square approaches. The total selected number of genes ranges from 1800-2700.The experimental results show that shows that logistic regression produces the highest accuracy with the fisher score for GSE38642 dataset with 90.23% and GSE13760 dataset with 61.90%. Feature selections with logistic regression, classification were used. The obtained accuracy result of logistic regression on two datasets based on fisher score feature selection was higher than Ch-2 feature selection. The accuracy results of two data were 90.23% and 61.90% respectively.
In 2020, Kopitar et al. [283] proposed an early detection of type 2 diabetes mellitus using machine learning based prediction models. This study compares machine learning-based prediction models (i.e. Glmnet, RF, XGBoost, LightGBM) to commonly used regression models for prediction of undiagnosed T2DM. With 6 months of data available, simple regression model performed with the lowest average RMSE of 0.838, followed by RF (0.842), LightGBM (0.846), Glmnet (0.859) and XGBoost (0.881). When more data were added, Glmnet improved with the highest rate (+ 3.4%). The aim of this study was to investigate whether novel machine learning-based approaches offered any advantages over standard regression techniques in the early prediction of impaired fasting glucose (IFG) and fasting plasma glucose level (FPGL) values. Table 6 represents the related work on diabetes based on regression technique.
3.3 Un-supervised learning (clustering technique)
In 2010, Paul et al. [235] introduced a technique how to use the background knowledge of medical domain in clustering process to predict the likelihood of diseases. To find the likelihood of diseases, it proposed constraint k-Means-Mode clustering algorithm. The proposed method also gives much better accuracy when compared to the k-means and K-Mode with about 77-78% over k-means and about 82-83% over k-mode. The developed algorithm can handle both continuous and discrete data and perform clustering based on anticipated likelihood attributes with core attributes of disease in data point. We have demonstrated its effectiveness by testing it for a real world patient data set.
In 2011, Hazemi et al. [236] proposed a grid-based interactive diabetes system. In this work agglomerative clustering algorithm is utilized as primary algorithm to focus medical researcher in the findings to predict the implication of the undertaken diabetes patient. This focusing was clearly shown that the grouped (red) line; which represented the optimized view of blood sugar changes over the newly selected period of time; was providing netted view of blood sugar measurements than the measurements (blue) line. GIDS was tested to study a basic history of a diabetes patient who was under supervision for less than a month. The test was performed to check two functions provided by GIDS which are changing the basic algorithm that GIDS used (Chronological Clustering Algorithm) and changing the full view of the supervision period in the time domain in the study.
In 2013, Khanna et al. [234] introduced an integrated approach towards the prediction of the likelihood of diabetes. This paper performs classification on diabetes dataset taken from SGPGI, Lucknow (A super specialty hospital in Lucknow, Uttar Pradesh, India). It predicts an unknown class label for a given set of data and helpful to find out whether the class label for the dataset under consideration would be of low risk, medium risk or high risk. The classifier is further trained on the basis of weights assigned to different attributes which are generated by means of expert guidelines. The accuracy of classifier is verified by kappa statistics and accuracy, evaluation criteria for classifiers.
In 2015, Flynt et al. [243] introduced a model-based clustering approach for the likelihood of diabetic. This work consists of model-based clustering, an unsupervised learning approach, to fid latent clusters of similar US counties based on a set of socioeconomic, demographic, and environmental variables chosen through the process of variable selection. Then use Analysis of Variance and Post-hoc Tukey comparisons to examine differences in rates of obesity and diabetes for the clusters from the resulting clustering solution. The results of the cluster analysis can be used to identify two sets of counties with significantly lower rates of diet-related chronic disease than those observed in the other identified clusters.
In 2017, Bhatia et al. [245] proposed a hybrid based clustering technique in diabetic prediction. In this research work, K-means has been used for removal of the inconsistency found in the data and for optimal feature selection, genetic algorithm is used with SVM for the purpose of classification. K-means is an optimized hierarchical clustering method which aims at reduction of computational cost. The application of the proposed hybrid clustering model applied to a Pima Indians Diabetes dataset shows increase in accuracy by 1.351% and in both sensitivity and positive predicted value by 2.0411%. The proposed model attains better results in comparison to the already existing models in the literature.
In 2018, Ahlqvist et al. [247] presents k-means and hierarchical clustering technique in prediction of diabetes. The clusters were based on six variables (glutamate decarboxylase antibodies, age at diagnosis, BMI, HbA1c, and homoeostatic model assessment 2 estimates of β-cell function and insulin resistance), and were related to prospective data from patient records on development of complications and prescription of medication. It identified five replicable clusters of patients with diabetes, which had significantly different patient characteristics and risk of diabetic complications. In particular, individuals in cluster 3 (more resistant to insulin) had significantly higher risk of diabetic kidney disease than individuals in clusters 4 and 5, but had been prescribed similar diabetes treatment. Cluster 2 (insulin deficient) had the highest risk of retinopathy. In support of the clustering, genetic associations in the clusters differed from those seen in traditional type 2 diabetes.
In 2020, Nguyen et al. [252] presents a Binning Approach based on Classical Clustering for Type 2 Diabetes Diagnosis. In this study, we propose a method combining K-means clustering algorithm and unsupervised binning approaches to improve the performance in metagenome-based disease prediction. We illustrate by experiments on metagenomic datasets related to Type 2 Diabetes that the proposed method embedded clusters generated by K-means allows to increase the performance in prediction accuracy reaching approximately or more than 70%. Table 7 represents the related work on diabetes based on clustering technique.
3.4 Un-supervised learning (association rule)
In 2000, Hsu et al. [284] proposed a knowledge discovery system for the diabetic patient database, the interesting issues that have surfaced, as well as the lessons we have learnt from this application. The proposed work uses classification with association rule mining (CBA) technique to find all such patterns. It uses minimum support of 1% and minimum confidence of 50% as suggested by the doctors to mine association rules. Approximately 700 rules are generated in total. The result based on 200,000 screening diabetic records suggested that the proposed exploration, mining methodology aims to give the doctors a better understanding of their data and the discovered patterns by helping the doctors to step through the massive amount of information in stages. Table 8 represents the related work on diabetes based on association rule technique.
4 Review of deep learning technique in diabetes prediction
4.1 Convolutional Neural Network (CNN)
A Convolutional Neural Network (ConvNet/CNN) is a Deep Learning algorithm which can take in an input image, assign importance (learnable weights and biases) to various aspects/objects in the image and be able to differentiate one from the other. Table 9 represents the related work on diabetes based on CNN technique.
4.2 Recurrent neural networks (RNN)
RNNs are a powerful and robust type of neural network, and belong to the most promising algorithms in use because it is the only one with an internal memory. Because of their internal memory, RNN’s can remember important things about the input they received, which allows them to be very precise in predicting what’s coming next. This is why they're the preferred algorithm for sequential data like time series, speech, text, financial data, audio, video, weather and much more. Recurrent neural networks can form a much deeper understanding of a sequence and its context compared to other algorithms. Table 10 represents the related work on diabetes based on RNN technique.
4.3 Artificial Neural Network (ANN)
An Artificial Neural Network is an information processing technique. It works like the way human brain processes information. ANN includes a large number of connected processing units that work together to process information. Table 11 represents the related work on diabetes based on ANN technique.
4.4 Long Short-Term Memory Networks (LSTMs)
Long Short Term Memory is a kind of recurrent neural network. In RNN output from the last step is fed as input in the current step. It tackled the problem of long-term dependencies of RNN in which the RNN cannot predict the word stored in the long term memory, but can give more accurate predictions from the recent information. Table 12 represents the related work on diabetes based on LSTM technique.
4.5 Multilayer Perceptron (MLP)
Multilayer Perceptron (MLP) is a class of feed-forward artificial neural networks. An MLP consists of three main layers of nodes — an input layer, a hidden layer, and an output layer. In the hidden and the output layer, every node is considered as a neuron that uses a nonlinear activation function. MLP uses a supervised learning technique called back propagation for training. When a neural network is initialized, weights are set for each neuron. Back propagation helps in adjusting the weights of the neurons to obtain output closer to the expected. Table 13 represents the related work on diabetes based on MLP technique.
4.6 Autoencoder (AE)
Autoencoder is a type of neural network where the output layer has the same dimensionality as the input layer. In simpler words, the number of output units in the output layer is equal to the number of input units in the input layer. An autoencoder replicates the data from the input to the output in an unsupervised manner and is therefore sometimes referred to as a replicator neural network. Table 14 represents the related work on diabetes based on AE technique.
4.7 Radial Basis Function (RBF)
A Radial Basis Function (RBF) neural network has an input layer, a hidden layer and an output layer. The neurons in the hidden layer contain Gaussian transfer functions whose outputs are inversely proportional to the distance from the center of the neuron. RBF networks are similar to K-Means clustering and PNN/GRNN networks. The main difference is that PNN/GRNN networks have one neuron for each point in the training file, whereas RBF networks have a variable number of neurons that is usually much less than the number of training points. For problems with small to medium size training sets, PNN/GRNN networks are usually more accurate than RBF networks, but PNN/GRNN networks are impractical for large training sets. Table 15 represents the related work on diabetes based on RBF technique.
5 Discussion and comparison
In this section, we mainly focus on comparative analysis of several machines and deep learning based diabetic prediction approach , including SVM,KNN, and DT (machine learning) with CNN,RNN and MLP (deep learning) .It also discusses the performance of various machines and deep learning approach for prediction of diabetic disease. Deep learning is computer software that mimics the network of neurons in a brain. It is a subset of machine learning and is called deep learning because it makes use of deep neural networks. The machine uses different layers to learn from the data. The depth of the model is represented by the number of layers in the model. Deep learning is the new state of the art in term of AI.
5.1 SVM vs. CNN
In 2016, Abdillah et al. [45] proposed a machine learning approach using support vector machines with kernel radial basis function (SVM-RBF) to predict diabetes. The Pima Indian diabetes dataset was used to validate the effectiveness of the proposed work. In order to achieve high classification performance, 10-fold cross-validation was used to build a model and search for the optimal parameters. The results of SVM-RBF using 10-fold cross validation was obtained from 500 training data with optimal parameter , which yields accuracy, sensitivity, specificity, and AUROC of 80.22%, 82.56%, 79.12%, and 0.8084 respectively. In the same year Zhu et al. [305] presents a deep learning approach based on CNN to find a patient similarity evaluation framework based on temporal matching of longitudinal patient EHRs (Electronic Health Records). The results of clustering, the deep model with feature embedding is clearly superior to others. On DATASET-I, the deep embedding model achieves an average Rand index of 0.9887, comparing with the second best one with 0.6796. Measured by Purity and NMI, it can achieve the performances of 0.9882 and 0.9516, separately, which also outperforms others with a margin.
In 2018, Dagliati et al. [59] proposed a machine learning approach based on SVM to predict diabetic complications. As far as the choice of the classification method is concerned, AUC values are higher for SVMs and RF when the data sets are balanced. However, SVMs and RF models are harder to interpret, especially considering that our final goal is the model application into clinical practice. In the same year Swapna et al. [310] proposed a deep learning based diabetic prediction model based on CNN approach. The proposed work consists of long short-term memory (LSTM), convolutional neural network (CNN) and its combinations for extracting complex temporal dynamic features of the input HRV data. These features are passed into a support vector machine (SVM) for classification. It has obtained the performance improvement of 0.03% and 0.06% in CNN and CNNLSTM architecture respectively, compared to earlier work without using SVM. The classification system proposed can help the clinicians to diagnose diabetes using ECG signals with a very high accuracy of 95.7%.
In 2019, Alirezaei et al. [66] present a machine learning approach based on SVM to predict diabetic complications. In this work, a method based on the k-means clustering algorithm is first utilized to detect and delete outliers. Then in order to select significant and effective features, four bi-objective meta-heuristic algorithms are employed to choose the least number of significant features with the highest classification accuracy using support vector machines (SVM). The results, based on PIMA Indian Type-2 diabetes dataset concluded that the multi-objective firefly (MOFA) and multi-objective imperialist competitive algorithm (MOICA) with 100% classification accuracy, outperform the non-dominated sorting genetic algorithm (NSGA-II) and multi-objective particle swarm optimization (MOPSO) with the accuracies of 98.2% and 94.6%, respectively.Sun et al. [313] proposed a Neural Network Method (CNN) based deep learning approach to build a diagnostic model, in this work, the CNN model is combined with the BN layer to prevent the dispersion of the gradient, speed up the training speed and improve the accuracy of the model. The experiments show that this method can achieve a training accuracy of 99.85% and a testing accuracy of 97.56%, which is more than 2% higher than that of using logistic regression.
In 2020, Harimoorthy et al. [79] proposed a multi disease prediction model using machine learning approach based on improved SVM-Radial bias technique. In this work, a general architecture has proposed for predicting the disease in the healthcare industry. This system was experimented using with reduced set features of Chronic Kidney Disease, Diabetes and Heart Disease dataset using improved SVM-Radial bias kernel method, and also this system has compared to other machine learning techniques such as SVM-Linear, SVM-Polynomial, Random forest and Decision tree in R studio. The performance of all these machine learning algorithms has evaluated with accuracy, misclassification rate, precision, sensitivity and specificity. From the experiment results, improved SVM-Radial bias kernel technique produces accuracy as 98.3%, 98.7% and 89.9% in Chronic Kidney Disease, Diabetes and Heart Disease dataset respectively. Ismail et al. [316] proposed a Remote health monitoring application with the advent of Internet of Things (IoT) technologies. The proposed work uses a deep learning approach based on CNN. In this framework, develop a CNN-regular pattern discovery model for data classification. First, the most important health-related factors are selected in the first hidden layer, then in the second layer, a correlation coefficient analysis is conducted to classify the positively and negatively correlated health factors. Moreover, regular patterns’ behaviours are discovered through mining the regular pattern occurrence among the classified health factors. The accuracy of diagnosis and referral of our model reached 80.43%; 80.85%; 91.49%; 82.61%; 95.60% with a testing dataset, respectively.
5.2 KNN vs. RNN
In 2018, Dwivedi et al. [385] introduced a computational intelligence technique for diabetes mellitus prediction. The proposed model uses a machine learning approach based on KNN (K-Nearest Neighbor). Clearly indicates that the ANN and the logistic model predicts the highest number of true positive (430 and 443 out of 500, respectively), where naïve Bayes predicts the highest number of true negative (179 out of 268). Naïve Bayes also predicts lowest number of false positive (89) whereas logistic regression predicts the lowest number of false negative. Overall, naïve Bayes has lowest type I error and logistic regression has lowest type II. Chen et al. [330] proposed a new deep learning technique, which is based on the Dilated Recurrent Neural Network (DRNN) model, is proposed to predict the future glucose levels for prediction horizon (PH) of 30 minutes. The result reveals that using the dilated connection in the RNN network, it can improve the accuracy of short-time glucose predictions significantly (RMSE = 19.04 in the blood glucose level prediction (BGLP) on and only on all data points provided). Lastly, in order to improve the performance of DRNN model, the first-order linear interpolation and first-order extrapolation are applied to the training and testing set, respectively.
In 2019, Alehegn et al. [188] introduced a MLTs (Machine Learning Techniques) that can act as a savior for early diagnosis and prediction of DM. ML is another side of Artificial Intelligence so that be used for prediction, recommendation and recovery from disease in early stages. The system proposed in this work makes use of two datasets viz. PIDD (Pima Indian Diabetes Dataset) and 130_US hospital diabetes data sets. Techniques used for datasets analysis are Random Forest, KNN, Naïve Bayes, and J48. The ensemble approach facilitates in achieving better results. The accuracy of proposed ensemble approach is 93.62% for PIDD and 88.56% for the 130_US hospital dataset. It is also observed that when dataset becomes large the accuracy of the proposed algorithm is not good relatively. NB and J48 prediction algorithm are better for large datasets analysis. KNN technique is not good for large dataset analysis. Li et al. [406] present a deep learning model that is capable of forecasting glucose levels with leading accuracy for simulated patient cases. This work is based on multi-layer convolutional recurrent neural network (CRNN) architecture. The proposed CRNN method showed superior performance in forecasting BG levels (RMSE and MARD) in the in silico and clinical experiments. The results achieved a mean RMSE = 9.38mg/dL in silico using the proposed method, and it is the best amongst other algorithms, including SVR, LVX and 3rd order ARX.
In 2018, Sarker et al. [197] proposed an optimal K-Nearest Neighbor (KNN) Learning based Diabetes Mellitus Prediction and analysis for eHealth Services. In this model select 5 baseline classification methods, such as Adaptive Boosting (AdaBoost), Logistic Regression (LR), Naive Bayes (NB), Support Vector Machines (SVM), and Decision Tree (DT) that are frequently used to analyze health data. The results show that proposed Opt-KNN based disease prediction model outperforms the traditional KNN based model. Opt-KNN based model gives better prediction accuracy in terms of precision, recall, f-measure, ROC area. This results show that Opt-KNN is more effective that traditional KNN in terms of prediction accuracy and minimize the additional effort for assuming the K-value. Zhou et al. [326] proposed a deep learning approach using RNN that can help not only to predict the occurrence of diabetes in the future, but also to determine the type of the disease that a person experiences. The experimental results show the effectiveness and adequacy of the proposed DTP model. The best result for the diabetes type dataset was 94.021 74% and that for the Pima Indians diabetes dataset was 99.411 2%. The experiments proved that proposed model can perform well on different types of data. The proposed model not only can predict if a person will be diabetic in the future, but also can determine and predict the specific type of the disease, type 1 or type 2.
In 2021, Patra et al. [200] introduced a machine learning based diabetic prediction model using standard deviation K-Nearest Neighbor (SDKNN). The proposed technique is based on a new distance calculation formula to find the nearest neighbor in KNN. The work consists of two segments, in the first segment standard deviation of attributes is used as power for calculating K-nearest neighbor to improved classification accuracy and in the second segment, based on mean of standard deviation attributes ,distance in KNN is processed to further improve the classification accuracy. This concept is applied on Pima Indian Diabetes Dataset (PIDD). The analysis is carried out on data set by splitting 90% of training data and 10% of testing data. The proposed approach achieved an accuracy rate of 83.2%, which shows better improvement as compared to the other technique.
Rabby et al. [328] presents a novel approach to predicting the blood glucose level with a stacked long short term memory (LSTM) based deep recurrent neural network (RNN) model considering sensor fault. In this work Kalman smoothing technique is used for the correction of the inaccurate CGM readings due to sensor error. Results demonstrate that the RNN model with stacked LSTM layered architecture performs better than RNN with a single LSTM layer for all of the cases based on OhiT1DM dataset. The proposed approach is more generalized as the prediction RMSE for all six patients is uniformly improved. As a consequence, it does not further experiment with traditional machine learning approaches. Proposed approach provides more reliable predictions than traditional methods while it assumed fngerstick BG readings as the ground truth in our experiment.
5.3 DT vs. MLP
In 2018, Zou et al. [155] proposed a diabetes mellitus prediction model using different machine learning technique. In this study, it uses decision tree, random forest and neural network to predict diabetes mellitus. The dataset is the hospital physical examination data in Luzhou, China. It contains 14 attributes. In this study, five-fold cross validation was used to examine the models. Principal component analysis (PCA) and minimum redundancy maximum relevance (mRMR) to reduce the dimensionality. The results showed that prediction with random forest could reach the highest accuracy (ACC = 0.8084) when all the attributes were used. In the Luzhou dataset, J48 (Decision Tree) has the best performance. But the results are not better than using all features. In the Pima Indians dataset, this method, which used RF as the classifier, has the best performance. Alfian et al. [371] developed a personalized healthcare monitoring system for diabetic patients by utilizing deep learning approach based on multi-layer perceptron (MLP). The results show that MLP achieved the highest accuracy (77.083%) compared to 73.046%, 76.6927%, 76.562%, and 76.0417% for Random Forest, Naïve Bayes, SVM, and Logistic Regression, respectively. These results show that for a small number of features (2 h glucose tolerance, diastolic blood pressure, body mass index, and age), the MLP algorithm achieved the highest accuracy of prediction compared to other models.
In 2019, Hebbar et al. [161] introduced a decision tree (DT) based prediction model for diabetic patients. Decision tree and random forest algorithms are applied on data to learn the class model. The optimal model is selected. Then the chosen model is promoted for testing with the test set of data. DRAP makes the classification and prediction based on the feature set mainly consisting of BMI, age, blood pressure, insulin level, and glucose level. The model used to modify decision tree, and random forest algorithm for learning, classification and prediction. Experimental study of the real-life data set has shown promising results and DRAP - yield the accuracy of 72% and 75% for decision tree, and random forest respectively. Mohapatra et al. [372] introduced a deep learning based diabetes prediction model using multi-layer perceptron (MLP). In this work, MLP is used for classification of pregnant women. Proposed technique has been applied on the diabetes database of PIMA for Indian people and is collected from University of California (UCI). Total l768 lady patients are considered for the experiment It is found that 268 are suffering from diabetes and rest 500 cases are in healthy. Also, it is verified for missing data and found no missing is there. The experiment results achieved 77.5% of classification accuracy, using the proposed approach.
In 2020, Maniruzzaman et al. [168] proposed a decision tree (DT) based machine learning approach for diabetic prediction. Moreover, LR-based model has been adopted to determine the high risk factors of diabetes disease. The high risk factors have been selected based on p-values and odds ratio (OR). Moreover, four classifiers have been also adapted and compared their performance based on ACC, SE, PPV, NPV, FM, and AUC, respectively. The dataset consists of 6561 respondents with 657 diabetic and 5904 controls. The overall ACC of ML-based system is 90.62%. The combination of LR-based feature selection and RF-based classifier gives 94.25% ACC and 0.95 AUC for K10 protocol. Bani-Salameh et al. [373] introduced a model based on the Multi-Layer Perceptron Neural Network (MLP). The main objective of this research is to benefit from ANN’s prediction capabilities. Examine whether an MLP neural network can help to precisely predict if patients are diabetes and/or suffer from blood pressure problems. Also, help determine the factor which has a high influence on these diseases. This study presents a prediction method for both diabetes and blood pressure by using ANNs. Python programming language was used to build the neural network model, test its accuracy, and compare it with other neural networks and classifiers. The model predicted the two diseases with correct classification rate (CCR) of 77.6% for diabetes and 68.7% of hypertension. The results indicate that MLP correctly predicts the probability of being diseased or not, and the performance can be significantly increased compared with both SVM and KNN. This shows MLP’s effectiveness in early disease prediction.
In 2021, Taser et al. [172] introduced the application of Bagging and Boosting Approaches Using Decision Tree-Based Algorithms in Diabetes Risk Prediction. The proposed model consists of six different decision tree based (DTB) classifiers were implemented on experimental data for diabetes prediction. This work also compares applied individual implementation, bagging, and boosting of DTB classifiers in terms of accuracy rates. The results indicate that the bagging and boosting approaches outperform the individual DTB classifiers, and real Adaptive Boosting (AdaBoost) and bagging using Naive Bayes Tree (NBTree) present the best accuracy score of 98.65%. In this study, bagging and boosting approaches using DTB algorithms were implemented in the experimental data to predict diabetes risk at an early stage. Thyde et al. [386] introduced a model which detects Type 2 Diabetes Patients using deep learning (MLP) approach. In this study, it explores how deep learning (DL) based on CGM data can be used for detecting adherence to once-daily basal insulin injections. It further considered a multilayered feed-forward neural network based on multilayer perceptrons (MLPs) and CNNs, the latter based on the raw CGM as input to the model. The T2D modified version of the MVP model was successfully used to simulate a large amount of realistic CGM data. The data were used to develop methods for treatment adherence detection. The automatically extracted features based on DL methods with added expert-dependent features performed best with an accuracy of 79.8% ± 0.5% 16 hours after TOI.
5.4 Discussion
5.4.1 Machine learning based
This section mainly focuses on a comparison of results for different machine learning approach for prediction of diabetic disease, including SVM, DT, KNN, NB and RF. There are various types of machine learning approach have been utilized by different researcher for prediction of diabetic disease. The machine learning technique is able to solve these issues which are faced by the doctors to diagnose properly the diabetic patients, it also helps the patient for early detection of diabetics, so that by taking the prior precaution the disease of diabetes can be minimized. The related works on machine learning based diabetic prediction model was proposed by [92, 171, 198, 231, 218]. The related work on diabetics using different ML and DL techniques from year 2014 to 2023 is shown in comparative graph which is depicted in fig 5.8. It is evident from the comparative graph of the year 2023 that ML techniques ([409] for SVM, [412] for DT, [414] for KNN, [417] for NB, and [419] for RF) as well as DL techniques ([429] for ANN, [426] for RNN, [424] for CNN, [434] for MLP, and [438] for RBF) were utilized for diabetic prediction.
In 2021, Dinesh et al. [92] proposed a Diabetes Mellitus Prediction System Using Hybrid KPCA-GA-SVM Feature Selection Techniques. The proposed work implements the Kernel Principal Component Analysis for dimensionality reduction. Genetic Algorithm to select the relevant and optimal features from the dataset. Then at the last Support Vector Machine is used as a classifier to classify the diabetes mellitus data. The proposed KPCA-GA-SVM obtains accuracy of 99.53% and also reduced feature size compared to GA-SVM of 98.79% accuracy. It is also proved that proposed algorithm performs better in terms of sensitivity (96.4%), specificity (94%), accuracy (97.3%) and MCC (89.3%) compared to other classification algorithms.
In 2020, Haq et al. [171] introduced an intelligent machine learning approach for effective recognition of diabetes in E- Healthcare uses clinical data. In this work, a filter method based on the Decision Tree (Iterative Dichotomiser 3) algorithm for highly important feature selection. Two ensemble learning algorithms, AdaBoost and Random Forest, are also used for feature selection and also compared the classifier performance with wrapper based feature selection algorithms. Classifier Decision Tree has been used for the classification of healthy and diabetic subjects. The experimental results show that the proposed feature selection algorithm selected features improve the classification performance of the predictive model and achieved optimal accuracy. The proposed method DT (ID3) +DT achieved 99% test accuracy, 99.8% accuracy with k-floods and 99.9% accuracy with LOSO validation. The accuracy rate achieved by the proposed method is higher than the accuracy rate achieved by authors in [92].
In 2021, Bhardwaj et al. [198] introduced a hierarchical severity level grading (HSG) system for the detection and classification of diabetic retinopathy (DR). In this work, SVM and KNN classification algorithm are utilized for the prediction of diabetic retinopathy. The proposed system achieves an overall accuracy of 98.10% by SVM classifier and 100% by kNN classifier. Hierarchal discrimination into further grades of abnormalities resulted in accuracy values of 95.68% and 92.60% with SVM classifier using Gaussian kernel and, 97.90% and 95.30% employing fine kNN classifier. Gaussian RBF kernel for SVM classifier and fine kNN classifier provides better performance in terms of diferent indices due to the robustness of these classifiers for non-linear classifcation problems. The accuracy rate achieved by the proposed method is higher than the accuracy rate achieved by authors in [171] and as well as by authors in [92].
In 2020, Rghioui et al. [218] presents an intelligent architecture for the surveillance of diabetic disease that will allow physicians to remotely monitor the health of their patients through sensors integrated into smartphones and smart portable devices. The classification algorithms used in the study were the naive Bayes (NB), J48, random tree, ZeroR, SMO (sequential minimal optimization), and OneR algorithms. Results demonstrated that the J48 algorithm exhibited excellent classification, with the highest accuracy of 99.17%, a sensitivity of 99.47% and a precision of 99.32%.NB achieved an accuracy of 85.16%, which is lower than the accuracy rate achieved by authors in [171] and as well as by authors in [92] and [198].
In 2021, Wang et al. [231] present an exploratory study on classification of diabetes mellitus through a combined Random Forest Classifier. This study explored different supervised classifiers, combined with SVM-SMOTE and two feature dimensionality reduction methods (Logistic stepwise regression and LAASO) to classify the diabetes survey sample data by unbalanced categories and complex related factors. Analysis and discussion of the classification results of 4 supervised classifiers based on 4 data processing methods. According to the result, Random Forest Classifier is combining SVM-SMOTE resampling technology and LASSO feature screening method (Accuracy= 0.890, Precision = 0.869, Recall = 0.919, F1-Score = 0.893, AUC= 0.948) proved the best way to tell those at high risk of DM. Besides, the combined algorithm helps enhance the classification performance for prediction of high-risk people of DM. The results showed that the Random Forest classifier combining with SVM-SMOTE and LASSO feature reduction method performs best in telling high-risk patients of DM from ordinary individuals. The proposed approach achieved a higher accuracy level than authors in [218], but it is less than the accuracy level of authors proposed by [198] and authors in [171] and [92].
5.4.2 Disadvantage of existing (machine learning) techniques
This section mainly focuses on disadvantages of existing work for different machine learning approach for prediction of diabetic disease, including SVM, DT, KNN, NB and RF. Diabetes Mellitus Prediction System Using Hybrid KPCA-GA-SVM Feature Selection Techniques proposed by Dinesh et al. [92] shows promising results, the hybrid approach combining Kernel Principal Component Analysis (KPCA), Genetic Algorithm (GA), and Support Vector Machine (SVM) introduces increased complexity to the model. This complexity may lead to higher computational costs, longer training times, and increased resource requirements, especially for large datasets. The hybrid approach may sacrifice interpretability for improved performance. With multiple layers of feature selection and dimensionality reduction techniques, it may become challenging to interpret how individual features contribute to the prediction model. The high accuracy reported by the proposed method could potentially indicate overfitting, especially if the model is trained and evaluated on the same dataset. It is essential to validate the model's performance on unseen data to ensure that it generalizes well to new instances.
While the intelligent machine learning approach introduced by Haq et al. [171] for the recognition of diabetes in E-Healthcare demonstrates impressive results, while the feature selection algorithms like Decision Tree (Iterative Dichotomiser 3), AdaBoost, and Random Forest are powerful techniques for selecting relevant features, they may introduce bias in feature selection. The choice of features and the criteria used to select them can influence the model's performance and may not always capture the most informative features for diabetes recognition across diverse datasets or populations. Ensemble learning methods such as AdaBoost and Random Forest can be computationally expensive, especially when dealing with large datasets or a high number of features. The performance of the proposed method heavily relies on the quality and representativeness of the clinical data used for training and evaluation. Inaccurate or incomplete data, common in healthcare datasets, can lead to biased model predictions and undermine the effectiveness of the proposed approach.
While the hierarchical severity level grading (HSG) system proposed by Bhardwaj et al. [198] for the detection and classification of diabetic retinopathy (DR) shows impressive results, the reported high accuracy rates, especially 100% accuracy by the kNN classifier, raise concerns about potential overfitting or bias in the model. It is essential to validate the model's performance on unseen data to ensure that it generalizes well to new instances and diverse populations. The complex decision boundaries learned by SVM with Gaussian RBF kernel and kNN may make it challenging to understand the underlying factors contributing to the classification decisions, limiting the clinical interpretability of the model's predictions. SVM with Gaussian RBF kernel and kNN classifiers can be computationally expensive, especially when dealing with large datasets or high-dimensional feature spaces.
While the intelligent architecture presented by Rghioui et al. [218] for the surveillance of diabetic disease using smartphone and smart portable devices shows promising results, the effectiveness of the surveillance system heavily relies on the features extracted from sensors integrated into smartphones and smart portable devices. The limited set of features may not capture all relevant aspects of diabetic health, potentially leading to incomplete or inaccurate monitoring of the disease. The accuracy and reliability of the sensor data collected from smartphones and smart portable devices are critical for the effectiveness of the surveillance system. Inaccurate or noisy sensor data may lead to erroneous classifications and misinterpretations of diabetic health status, compromising the reliability of the system. The performance of the surveillance system may vary across different patient populations and demographic groups. The effectiveness of the classification algorithms and the accuracy of the surveillance system may depend on factors such as age, gender, ethnicity, and comorbidities, which may not be adequately addressed in the study. Remote monitoring of patients' health using smartphones and smart portable devices raises privacy and security concerns regarding the collection, storage, and transmission of sensitive health data. Ensuring compliance with data protection regulations and implementing robust security measures is crucial to maintain patient confidentiality and prevent unauthorized access to health information.
While the study by Wang et al. [231] on the classification of diabetes mellitus through a combined Random Forest Classifier offers promising results, dealing with unbalanced categories in the dataset poses challenges for classification algorithms. While methods like SVM-SMOTE can help address class imbalance by oversampling minority classes, they may also introduce biases or overfitting, particularly if not applied carefully or if the underlying assumptions of the data distribution are not met. The combined approach of Random Forest Classifier with SVM-SMOTE and LASSO feature reduction method may increase the computational complexity of the classification model, especially for large datasets or high-dimensional feature spaces. This could limit the scalability of the model, particularly in real-time or resource-constrained environments. While the reported performance metrics (e.g., Accuracy, Precision, Recall, F1-Score, AUC) indicate strong classification performance, there is a risk of overfitting, particularly if the model is not properly validated on unseen data. It is essential to assess the generalizability of the model across different datasets and patient populations to ensure its reliability in diverse clinical settings.
5.4.3 Deep learning based
This section mainly focuses on a comparison of results of different deep learning approach for prediction of diabetic disease, including CNN, RNN, LSTM, MLP and RBF. There are various types of deep learning approach have been utilized by different researcher for prediction of diabetic disease. Deep learning technique is able to solve these issues which are faced by the doctors to diagnose properly the diabetic patients, it also helps the patient for early detection of diabetics, so that by taking the prior precaution the disease of diabetes can be minimized. The related works on deep learning based diabetic prediction model was proposed by [82, 316, 327, 352, 373].
In 2020, Ismail et al. [316] proposed a remote health monitoring system using a deep learning approach based on CNN. In this work, first, the most important health-related factors are selected in the first hidden layer, then in the second layer, a correlation coefficient analysis is conducted to classify the positively and negatively correlated health factors. By exploiting such knowledge of the regular correlated algorithm, the proposed model demonstrated competitive analysis performance on 4,759,777 medical records. The accuracy of diagnosis and referral of our model reached 80.43%; 80.85%; 91.49%; 82.61%; 95.60% with a test dataset, respectively. Regarding the performance study of the proposed model, it provides knowledge related to regular-correlated health parameters of obesity, high blood pressure, and diabetes.
In 2020, Zhu et al. [327] introduce a deep learning model based on a dilated recurrent neural network (DRNN) to provide 30-min forecasts of future glucose levels. The proposed approach outperforms existing glucose forecasting algorithms, including autoregressive models (ARX), support vector regression (SVR) and conventional neural networks for predicting glucose (NNPG) (e.g. RMSE = NNPG, 22.9 mg/dL; SVR, 21.7 mg/dL; ARX, 20.1 mg/dl; DRNN, 18.9 mg/dL on the OhioT1DM dataset). The results suggest that dilated connections can improve glucose forecasting performance efficiency. Compared with the standard RNNs, the recurrent layers in the DRNN model exponentially increase dilation to expand their receptive fields and improve the prediction accuracy. Proposed model results show that the DRNN model achieves the best performance with the smallest RMSE, MARD and time lag. Therefore, it is believed the DRNN model is a promising approach to achieve good BG prediction and has great potential for future research in diabetes management.
In 2020, Carrillo-Moreno et al. [352] a glucose predictor based on long short-term memory (LSTM) neural networks is designed. Different prediction times and input dimensions have been evaluated in order to provide the best prediction to patients. The main goal of this paper is to design and implement a set of predictors with the aim of predicting accurately the glucose level of type 1 diabetes and improving previous results in the literature. First, a main predictor model development will consist of an LSTM fed with previous values of glucose, insulin bolus and meal intake. Based on the main predictive model, a set of predictors specialized in forecasting for different PHs will be deployed. Twelve models have been deployed to achieve the objective of predicting glucose concentration in patients with type 1 diabetes. These models may be grouped by their PH: 5 min, 15 min, 30 min and 45 min. The predictors with a PH of 5 min are the most accurate models, but they do not provide enough time to anticipate therapeutic actions to predict adverse events (hypoglycemia or hyperglycemia) due to the delayed action of insulin infusions. Consequently, the predictors with a PH of 5 min are not useful in a clinical scenario.
In 2020, Bani-Salameh et al. [373] present a Prediction model of Diabetes and Hypertension using Multi-Layer Perceptron (MLP) Neural Networks. The inputs of the network were the factors for each disease, while the output was the prediction of the disease’s occurrence. The model performance was compared with other classifiers Support Vector Machine (SVM) and K-Nearest Neighbors (KNN). It used performance metric measures to assess the accuracy and performance of MLP. The model predicted the two diseases with correct classification rate (CCR) of 77.6% for diabetes and 68.7% of hypertension. The results indicate that MLP correctly predicts the probability of being diseased or not, and the performance can be significantly increased compared with both SVM and KNN. This shows MLP’s effectiveness in early disease prediction. The results indicate 77.6% accuracy in diabetes and 68% in hypertension. Also, F1-score and MCC values show improvement for MLP compared with both of the two algorithms. This shows that the used method is effective and improves disease prediction.
In 2020, Nnamoko et al. [82] presents a RBF based deep learning approach for prediction of diabetes. Experiments with Naïve Bayes, SVM-RBF, C4.5 and RIPPER show that proposed selective data preprocessing method applied to C4.5 decision tree produced better results than the other three classifiers with 89.5% Accuracy, 90% Precision, 89.4% Recall, 89.5% F-score and 83.5% Kappa. These results are also better than baseline experiments conducted with AdaBoostM1 and Random Forest. The experiment results show that SVM-RBF trained with IQRd +SMOTEd produced the best results. To experimentally demonstrate the significance of this improvement, over the best performing models from the other classifiers, including the baseline models – AdaBoostM1 and Random Forest; it conducted a McNemar's test to compare their predictions. The performance of SVM-RBF trained with IQRd +SMOTEd data led to significant improvement in all but one classifier, i.e., Random Forest. Nevertheless, the result is a clear indication that given the right classifier, models trained in the selective data preprocessing method presented in this study generally responds positively to class imbalance and outliers.
5.4.4 Disadvantage of existing (deep learning) techniques
This section mainly focuses on disadvantages of existing work for different deep learning approach for prediction of diabetic disease, including CNN, RNN, LSTM, MLP and RBF. While the remote health monitoring system proposed by Ismail et al. [316] using a deep learning approach based on CNN demonstrates competitive analysis performance, Deep learning models, such as convolutional neural networks (CNNs), are often considered black-box models due to their complex architectures and numerous parameters. While the model may achieve high accuracy, understanding the underlying factors and features driving the predictions can be challenging. Deep learning models, particularly CNNs, can be computationally expensive and resource-intensive, especially when dealing with large datasets and complex architectures. While the reported accuracy rates on the test dataset are promising, it is essential to validate the performance of the model on independent datasets and real-world clinical scenarios. While the reported accuracy rates on the test dataset are promising, it is essential to validate the performance of the model on independent datasets and real-world clinical scenarios.
While the dilated recurrent neural network (DRNN) proposed by Zhu et al.[327] shows promising results in forecasting glucose levels, like any model, deep learning models, especially those incorporating advanced architectures like DRNNs, can be computationally expensive to train and deploy. Deep learning models, including DRNNs, often require large amounts of data to generalize well and make accurate predictions. Limited data availability or poor data quality could hinder the performance of the model. Deep learning models are often criticized for their lack of interpretability. It can be difficult to understand how the model arrives at its predictions, which may be crucial in healthcare applications where interpretability is important for trust and acceptance.
The glucose predictor based on Long Short-Term Memory (LSTM) neural networks designed by Carrillo-Moreno et al.[352] presents several advantages in predicting glucose levels for type 1 diabetes patients. However, there are also some disadvantages or limitations noted in the study, the predictors with a prediction horizon (PH) of 5 minutes, which are the most accurate models, are deemed not useful in a clinical scenario because they do not provide enough time to anticipate therapeutic actions for adverse events such as hypoglycemia or hyperglycemia. The model relies on inputs such as previous glucose levels, insulin bolus, and meal intake. While these features are important for predicting glucose levels, inaccuracies or missing data in these inputs could affect the reliability and accuracy of the predictions. LSTM neural networks, while powerful for sequence prediction tasks, can be complex and difficult to interpret. Understanding how the model arrives at its predictions, especially in healthcare settings where interpretability is crucial, may pose challenges.
While the prediction model of Diabetes and Hypertension using Multi-Layer Perceptron (MLP) Neural Networks presented by Bani-Salameh et al. [373] demonstrates promising results, MLP neural networks are often criticized for their lack of interpretability. Understanding how the model makes predictions can be challenging, especially in clinical settings where interpretability is crucial for decision-making and trust in the model's output. The performance of the MLP model heavily relies on the quality of the input data and the selection of relevant features. Inadequate or biased data, as well as irrelevant features, could lead to inaccurate predictions and diminish the model's effectiveness. Training MLP neural networks, especially with large datasets and complex architectures, can be computationally expensive and time-consuming. This may pose challenges in resource-constrained environments or real-time prediction scenarios. MLP models have several hyper parameters that need to be tuned to achieve optimal performance. Finding the right combination of hyper parameters requires extensive experimentation and computational resources. The performance of the MLP model may heavily depend on the distribution and diversity of the training data. Models trained on limited or biased datasets may not generalize well to new patients or different populations.
While the RBF-based deep learning approach presented by Nnamoko et al. [82] shows promising results for predicting diabetes, RBF-based deep learning models, like other deep learning approaches, can be complex and difficult to interpret. Understanding the underlying mechanisms and decision-making processes of such models may pose challenges, especially in clinical settings where interpretability is crucial. RBF-based deep learning models have several hyper parameters that need to be tuned to achieve optimal performance. Finding the right combination of hyper parameters requires extensive experimentation and computational resources. Predictive models for diseases such as diabetes raise ethical considerations related to patient privacy, consent, and the potential consequences of false positives or false negatives. Ensuring the responsible use of predictive models in healthcare is essential. The performance of the RBF-based deep learning model may vary across different populations or healthcare settings. Models trained on specific datasets may not generalize well to new patients or diverse demographic groups. While the RBF-based deep learning model demonstrates improved prediction accuracy, its integration into existing clinical workflows and electronic health record systems may pose technical challenges. While the model may achieve high accuracy and performance metrics, the clinical significance of these results should be carefully interpreted. High accuracy does not necessarily guarantee improved patient outcomes, and the model's predictions should be validated in real-world clinical settings.
5.5 Motivation and hypothesis
The main focus of this proposed work is to discuss the previous work done in the area of prediction of diabetics, including Type-1, Type-2, and Gestational diabetes based on different machine and deep learning approach. Although the prediction model of diabetic disease based on the machine and deep learning approach has greatly improved over the last decade, different prediction algorithms and its approach of implementation has been improved in terms of Classification accuracy, enhance the Sensitivity rate, enhance Specificity rate and provide overall true diabetic prediction model. , a lot of work still needs to be done to make the prediction process more practicable so that it can be easily applied on different application of health care system. In addition, this review also compared the different research work based on different machine and deep learning based algorithms and discuss the different performance parameter like TP (number of diabetes patients detected as a patient), FP (number of healthy persons detected as a patient), TN (number of healthy persons detected as healthy), and FN (number of patients detected as healthy), different algorithm are used in prediction model and the different classifier are used in the identification process. It also discusses explicitly the factor that may affect the prediction model of the system.
Various researchers have used different algorithm in their research work for data collection, feature extraction, prediction methodology, classification and performance evaluation to measure the accuracy and strength of the system. It is very difficult to directly compare and contrast many of these studies in terms of their accuracy system and performance, as their method for evaluating the performance differs depending upon the aim of the study. The success of any predictive model for diabetic system is mainly depends on the algorithm used, which is compulsory to combine the information presented by multiple domain experts. The main purpose of any predictive model is to determine the best set of experts in a given problem domain and implement an appropriate function that can optimally combine the decision produced by individual experts.
Interest in machine learning for healthcare has grown immensely, including work in diagnosing diabetic retinopathy, cancer detection, heart failure, and hypertensions. Despite these advances, the direct application of machine learning to healthcare remains fraught with pitfalls. Many of these challenges stem from the nominal goal in healthcare to make personalized predictions using data generated and managed via the medical system, where data collection's primary purpose is to support care, rather than facilitate subsequent analysis. In tackling healthcare tasks, there are factors that should be considered carefully in the design and evaluation of machine learning projects: causality, missingness, and outcome definition. These considerations are important across both modeling frameworks (e.g., supervised vs. unsupervised), and learning targets (e.g., classification vs. regression). Even if all important variables are included in a health care dataset, it is likely that many observations will be missing. Truly complete data are often impractical due to cost and volume. Learning from incomplete, or missing, data has received little attention in the machine learning community. Obtaining reliable outcomes for learning is an important step in defining tasks. Outcomes are often used to create the gold-standard labels needed for supervised prediction tasks, but are crucial in other settings as well, e.g., to ensure well-defined cohorts in a clustering task. There are three key factors to consider with outcome definitions: creating reliable outcomes, understanding the relevance of an outcome clinically, and the subtlety of label leakage.
Machine learning is a general-purpose method of artificial intelligence that can learn relationships from the data without the need to define them a priori. The major appeal is the ability to derive predictive models without a need for strong assumptions about the underlying mechanisms, which are usually unknown or insufficiently defined. The typical machine learning workflow involves four steps: data harmonization, representation learning, model fitting and evaluation. For decades, constructing a machine learning system required careful engineering and domain expertise to transform the raw data into a suitable internal representation from which the learning subsystem, often a classifier, could detect patterns in the data set.
Deep learning is different from traditional machine learning in how representations learn from the raw data. In fact, deep learning allows computational models that are composed of multiple processing layers based on neural networks to learn representations of data with multiple levels of abstraction. The major differences between deep learning and traditional artificial neural networks (ANNs) are the number of hidden layers, their connections and the capability learn meaningful abstractions of the inputs. In fact, traditional ANNs are usually limited to three layers and are trained to obtain supervised representations that are optimized only for the specific task and are usually not generalizable. Differently, every layer of a deep learning system produces a representation of the observed patterns based on the data it receives as inputs from the layer below, by optimizing a local unsupervised criterion. The key aspect of deep learning is that these layers of features are not designed by human engineers, but they are learned from data using a general purpose learning procedure.
More recently deep learning has been applied to process, aggregated EHRs, including both structured (e.g. diagnosis, medications, laboratory tests) and unstructured (e.g. free-text clinical notes) data. In particular, a common approach is to show that deep learning obtains better results than conventional machine learning models with respect to certain metrics, such as Area under the Receiver Operating Characteristic Curve, accuracy and F-score.
Several works applied deep learning to predict diseases from the patient clinical status. Cheng et al. [302] used a four-layer CNN to predict congestive heart failure and chronic obstructive pulmonary disease and showed significant advantages over the baselines. RNNs with long short-term memory (LSTM) hidden units, pooling and word embedding were used in DeepCare [387], an end-to-end deep dynamic network that infers current illness states and predicts future medical outcomes.
The authors also proposed to moderate the LSTM unit with a decay effect to handle irregular timed events (which are typically in longitudinal EHRs). Moreover, they incorporated medical interventions in the model to dynamically shape the predictions. DeepCare was evaluated for disease progression modeling, intervention recommendation and future risk prediction of diabetes and mental health patient cohorts. RNNs with gated recurrent unit (GRU) were used by Choi et al. [388] to develop Doctor AI, an end-to-end model that uses patient history to predict diagnoses and medications for subsequent encounters. The evaluation showed significantly higher recall than shallow baselines and good generalizability by adapting the resulting model from one institution to another without losing substantial accuracy. Differently, Miotto et al. [389] proposed to learn deep patient representations from the EHRs using a three-layer Stacked Denoising Autoencoder (SDA). They applied this novel representation on disease risk prediction using random forest as classifiers. The evaluation was performed on 76 214 patients comprising 78 diseases from diverse clinical domains and temporal windows (up to a 1 year). The results showed that the deep representation leads to significantly better predictions than using raw EHRs or conventional representation learning algorithms (e.g. Principal Component Analysis (PCA), k-means). Moreover, they also showed that results significantly improve when adding a logistic regression layer on top of the last AE to fine-tune the entire supervised network [390]. Similarly, Liang et al. [391] used RBMs to learn representations from EHRs that revealed novel concepts and demonstrated better prediction accuracy on a number of diseases.
Deep learning was also applied to model continuous time signals, such as laboratory results, toward the automatic identification of specific phenotypes. For example, Lipton et al. [392] used RNNs with LSTM to recognize patterns in multivariate time series of clinical measurements. Specifically, they trained a model to classify 128 diagnoses from 13 frequently, but irregularly sampled clinical measurements from patients in pediatric intensive unit care. The results showed significant improvements with respect to several strong baselines, including multilayer perceptron trained in hand-engineered features. Che et al. [393] used SDAs regularized with a prior knowledge based on ICD-9s for detecting characteristic patterns of physiology in clinical time series. Lasko et al. [394] used a two-layer stacked AE (without regularization) to model longitudinal sequences of serum uric acid measurements to distinguish the uric-acid signatures of gout and acute leukemia. Razavian et al. [395] evaluated CNNs and RNNs with LSTM units to predict disease onset from laboratory test measures alone, showing better performances than logistic regression with hand-engineered, clinically relevant features. Neural language deep models were also applied to EHRs, in particular to learn embedded representations of medical concepts, such as diseases, medications and laboratory tests that could be used for analysis and prediction [396]. As an example, Tran et al. [397] used RBMs to learn abstractions about ICD-10 codes on a cohort of 7578 mental health patients to predict suicide risk. A deep architecture based on RNNs also obtained promising results in removing protected health information from clinical notes to leverage the automatic de-identification of free-text patient summaries [398]. The prediction of unplanned patient readmissions after discharge recently received attention as well. In this domain, Nguyen et al. [399] proposed Deepr, an end-to-end architecture based on CNNs, which detects and combines clinical motifs in the longitudinal patient EHRs to stratify medical risks. Deepr performed well in predicting readmission within 6 months and was able to detect meaningful and interpretable clinical patterns.
5.6 Challenges and opportunities
Even though the promising outcome obtained using deep architectures, there stay a few strange difficulties confronting the clinical use of deep learning to health care. Specifically, the following main issues should be considered:
-
Volume of data: Deep learning consists of a set of highly comprehensive computational models. One typical example is fully connected multi-layer neural networks, where tons of network parameters need to be estimated properly. The basis to achieve this goal is the availability of huge amounts of data. In fact, while there are no hard guidelines about the minimum number of training documents, a general rule of thumb is to have at least about 10 × the number of samples as parameters in the network. This is also one of the reasons why deep learning is so successful in domains where huge amount of data can be easily collected (e.g. computer vision, speech, natural language). However, health care is a different domain; in fact, we only have approximately 7.5 billion people all over the world (as per September 2016), with a great part not having access to primary health care. Consequently, from a big data perspective, the amount of medical data that is needed to train an effective and robust, deep learning model would be much more comparable with other media.
-
Data quality: Unlike other domains where the data are clean and well-structured, health care data are highly heterogeneous, ambiguous, noisy and incomplete. Training a good deep learning model with such massive and variegate data sets is challenging and needs to consider several issues, such as data sparsity, redundancy and missing values.
-
Temporality: The diseases are always progressing and changing over time in a nondeterministic way. However, many existing deep learning models, including those already proposed in the medical domain, assume static vector-based inputs, which cannot handle the time factor in a natural way. Designing deep learning approaches that can handle temporal health care data is an important aspect that will require the development of novel solutions.
-
Domain complexity: Different from other application domains (e.g. image and speech analysis), the problems in biomedicine and health care are more complicated. The diseases are highly heterogeneous and for most of the diseases there is still no complete knowledge of their causes and how they progress. Moreover, the number of patients is usually limited in a practical clinical scenario and we cannot ask for as many patients as we want.
-
Interpretability: Although deep learning models have been successful in quite a few application domains, they are often treated as black boxes. While this might not be a problem in other more deterministic domains such as image annotation (because the end user can objectively validate the tags assigned to the images), in health care, not only the quantitative algorithmic performance is important, but also the reason why the algorithm works is relevant.
All these challenges introduce several opportunities and future research possibilities to improve the field. Therefore, with all of them in mind, we point out the following directions, which we believe would be promising for the future of deep learning in health care.
-
Feature enrichment: Because of the limited amount of patients in the world, we should capture as many features as possible to characterize each patient and find novel methods to jointly process them. The data sources for generating those features need to include, but not to be limited to, EHRs, social media (e.g. there is prior research leveraging patient-reported information on social media for pharmacovigilance), wearable devices, environments, surveys, online communities, genome profiles, and omics data such as proteome and so on. The effective integration of such highly heterogeneous data and how to use them in a deep learning model would be an important and challenging research topic.
-
Federated inference: Each clinical institution possesses its own patient population. Building a deep learning model by leveraging the patients from different sites without leaking their sensitive information becomes a crucial problem in this setting. Consequently, learning deep model in this federated setting in a secure way will be another important research topic, which will interface with other mathematical domains, such as cryptography (e.g. homomorphic encryption and secure multiparty computation).
-
Model privacy: Privacy is an important concern in scaling up deep learning (e.g. through cloud computing services). Machine Learning (ML)-as-a-service (i.e. ‘predictive analytics’) on a set of common models including deep neural networks. The deployment of intelligent tools for next-generation health care needs to consider these risks and attempt to implement a differential privacy standard.
-
Incorporating expert knowledge: The existing expert knowledge of medical problems is invaluable for health care problems. Because of the limited amount of medical data and their various quality problems, incorporating the expert knowledge into the deep learning process to guide it toward the right direction is an important research topic. For example, the online medical encyclopedia and PubMed abstracts should be mined to extract reliable content that can be included in the deep architecture to leverage the overall performances of the systems. Also semi-supervised learning, an effective scheme to learn from the large amount of unlabeled samples with only a few labeled samples, would be of great potential because of its capability of leveraging both labeled (which encodes the knowledge) and unlabeled samples .
-
Temporal modeling: Considering that the time factor is important in all kinds of health care-related problems, in particular in those involving EHRs and monitoring devices, training a time-sensitive deep learning model is critical for a better understanding of the patient condition and for providing timely clinical decision support. Thus, temporal deep learning is crucial for solving health care problems. It expects that RNNs as well as architectures coupled with memory and attention mechanisms will play a more significant role toward better clinical deep architectures.
-
Interpretable modeling: Model performance and interpretability are equally important for health care problems. Clinicians are unlikely to adopt a system they cannot understand. Deep learning models are popular because of their superior performance. Yet, how to explain the results obtained from these models and how to make them more understandable is of key importance toward the development of trustable and reliable systems. Deep learning methods are powerful tools that allow computers to learn from the data, so that they can come up with ways to create smarter applications. These approaches have already been used in a number of applications, especially for computer vision and natural language processing. In fact, processing medical data with multi-layer neural networks increased the predictive power for several specific applications in different clinical domains.
5.7 Research question or hypothesis
-
What extend the research work has to be done in the diabetic prediction model based on different machine and deep learning algorithms from the last two decades?
-
What different types of algorithms are used by the researcher in their work in the prediction model?
-
What are the different performance parameters that can influence the performance of a prediction model?
-
How to enhance the accuracy rate of the prediction model even in the presence of not enough medical data?
-
What are the different optimization technique are used in a prediction model to make the system more robust?
-
What are the different classification algorithm are used in the prediction model to distinguish between diabetic and non-diabetic patients?
-
There is a very limited work in the Autoencoder (AE) based deep learning approach for diabetic prediction model from the last decade, How can improved the performance of the prediction model using AE?
-
How we can make the diabetic prediction model is more reliable and secure?
-
How can we improve the accuracy rate of diabetic prediction model based on different clustering approach?
-
How can we enhance the accuracy rate in a Smartphone enabled diabetic prediction model by using different machines and deep learning approach?
-
How the diabetic prediction model based on different association rule approach can be improved the accuracy rate?
-
How cloud based diabetic prediction model make more stable in terms of classification accuracy, Sensitivity rate, Specificity rate and provides overall true diabetic prediction model.?
-
In future work, what classification algorithms and technique, it should apply to make the prediction model more secure and reliable.
-
How can we make deep learning and machine learning based diabetic prediction model more secured using different cryptography technique? (Fig. 5).
6 Conclusion
Machine and deep learning based algorithms play an important role to enhance the overall performance of the diabetic prediction system, in which different classification algorithms are utilized efficiently to form a better prediction system. The proper use of machine or deep learning based algorithms is very important in the diabetic prediction system because it can affect the overall performance and accuracy level of the systems. In designing a diabetic-prediction based system, it is very important, how it can design the classifier for the detection of Diabetes disease with optimal cost and better performance.
This paper is an in-depth study on diabetic prediction strategy, including machine learning (Supervise learning like SVM, KNN, RF, DT, NB and Regression, Unsupervised learning like clustering and association rule) and deep learning (CNN, RNN, AE, RBF and ANN) approaches and its applications in the areas of user diabetic based health prediction model. The main reason behind the success of any health related prediction system are totally depends on the machine and deep learning based algorithm methodology. The main focus of this study is to discuss the methodology and approaches or algorithms used in different diabetic prediction model to enhance the performance of the system and compare the results of related works based on the different diabetic prediction system. Although the prediction model of diabetic disease based on the machine and deep learning approach has greatly improved over the last decade, different prediction algorithms and its approach of implementation has been improved in terms of Classification accuracy, enhance the Sensitivity rate, enhance Specificity rate and provide overall true diabetic prediction model., a lot of work still needs to be done to make the prediction process more practicable so that it can be easily applied on different application of health care system.
It is also discussing the strength and weaknesses of various research works in the area of the diabetic prediction system and providing a comprehensive review of the system based on machine and deep learning. Furthermore, this review paper provides a comparative discussion which represents the different algorithms used in the previous research works since from the year 2010 to 2021 in the area of the diabetic prediction model which can help to find out how proper used a machine and deep learning based algorithm in the diabetic prediction model improved from year to year and in the future how it can make the model more secure by using the proper technique. It discusses the classification of machine learning based algorithms utilized in the area of diabetic prediction systems, namely supervised and unsupervised. The main problem associated with mostly diabetic-prediction system is that the optimization process which required extra time for computation and this will affect the prediction model, choosing an inappropriate optimization technique may result in a very low accuracy rate and affect the overall performance of the model. Hence this review paper is an in depth study of various machine and affect the overall performance of the model. Hence this review paper is an in depth study of various machines and deep learning based algorithms used in the field of the diabetic prediction system and has made clear why more research needs to be done to find a solution to the stated problems found in various diabetic prediction system, also the shortcoming of the various prediction techniques.
Data availability
Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.
References
Bloomgarden Z (2016) Questioning glucose measurements used in the International Diabetes Federation (IDF) Atlas. J Diab 8(6):746–747. https://doi.org/10.1111/1753-0407.12453
Ming Z, Wang X, Zhu X (2014) Understanding diabetes from the diagnosis of diabetes mellitus. J Diagn Concept Pract 2:226–228
Rajesh K, Sangeetha V (2012) Application of Data Mining Methods and Techniques for Diabetes Diagnosis. Int J Eng Innov Technol 2(3):224–229
Kaveeshwar SA, Cornwall J (2014) The current state of diabetes mellitus in India. Australas Med J 7(1):45
Wild S, Roglic G, Green A, Sicree R, King H (2004) Global prevalence of diabetes: estimates for the year 2000 and projections for 2030. Diabetes Care 27(5):1047–1053
Shaw JE, Simpson RW (2009) Prevention of type 2 diabetes. Diabetes and Exercise. Springer, pp 55–62
IDF Diabetes Atlas - 8th Edition. Available from: http://www.diabetesatlas.org/across-the-globe.html. Accessed 31 Dec. 2017
Kaur P, Sharma M (2018) Analysis of data mining and soft computing techniques in prospecting diabetes disorder in human beings: a review. Int J Pharm Sci Res 9:2700–2719
Sun YL, Zhang DL (2019) Machine learning techniques for screening and diagnosis of diabetes: a survey. Tehnički Vjesnik 26(3):872–880
Yoo I, Alafaireet P, Marinov M, Pena-Hernandez K, Gopidi R, Chang JF, Hua L (2012) Data mining in healthcare and biomedicine: a survey of the literature. J Med Syst 36(4):2431–2448
Fatima M, Pasha M (2017) Survey of machine learning algorithms for disease diagnostic. J Intell Learn Syst Appl 9(01):1
Kaur H, Kumari V (2020) Predictive modelling and analytics for diabetes using machine learning approach. Appl Comput Inform 18(1/2):90–100
Javitt JC, Aiello LP, Chiang Y, Ferris FL, Canner JK, Greenfield S (1994) Preventive eye care in people with diabetes is cost-saving to the federal government: implications for health-care reform. Diabetes Care 17(8):909–917
Mendonca AM, Campilho AJ, Nunes JM (1999) Automatic segmentation of microaneurysms in retinal angiograms of diabetic patients. In Proceedings 10th International Conference on Image Analysis and Processing. IEEE. pp 728-733
Cree MJ, Olson JA, McHardy KC, Forrester JV, Sharp PF (1996) Automated microaneurysm detection. In Proceedings of 3rd IEEE International Conference on Image Processing. vol. 3. IEEE. pp 699-702
Zhang X, Chutatape O (2005) A SVM approach for detection of hemorrhages in background diabetic retinopathy. In Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005. vol. 4. IEEE. pp 2435-2440
Stoean R, Stoean C, Preuss M, El-Darzi E, Dumitrescu D (2006) Evolutionary support vector machines for diabetes mellitus diagnosis. In 2006 3rd International IEEE Conference Intelligent Systems. IEEE. pp 182-187
Balakrishnan S, Narayanaswamy R, Savarimuthu N, Samikannu R (2008) SVM ranking with backward search for feature selection in type II diabetes databases. In 2008 IEEE International Conference on Systems, Man and Cybernetics. IEEE. pp 2628-2633
Polat K, Güneş S, Arslan A (2008) A cascade learning system for classification of diabetes disease: Generalized discriminant analysis and least square support vector machine. Expert Syst Appl 34(1):482–487
Wu J, Diao YB, Li ML, Fang YP, Ma DC (2009) A semi-supervised learning based method: Laplacian support vector machine used in diabetes disease diagnosis. Interdiscip Sci: Comput Life Sci 1(2):151–155
Yu W, Liu T, Valdez R, Gwinn M, Khoury MJ (2010) Application of support vector machine modeling for prediction of common diseases: the case of diabetes and pre-diabetes. BMC Med Inform Decis Mak 10(1):1–7
Barakat N, Bradley AP, Barakat MNH (2010) Intelligible support vector machines for diagnosis of diabetes mellitus. IEEE Trans Inf Technol Biomed 14(4):1114–1120
Çalişir D, Doğantekin E (2011) An automatic diabetes diagnosis system based on LDA-Wavelet Support Vector Machine Classifier. Expert Syst Appl 38(7):8311–8315
Gupta S, Kumar D, Sharma A (2011) Performance analysis of various data mining classification techniques on healthcare data. Int J Comput Sci Inform Technol 3(4):155–169
Marling C, Wiley M, Cooper T, Bunescu R, Shubrook J, Schwartz F (2011) The 4 diabetes support system: a case study in CBR research and development. In International Conference on Case-Based Reasoning. Springer, Berlin, Heidelberg. pp 137-150
Zolfaghari R (2012) Diagnosis of diabetes in female population of pima indian heritage with ensemble of bp neural network and svm. Int J Comput Eng Manag 15:2230–7893
Giveki D, Salimi H, Bahmanyar G, Khademian Y (2012) Automatic detection of diabetes diagnosis using feature weighted support vector machines based on mutual information and modified cuckoo search. arXiv preprint arXiv:1201.2173
Hashim MF, Hashim SZM (2012) Comparison of clinical and textural approach for Diabetic Retinopathy grading. In 2012 IEEE International Conference on Control System, Computing and Engineering. IEEE. pp 290-295
Karatsiolis S, Schizas CN (2012) Region based Support Vector Machine algorithm for medical diagnosis on Pima Indian Diabetes dataset. In 2012 IEEE 12th International Conference on Bioinformatics & Bioengineering (BIBE). IEEE. pp 139-144
Kumari VA, Chitra R (2013) Classification of diabetes disease using support vector machine. Int J Eng Res Appl 3(2):1797–1801
Farran B, Channanath AM, Behbehani K, Thanaraj TA (2013) Predictive models to assess risk of type 2 diabetes, hypertension and comorbidity: machine-learning algorithms and validation using national health data from Kuwait—a cohort study. BMJ Open 3(5):e002457
Mansour RF, Abdelrahim EM and Al-Johani AS (2013) Identification of diabetic retinal exudates in digital color images using support vector machine
Tapak L, Mahjub H, Hamidi O, Poorolajal J (2013) Real-data comparison of data mining methods in prediction of diabetes in Iran. Healthc Inform Res 19(3):177
Anthimopoulos MM, Gianola L, Scarnato L, Diem P, Mougiakakou SG (2014) A food recognition system for diabetic patients based on an optimized bag-of-features model. IEEE J Biomed Health Inform 18(4):1261–1271
Choi SB, Kim WJ, Yoo TK, Park JS, Chung JW, Lee YH, Kang EU, Kim DW (2014) Screening for prediabetes using machine learning models. Comput Math Methods Med 2014(1):618976
Roychowdhury S, Koozekanani DD, Parhi KK (2014) DREAM: diabetic retinopathy analysis using machine learning. IEEE J Biomed Health Inform 18(5):1717–1728
Cai L, Wu H, Li D, Zhou K, Zou F (2015) Type 2 diabetes biomarkers of human gut microbiota selected via iterative sure independent screening method. PLoS One 10(10):e0140827
Jaya T, Dheeba J, Singh NA (2015) Detection of hard exudates in colour fundus images using fuzzy support vector machine-based expert system. J Digit Imaging 28(6):761–768
Arjun C, Anto M (2015) Diagnosis of diabetes using support vector machine and ensemble learning approach. Int J Eng Appl Sci 2(11):257790
Kang S, Kang P, Ko T, Cho S, Rhee SJ, Yu KS (2015) An efficient and effective ensemble of support vector machines for anti-diabetic drug failure prediction. Expert Syst Appl 42(9):4265–4273
Ramanathan TT, Sharma D (2015) An SVM-Fuzzy Expert System design for diabetes risk classification. Int J Comput Sci Inform Technol 6(3):2221–2226
Santhanam T, Padmavathi MS (2015) Application of K-means and genetic algorithms for dimension reduction by integrating SVM for diabetes diagnosis. Procedia Comput Sci 47:76–83
Sowjanya K, Singhal A, Choudhary C (2015) MobDBTest: A machine learning based system for predicting diabetes risk using mobile devices. In 2015 IEEE International Advance Computing Conference (IACC). IEEE. pp 397-402
Tafa Z, Pervetica N, Karahoda B (2015) An intelligent system for diabetes prediction. In 2015 4th Mediterranean Conference on Embedded Computing (MECO). IEEE. pp 378-382
Abdillah AA, Suwarno S (2016) Diagnosis of diabetes using support vector machines with radial basis function kernels. Int J Technol 7(5)
Bano S, Khan MNA (2016) A Framework to Improve Diabetes Prediction using k-NN and SVM. Int J Comput Sci Inform Sec 14(11):450
Gill NS, Mittal P (2016) A computational hybrid model with two level classification using SVM and neural network for predicting the diabetes disease. J Theor Appl Inf Technol 87(1):1–10
Huang YP, Nashrullah M (2016) SVM-based Decision Tree for medical knowledge representation. In 2016 International Conference on Fuzzy Theory and Its Applications (iFuzzy). IEEE. pp 1-6
Kose U, Guraksin GE, Deperlioglu O (2016) Cognitive development optimization algorithm based support vector machines for determining diabetes. Broad Res Artif Intell Neurosci 7(1):80–90
Malik S, Khadgawat R, Anand S, Gupta S (2016) Non-invasive detection of fasting blood glucose level via electrochemical measurement of saliva. Springerplus 5(1):1–12
Negi A, Jaiswal V (2016) A first attempt to develop a diabetes prediction method based on different global datasets. In 2016 Fourth International Conference on Parallel, Distributed and Grid Computing (PDGC). IEEE. pp 237-241
Osman AH, Aljahdali HM (2017) Diabetes disease diagnosis method based on feature extraction using K-SVM. Int J Adv Comput Sci Appl 8(1)
Carrera EV, González A, Carrera R (2017) Automated detection of diabetic retinopathy using SVM. In 2017 IEEE XXIV international conference on electronics, electrical engineering and computing (INTERCON). IEEE. pp 1-4
Khalil RM, Al-Jumaily A (2017) Machine learning based prediction of depression among type 2 diabetic patients. In 2017 12th international conference on intelligent systems and knowledge engineering (ISKE). IEEE. pp 1-5
Rathore A, Chauhan S, Gujral S (2017) Detecting and predicting diabetes using supervised learning: an approach towards better healthcare for women. Int J Adv Res Comput Sci 8(5)
Wang Y, Liu ZP (2017) Identifying biomarkers of diabetes with gene co expression networks. In 2017 Chinese Automation Congress (CAC). IEEE. pp 5283-5286
Zhang J, Xu J, Hu X, Chen Q, Tu L, Huang J, Cui J (2017) Diagnostic method of diabetes based on support vector machine and tongue images. BioMed Res Int 2017(1):7961494
Cui S, Wang D, Wang Y, Yu PW, Jin Y (2018) An improved support vector machine-based diabetic readmission prediction. Comput Methods Prog Biomed 166:123–135
Dagliati A, Marini S, Sacchi L, Cogni G, Teliti M, Tibollo V, … Bellazzi R (2018) Machine learning methods to predict diabetes complications. J Diabetes Sci Technol 12(2):295–302
Joshi TN, Chawan PM (2018) Logistic regression and svm based diabetes prediction system. Int J Technol Res Eng 5:4347–4350
Rao NM, Kannan K, Gao XZ, Roy DS (2018) Novel classifiers for intelligent disease diagnosis with multi-objective parameter evolution. Comput Electr Eng 67:483–496
Mule DB, Chowhan SS, Somwanshi DR (2018) Detection and classfication of non-proliferative diabetic retinopathy using retinal images. In International Conference on Recent Trends in Image Processing and Pattern Recognition. Springer, Singapore. pp 312-320
Abdullah AS, Gayathri N, Selvakumar S, Kumar SR (2018) Identification of the Risk Factors of Type II Diabetic Data Based Support Vector Machine Classifiers upon Varied Kernel Functions. In Computational Vision and Bio Inspired Computing. Springer, Cham. pp 496-505
Sisodia D, Sisodia DS (2018) Prediction of diabetes using classification algorithms. Procedia Comput Sci 132:1578–1585
Tsao HY, Chan PY, Su ECY (2018) Predicting diabetic retinopathy and identifying interpretable biomedical features using machine learning algorithms. BMC Bioinform 19(9):111–121
Alirezaei M, Niaki STA, Niaki SAA (2019) A bi-objective hybrid optimization algorithm to reduce noise and data dimension in diabetes diagnosis using support vector machines. Expert Syst Appl 127:47–57
Bernardini M, Romeo L, Misericordia P, Frontoni E (2019) Discovering the type 2 diabetes in electronic health records using the sparse balanced support vector machine. IEEE J Biomed Health Inform 24(1):235–246
Raj RS, Sanjay DS, Kusuma M, Sampath S (2019) Comparison of support vector machine and Naive Bayes classifiers for predicting diabetes. In 2019 1st International Conference on Advanced Technologies in Intelligent Control, Environment, Computing & Communication Engineering (ICATIECE). IEEE. pp. 41-45
He K, Huang S, Qian X (2019) Early detection and risk assessment for chronic disease with irregular longitudinal data analysis. J Biomed Inform 96:103231
Karkuzhali S, Manimegalai D (2019) Distinguising Proof of Diabetic Retinopathy Detection by Hybrid Approaches in Two Dimensional Retinal Fundus Images. J Med Syst 43(6):1–12
Abbas HT, Alic L, Erraguntla M, Ji JX, Abdul-Ghani M, Abbasi QH, Qaraqe MK (2019) Predicting long-term type 2 diabetes with support vector machine using oral glucose tolerance test. PLoS One 14(12):e0219636
Lokuarachchi D, Muthumal L, Gunarathna K, Gamage TD (2019) Detection of red lesions in retinal images using image processing and machine learning techniques. In 2019 Moratuwa Engineering Research Conference (MERCon). IEEE. pp 550-555
Aminah R, Saputro AH (2019) Application of machine learning techniques for diagnosis of diabetes based on iridology. In 2019 International Conference on Advanced Computer Science and information Systems (ICACSIS). IEEE. pp 133-138
Qomariah DUN, Tjandrasa H, Fatichah C (2019) Classification of diabetic retinopathy and normal retinal images using CNN and SVM. In 2019 12th International Conference on Information & Communication Technology and System (ICTS). IEEE. pp 152-157
Selvathi D, Suganya K (2019) Support vector machine based method for automatic detection of diabetic eye disease using thermal images. In 2019 1st International Conference on Innovations in Information and Communication Technology (ICIICT). IEEE. pp 1-6
Sneha N, Gangil T (2019) Analysis of diabetes mellitus for early prediction using optimal features selection. J Big Data 6(1):1–19
Hao Y, Cheng F, Pham M, Rein H, Patel D, Fang Y, … Wang Y (2019) A Noninvasive, Economical, and Instant-Result Method to Diagnose and Monitor Type 2 Diabetes Using Pulse Wave: Case-Control Study. JMIR MHealth UHealth 7(4):e11959
Azad C, Mehta AK, Mahto D, Yadav DK (2020) Support Vector Machine based eHealth Cloud System for Diabetes Classification. EAI Endorsed Trans Pervasive Health Technol 6(22):e3
Harimoorthy K, Thangavelu M (2020) Multi-disease prediction model using improved SVM-radial bias technique in healthcare monitoring system. J Ambient Intell Humaniz Comput 12(3):3715–3723
Jayabalan S, Pratheeksha PS, Bolar NS, Malavika NL (2020) Prediction of diabetic retinopathy using svm algorithm. J Crit Rev 7(14):1702–1711
Kazerouni F, Bayani A, Asadi F, Saeidi L, Parvizi N, Mansoori Z (2020) Type2 diabetes mellitus prediction using data mining algorithms based on the long-noncoding RNAs expression: a comparison of four data mining approaches. BMC Bioinformatics 21(1):1–13
Nnamoko N, Korkontzelos I (2020) Efficient treatment of outliers and class imbalance for diabetes prediction. Artif Intell Med 104:101815
Shuja M, Mittal S, Zaman M (2020) Effective prediction of type ii diabetes mellitus using data mining classifiers and SMOTE. In Advances in computing and intelligent systems. Springer, Singapore. pp 195-211
Mishra SK, Tiwari AK (2020) An Ensemble Approach for the Prediction of Diabetes. SAMRIDDHI 12(02):122–129
Viloria A, Herazo-Beltran Y, Cabrera D, Pineda OB (2020) Diabetes diagnostic prediction using vector support machines. Procedia Comput Sci 170:376–381
Wang X, Yang Y, Xu Y, Chen Q, Wang H, Gao H (2020) Predicting hypoglycemic drugs of type 2 diabetes based on weighted rank support vector machine. Knowl-Based Syst 197:105868
Srivastava AK, Kumar Y, Singh PK (2020) Computer aided diagnostic system based on SVM and K harmonic mean based attribute weighting method. Obes Med 19:100270
Xue J, Min F, Ma F (2020) Research on Diabetes Prediction Method Based on Machine Learning. In Journal of Physics: Conference Series. vol. 1684, no. 1. IOP Publishing. p 012062
Ahmad HF, Mukhtar H, Alaqail H, Seliaman M, Alhumam A (2021) Investigating Health-Related Features and Their Impact on the Prediction of Diabetes Using Machine Learning. Appl Sci 11(3):1173
Alabdulwahhab KM, Sami W, Mehmood T, Meo SA, Alasbali TA, Alwadani FA (2021) Automated detection of diabetic retinopathy using machine learning classifiers. Eur Rev Med Pharmacol Sci 25(2):583–590
Chaves L, Marques G (2021) Data Mining Techniques for Early Diagnosis of Diabetes: A Comparative Study. Appl Sci 11(5):2218
Dinesh MG, Prabha D (2021) Diabetes Mellitus Prediction System Using Hybrid KPCA-GA-SVM Feature Selection Techniques. J Phys Conf Ser. 1767(1):012001
Khanam JJ, Foo SY (2021) A comparison of machine learning algorithms for diabetes prediction. ICT Express
Reddy SS, Sethi N, Rajender R (2021) Discovering Optimal Algorithm to Predict Diabetic Retinopathy using Novel Assessment Methods. EAI Endorsed Trans Scalable Inf Syst 8(29):e1
Rodríguez-Rodríguez I, Rodríguez JV, Woo WL, Wei B, Pardo-Quiles DJ (2021) A Comparison of Feature Selection and Forecasting Machine Learning Algorithms for Predicting Glycaemia in Type 1 Diabetes Mellitus. Appl Sci 11(4):1742
Tang H, Zhang Y, Xiang B, Liu M, Hu J, Liu C (2021) Risk prediction of early diabetes mellitus based on combination model. In MATEC Web of Conferences. vol. 336, EDP Sciences. p 07018
Hossain ME, Uddin S, Khan A (2021) Network analytics and machine learning for predictive risk modelling of cardiovascular disease in patients with type 2 diabetes. Expert Syst Appl 164:113918
Senthil Velmurugan N, Viveka T (2021) Performance analysis of ML algorithms on diabetes data. Int Adv Res J Sci Eng Technol 8(2):72–79
Suresh K, Obulesu O, Ramudu BV (2020) Diabetes Prediction using Machine Learning Techniques. Helix 10(02):136–142
Brisimi TS, Xu T, Wang T, Dai W, Adams WG, Paschalidis IC (2018) Predicting chronic disease hospitalizations from electronic health records: an interpretable classification approach. Proc IEEE 106(4):690–707
Baitharu TR, Pani SK, Dhal S (2015) Comparison of Kernel selection for support vector machines using diabetes dataset. J Comput Sci Appl 3(6):181–184
Breault JL, Goodall CR, Fos PJ (2002) Data mining a diabetic data warehouse. Artif Intell Med 26(1-2):37–54
Miyaki K, Takei I, Watanabe K, Nakashima H, Watanabe K, Omae K (2002) Novel statistical classification model of type 2 diabetes mellitus patients for tailormade prevention using data mining algorithm. J Epidemiol 12(3):243–248
Duhamel A, Nuttens MC, Devos P, Picavet M, Beuscart R (2003) A preprocessing method for improving data mining techniques. Appl Large Med Diab Database Stud Health Technol Inform 95:269–274
Huang Y, McCullagh P, Black N and Harper R (2004) Evaluation of outcome prediction for a clinical diabetes database. In International Symposium on Knowledge Exploration in Life Science Informatics (pp. 181-190). Springer, Berlin, Heidelberg
Harper PR (2005) A review and comparison of classification algorithms for medical decision making. Health Policy 71(3):315–331
Sigurdardottir AK, Jonsdottir H, Benediktsson R (2007) Outcomes of educational interventions in type 2 diabetes: WEKA data-mining analysis. Patient Educ Couns 67(1-2):21–31
Huang Y, McCullagh P, Black N, Harper R (2007) Feature selection and classification model construction on type 2 diabetic patients’ data. Artif Intell Med 41(3):251–262
Liou FM, Tang YC, Chen JY (2008) Detecting hospital fraud and claim abuse through diabetic outpatient services. Health Care Manag Sci 11(4):353–358
Toussi M, Lamy JB, Le Toumelin P, Venot A (2009) Using data mining techniques to explore physicians' therapeutic decisions when clinical guidelines do not provide recommendations: methods and example for type 2 diabetes. BMC Med Inform Decis Mak 9(1):1–12
Hische M, Luis-Dominguez O, Pfeiffer AF, Schwarz PE, Selbig J, Spranger J (2010) Decision trees as a simple-to-use and reliable tool to identify individuals with impaired glucose metabolism or type 2 diabetes mellitus. Eur J Endocrinol 163(4):565
Patil BM, Joshi RC, Toshniwal D (2010) Hybrid prediction model for type-2 diabetic patients. Expert Syst Appl 37(12):8102–8108
Ahmad A, Mustapha A, Zahadi ED, Masah N, Yahaya NY (2011) Comparison between neural networks against decision tree in improving prediction accuracy for diabetes mellitus. In International conference on digital information processing and communications. Springer, Berlin, Heidelberg. pp 537-545
Al Jarullah AA (2011) Decision tree discovery for the diagnosis of type II diabetes. In 2011 International conference on innovations in information technology. IEEE 303-307
Karegowda AG, Manjunath AS, Jayaram MA (2011) Application of genetic algorithm optimized neural network connection weights for medical diagnosis of pima Indians diabetes. Int J Soft Comput 2(2):15–23
Kelarev AV, Stranieri A, Yearwood JL, Jelinek HF (2012) Empirical study of decision trees and ensemble classifiers for monitoring of diabetes patients in pervasive healthcare. In 2012 15th International Conference on Network-Based Information Systems. IEEE. pp 441-446
Hemant P, Pushpavathi T (2012) A novel approach to predict diabetes by Cascading Clustering and Classification. In 2012 Third International Conference on Computing, Communication and Networking Technologies (ICCCNT'12). IEEE. pp 1-7
Hussein AS, Omar WM, Li X, Ati M (2012) Efficient chronic disease diagnosis prediction and recommendation system. In 2012 IEEE-EMBS Conference on Biomedical Engineering and Sciences. IEEE. pp 209-214
Rajesh K, Sangeetha V (2012) Application of data mining methods and techniques for diabetes diagnosis. Int J Eng Innov Technol 2(3):224–229
Li CP, Zhi XY, Jun MA, Zhuang CUI, Zhu ZL, Zhang C, Hu LP (2012) Performance comparison between logistic regression, decision trees, and multilayer perceptron in predicting peripheral neuropathy in type 2 diabetes mellitus. Chin Med J 125(5):851–857
Chen H, Tan C (2012) Prediction of type-2 diabetes based on several element levels in blood and chemometrics. Biol Trace Elem Res 147(1):67–74
Karegowda AG, Punya V, Jayaram MA, Manjunath AS (2012) Rule based classification for diabetic patients using cascaded k-means and decision tree C4. 5. Int J Comput Appl 45(12):45–50
Karthikeyani V, Begum IP, Tajudin K, Begam IS (2012) Comparative of data mining classification algorithm (CDMCA) in diabetes disease prediction. Int J Comput Appl 60(12)
Ameri H, Alizadeh S, Barzegari A (2013) Knowledge extraction of diabetics' data by decision tree method. J Healthc Adm 16(53):58–72
Meng XH, Huang YX, Rao DP, Zhang Q, Liu Q (2013) Comparison of three data mining models for predicting diabetes or prediabetes by risk factors. Kaohsiung J Med Sci 29(2):93–99
Karthikeyani V, Begum IP (2013) Comparison a performance of data mining algorithms (CPDMA) in prediction of diabetes disease. Int J Comput Sci Eng 5(3):205
Rahman RM, Afroz F (2013) Comparison of various classification techniques using different data mining tools for diabetes diagnosis. J Softw Eng Appl 6(03):85
Varma KV, Rao AA, Lakshmi TSM, Rao PN (2014) A computational intelligence approach for a better diagnosis of diabetic patients. Comput Electr Eng 40(5):1758–1765
Kaur G, Chhabra A (2014) Improved J48 classification algorithm for the prediction of diabetes. Int J Comput Appl 98(22):13–17
Seera M, Lim CP (2014) A hybrid intelligent system for medical data classification. Expert Syst Appl 41(5):2239–2249
Uppin S, Anusuya MA (2014) Expert system design to predict heart and diabetes diseases. Int J Sci EngTechnol 3(8):1054–1059
Ramezankhani A, Pournik O, Shahrabi J, Khalili D, Azizi F, Hadaegh F (2014) Applying decision tree for identification of a low risk population for type 2 diabetes. Tehran Lipid and Glucose Study. Diabetes Res Clin Pract 105(3):391–398
Bashir S, Qamar U, Khan FH, Javed MY (2014) An efficient rule-based classification of Diabetes using ID3, C4. 5, & CART ensembles. In 2014 12th International Conference on Frontiers of Information Technology. IEEE. pp 226-231
Habibi S, Ahmadi M, Alizadeh S (2015) Type 2 diabetes mellitus screening and risk factors using decision tree: results of data mining. Global J Health Sci 7(5):304
Kandhasamy JP, Balamurali SJPCS (2015) Performance analysis of classifier models to predict diabetes mellitus. Procedia Comput Sci 47:45–51
Iyer A, Jeyalatha S, Sumbaly R (2015) Diagnosis of diabetes using classification mining techniques. arXiv preprint arXiv:1502.03774
Vijayan VV, Anjali C (2015) Prediction and diagnosis of diabetes mellitus—A machine learning approach. In 2015 IEEE Recent Advances in Intelligent Computational Systems (RAICS). IEEE. pp 122-127
Thirumal PC, Nagarajan N (2015) Utilization of data mining techniques for diagnosis of diabetes mellitus-a case study. ARPN J Eng Appl Sci 10(1):8–13
Nai-arun N, Moungmai R (2015) Comparison of classifiers for the risk of diabetes prediction. Procedia Comput Sci 69:132–142
Heydari M, Teimouri M, Heshmati Z, Alavinia SM (2016) Comparison of various classification algorithms in the diagnosis of type 2 diabetes in Iran. Int J Diabetes Dev Ctries 36(2):167–173
Ahmed TM (2016) Developing a predicted model for diabetes type 2 treatment plans by using data mining. J Theor Appl Inf Technol 90(2):181
Ahmed TM (2016) Using data mining to develop model for classifying diabetic patient control level based on historical medical records. J Theor Appl Inf Technol 87(2):316
Daghistani T, Alshammari R (2016) Diagnosis of diabetes by applying data mining classification techniques. Int J Adv Comput Sci Appl 7(7):329–332
Orabi KM, Kamal YM, Rabah TM (2016) Early predictive system for diabetes mellitus disease. In Industrial Conference on Data Mining. Springer, Cham. pp 420-427
Perveen S, Shahbaz M, Guergachi A, Keshavjee K (2016) Performance analysis of data mining classification techniques to predict diabetes. Procedia Comput Sci 82:115–121
Pradeep KR, Naveen NC (2016) Predictive analysis of diabetes using J48 algorithm of classification techniques. In 2016 2nd International Conference on Contemporary Computing and Informatics (IC3I). IEEE. pp 347-352
Shetty SP, Joshi S (2016) A tool for diabetes prediction and monitoring using data mining technique. Int J Inform TechnolComput Sci 8(11):26–32
Songthung P, Sripanidkulchai K (2016) Improving type 2 diabetes mellitus risk prediction using classification. In 2016 13th International Joint Conference on Computer Science and Software Engineering (JCSSE). IEEE. pp 1-6
Srikanth P, Deverapalli D (2016) A critical study of classification algorithms using diabetes diagnosis. In 2016 IEEE 6th International Conference on Advanced Computing (IACC). IEEE. pp 245-249
Teimouri M, Farzadfar F, Alamdari MS, Hashemi-Meshkini A, Alamdari PA, Rezaei-Darzi E, … Zeynalabedini A (2016) Detecting diseases in medical prescriptions using data mining tools and combining techniques. Iranian J Pharmaceut Res 15(Suppl):113
Chen W, Chen S, Zhang H, Wu T (2017) A hybrid prediction model for type 2 diabetes using K-means and decision tree. In 2017 8th IEEE International conference on software engineering and service science (ICSESS). IEEE. pp 386-390
Kasbekar PU, Goel P, Jadhav SP (2017) A decision tree analysis of diabetic foot amputation risk in indian patients. Front Endocrinol 8:25
Sayadi M, Zibaeenezhad M, Taghi Ayatollahi SM (2017) Simple prediction of type 2 diabetes mellitus via decision tree modeling. Int Cardiovasc Res J 11(2):71–76
Yuvaraj N, SriPreethaa KR (2019) Diabetes prediction in healthcare systems using machine learning algorithms on Hadoop cluster. Clust Comput 22(1):1–9
Zou Q, Qu K, Luo Y, Yin D, Ju Y, Tang H (2018) Predicting diabetes mellitus with machine learning techniques. Front Genet 9:515
Kadhm MS, Ghindawi IW, Mhawi DE (2018) An accurate diabetes prediction system based on K-means clustering and proposed classification approach. Int J Appl Eng Res 13(6):4038–4041
Esmaily H, Tayefi M, Doosti H, Ghayour-Mobarhan M, Nezami H, Amirabadizadeh A (2018) A comparison between decision tree and random forest in determining the risk factors associated with type 2 diabetes. J Res Health Sci 18(2):412
Barhate R, Kulkarni P (2018) Analysis of classifiers for prediction of type ii diabetes mellitus. In 2018 Fourth International Conference on Computing Communication Control and Automation (ICCUBEA). IEEE. pp 1-6
Mahmud SH, Hossin MA, Ahmed MR, Noori SRH, Sarkar MNI (2018) Machine learning based unified framework for diabetes prediction. In Proceedings of the 2018 International Conference on Big Data Engineering and Technology. pp 46-50
Fiarni C, Sipayung EM, Maemunah S (2019) Analysis and prediction of diabetes complication disease using data mining algorithm. Procedia Comput Sci 161:449–457
Hebbar A, Kumar M, Sanjay HA (2019) DRAP: Decision Tree and Random Forest Based Classification Model to Predict Diabetes. In 2019 1st International Conference on Advances in Information Technology (ICAIT). IEEE. pp 271-276
Pei D, Zhang C, Quan Y, Guo Q (2019) Identification of potential type II diabetes in a Chinese population with a sensitive decision tree approach. J Diabetes Res 2019(1):4248218
Choudhury A, Gupta D (2019) A survey on medical diagnosis of diabetes using machine learning techniques. In Recent developments in machine learning and data analytics. Springer, Singapore. pp 67-78
Sun Y, Zhang D (2019) Diagnosis and analysis of diabetic retinopathy based on electronic health records. Ieee Access 7:86115–86120
Choubey DK, Kumar P, Tripathi S, Kumar S (2020) Performance evaluation of classification methods with PCA and PSO for diabetes. Netw Model Anal Health Inform Bioinform 9(1):1–30
Al-Zebari A, Sengur A (2019) Performance Comparison of Machine Learning Techniques on Diabetes Disease Detection. In 2019 1st International Informatics and Software Engineering Conference (UBMYK). IEEE. pp 1-4
Pei D, Yang T, Zhang C (2020) Estimation of Diabetes in a High-Risk Adult Chinese Population Using J48 Decision Tree Model. Diabetes Metab Syndr Obes 13:4621
Maniruzzaman M, Rahman MJ, Ahammed B, Abedin MM (2020) Classification and prediction of diabetes disease using machine learning paradigm. Health inform Sci Syst 8(1):1–14
Pranto B, Mehnaz S, Mahid EB, Sadman IM, Rahman A, Momen S (2020) Evaluating machine learning methods for predicting diabetes among female patients in bangladesh. Information 11(8):374
Tigga NP, Garg S (2020) Prediction of type 2 diabetes using machine learning classification methods. Procedia Comput Sci 167:706–716
Haq AU, Li JP, Khan J, Memon MH, Nazir S, Ahmad S, … Ali A (2020) Intelligent Machine Learning Approach for Effective Recognition of Diabetes in E-Healthcare Using Clinical Data. Sensors 20(9):2649
Taser PY (2021) Application of Bagging and Boosting Approaches Using Decision Tree-Based Algorithms in Diabetes Risk Prediction. In Multidisciplinary Digital Publishing Institute Proceedings. vol. 74, no. 1. p. 6
Chen T, Shang C, Su P, Keravnou-Papailiou E, Zhao Y, Antoniou G, Shen Q (2021) A Decision Tree-Initialised Neuro-fuzzy Approach for Clinical Decision Support. Artif Intell Med 111:101986
Emon MU, Zannat R, Khatun T, Rahman M, Keya MS (2021) Performance Analysis of Diabetic Retinopathy Prediction using Machine Learning Models. In 2021 6th International Conference on Inventive Computation Technologies (ICICT). IEEE. pp 1048-1052
Lee M, Gatton TM, Lee KK (2010) A monitoring and advisory system for diabetes patient management using a rule-based method and KNN. Sensors 10(4):3934–3953
Chikh MA, Saidi M, Settouti N (2012) Diagnosis of diabetes diseases using an artificial immune recognition system2 (AIRS2) with fuzzy k-nearest neighbor. J Med Syst 36(5):2721–2729
Aslam MW, Zhu Z, Nandi AK (2013) Feature generation using genetic programming with comparative partner selection for diabetes classification. Expert Syst Appl 40(13):5402–5412
Christobel YA, Sivaprakasam P (2013) A new classwise k nearest neighbor (CKNN) method for the classification of diabetes dataset. Int J Eng Adv Technol 2(3):396–200
NirmalaDevi M, Alias Balamurugan SA, Swathi UV (2013) An amalgam KNN to predict diabetes mellitus. In 2013 IEEE International Conference ON Emerging Trends in Computing, Communication and Nanotechnology (ICECCN). IEEE. pp 691-695
Sarwar A, Sharma V (2014) Comparative analysis of machine learning techniques in prognosis of type II diabetes. AI & Soc 29(1):123–129
Farahmandian M, Lotfi Y, Maleki I (2015) Data mining algorithms application in diabetes diseases diagnosis: A case study. Magnt Res Tech Rep 3(1):989–997
Hidalgo JI, Colmenar JM, Kronberger G, Winkler SM, Garnica O, Lanchares J (2017) Data based prediction of blood glucose concentrations using evolutionary methods. J Med Syst 41(9):1–20
Kumar PS, Pranavi S (2017) Performance analysis of machine learning algorithms on diabetes dataset using big data analytics. In 2017 International Conference on Infocom Technologies and Unmanned Systems (Trends and Future Directions)(ICTUS). IEEE. pp 508-513
Aiello EM, Toffanin C, Messori M, Cobelli C, Magni L (2018) Postprandial glucose regulation via KNN meal classification in type 1 diabetes. IEEE Control Syst Lett 3(2):230–235
Mittal K, Aggarwal G, Mahajan P (2019) Performance study of K-nearest neighbor classifier and K-means clustering for predicting the diagnostic accuracy. Int J Inf Technol 11(3):535–540
Dey SK, Hossain A, Rahman MM (2018) Implementation of a web application to predict diabetes disease: an approach using machine learning algorithm. In 2018 21st international conference of computer and information technology (ICCIT). IEEE. pp 1-5
Azrar A, Ali Y, Awais M, Zaheer K (2018) Data mining models comparison for diabetes prediction. Int J Adv Comput Sci Appl 9(8):320–323
Alehegn M, Joshi RR, Mulay P (2019) Diabetes analysis and prediction using random forest KNN Naïve Bayes and J48: An ensemble approach. Int J Sci Technol Res 8(9):1346–1354
Aminah R, Saputro AH (2019) Diabetes prediction system based on iridology using machine learning. In 2019 6th International Conference on Information Technology, Computer and Electrical Engineering (ICITACEE). IEEE. pp 1-6
Faruque MF, Sarker IH (2019) Performance analysis of machine learning techniques to predict diabetes mellitus. In 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE). IEEE. pp 1-4
Dahiwade D, Patle G, Meshram E (2019) Designing disease prediction model using machine learning approach. In 2019 3rd International Conference on Computing Methodologies and Communication (ICCMC). IEEE. pp 1211-1215
El-Sappagh S, Elmogy M, Ali F, Abuhmed T, Islam SM, Kwak KS (2019) A comprehensive medical decision–support framework based on a heterogeneous ensemble classifier for diabetes prediction. Electronics 8(6):635
Ali AMEER, Alrubei MA, Hassan LFM, Al-Ja’afari MA, Abdulwahed SH (2020) Diabetes classification based on KNN. IIUM Eng J 21(1):175–181
Garcia-Carretero R, Vigil-Medina L, Mora-Jimenez I, Soguero-Ruiz C, Barquero-Perez O, Ramos-Lopez J (2020) Use of a K-nearest neighbors model to predict the development of type 2 diabetes within 2 years in an obese, hypertensive population. Med Biol Eng Comput 58(5):991–1002
Gupta SC, Goel N (2020) Performance enhancement of diabetes prediction by finding optimum K for KNN classifier with feature selection method. In 2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT). IEEE. pp 980-986
Hassan AS, Malaserene I, Leema AA (2020) Diabetes Mellitus Prediction using Classification Techniques. Int J Innov Technol Explor Eng 9(5):2080–2084
Sarker IH, Faruque F, Alqahtani H, Kalim A (2018) K-nearest neighbor learning based diabetes mellitus prediction and analysis for eHealth services. EAI Endorsed Trans Scalable Inf Syst 7(26):e4–e4
Bhardwaj C, Jain S, Sood M (2021) Hierarchical severity grade classification of non-proliferative diabetic retinopathy. J Ambient Intell Humaniz Comput 12(2):2649–2670
Mohanty S, Mishra A, Saxena A (2021) Medical Data Analysis Using Machine Learning with KNN. In International Conference on Innovative Computing and Communications. Springer, Singapore. pp 473-485
Patra R (2021) Analysis and Prediction Of Pima Indian Diabetes Dataset Using SDKNN Classifier Technique. In IOP Conference Series: Materials Science and Engineering. vol. 1070, no. 1. IOP Publishing. p 012059
Shinde VD, Raut JR, Sharma Y (2021) Performance evaluation of various supervised machine learning algorithms for diabetes prediction. Eur J Mol Clin Med 7(8):4921–4925
Sopharak A, Dailey MN, Uyyanonvara B, Barman S, Williamson T, Nwe KT, Moe YA (2010) Machine learning approach to automatic exudate detection in retinal images from diabetic patients. J Mod Opt 57(2):124–135
Tama BA (2011) An early detection method of type-2 diabetes mellitus in public hospital. Telkomnika 9(2):287–294
Guo Y, Bai G, Hu Y (2012) Using bayes network for prediction of type-2 diabetes. In 2012 International Conference for Internet Technology and Secured Transactions. IEEE. pp 471-472
Leung RK, Wang Y, Ma RC, Luk AO, Lam V, Ng M, … Chan JC (2013) Using a multi-staged strategy based on machine learning and mathematical modeling to predict genotype-phenotype risk patterns in diabetic kidney disease: a prospective case–control cohort analysis. BMC Nephrol 14(1):1–9
Lee BJ, Ku B, Nam J, Pham DD, Kim JY (2013) Prediction of fasting plasma glucose status using anthropometric measures for diagnosing type 2 diabetes. IEEE J Biomed Health Inform 18(2):555–561
Huang GM, Huang KY, Lee TY, Weng JTY (2015) An interpretable rule-based diagnostic classification of diabetic nephropathy among type 2 diabetes patients. BMC Bioinform 16(1):1–10
Singh DAAG, Leavline EJ, Baig BS (2017) Diabetes prediction using medical data. J Comput Intell Bioinforma 10(1):1–8
Das H, Naik B, Behera HS (2018) Classification of diabetes mellitus disease (DMD): a data mining (DM) approach. In Progress in computing, analytics and networking. Springer, Singapore. pp 539-549
Insani MI, Alamsyah A, Putra AT (2018) Implementation of Expert System for Diabetes Diseases using Naïve Bayes and Certainty Factor Methods. Sci J Inform 5(2):185–193
Uddin S, Khan A, Hossain ME, Moni MA (2019) Comparing different supervised machine learning algorithms for disease prediction. BMC Med Inform Decis Mak 19(1):1–16
Birjais R, Mourya AK, Chauhan R, Kaur H (2019) Prediction and diagnosis of future diabetes risk: a machine learning approach. SN Appl Sci 1(9):1–8
Khan NS, Muaz MH, Kabir A, Islam MN (2019) A Machine Learning-Based Intelligent System for Predicting Diabetes. Int J Big Data Analytics Healthcare 4(2):1–20
Sonar P, JayaMalini K (2019) Diabetes prediction using different machine learning approaches. In 2019 3rd International Conference on Computing Methodologies and Communication (ICCMC). IEEE. pp 367-371
Nakra A, Duhan M (2019) Comparative Analysis of Bayes Net Classifier, Naive Bayes Classifier and Combination of both Classifiers using WEKA. IJ Inf Technol Comput Sci 11:38–45
Jackins V, Vimal S, Kaliappan M, Lee MY (2021) AI-based smart prediction of clinical disease using random forest classifier and Naive Bayes. J Supercomput 77(5):5198–5219
Priya KL, Kypa MSCR, Reddy MMS, Reddy GRM (2020) A Novel Approach to Predict Diabetes by Using Naive Bayes Classifier. In 2020 4th International Conference on Trends in Electronics and Informatics (ICOEI)(48184). IEEE. pp 603-607
Rghioui A, Lloret J, Harane M, Oumnad A (2020) A Smart Glucose Monitoring System for Diabetic Patient. Electronics 9(4):678
Khalilia M, Chakraborty S, Popescu M (2011) Predicting disease risks from highly imbalanced data using random forest. BMC Med Inform Decis Mak 11(1):1–13
Casanova R, Saldana S, Chew EY, Danis RP, Greven CM, Ambrosius WT (2014) Application of random forests methods to diabetic retinopathy classification analyses. PLoS One 9(6):e98587
Sabariah MMK, Hanifa SA and Sa'adah MS (2014) Early detection of type II Diabetes Mellitus with random forest and classification and regression tree (CART). In 2014 International Conference of Advanced Informatics: Concept, Theory and Application (ICAICTA). IEEE 238-242
Butwall M, Kumar S (2015) A data mining approach for the diagnosis of diabetes mellitus using random forest classifier. Int J Comput Appl 120(8)
Xu W, Zhang J, Zhang Q, Wei X (2017) Risk prediction of type II diabetes based on random forest model. In 2017 Third International Conference on Advances in Electrical, Electronics, Information, Communication and Bio-Informatics (AEEICB). IEEE. pp 382-386
Kumar NK, Vigneswari D, Krishna MV, Reddy GP (2019) An optimized random forest classifier for diabetes mellitus. In Emerging Technologies in Data Mining and Information Security. Springer, Singapore. pp 765-773
VijiyaKumar K, Lavanya B, Nirmala I, Caroline SS (2019) Random Forest Algorithm for the Prediction of Diabetes. In 2019 IEEE International Conference on System, Computation, Automation and Networking (ICSCAN). IEEE. pp 1-5
Kaur M, Gianey HK, Singh D, Sabharwal M (2019) Multi-objective differential evolution based random forest for e-health applications. Modern Phys Lett B 33(05):1950022
Alam MZ, Rahman MS, Rahman MS (2019) A Random Forest based predictor for medical data classification using feature ranking. Inform Med Unlocked 15:100180
Kaur P, Kumar R, Kumar M (2019) A healthcare monitoring system using random forest and internet of things (IoT). Multimed Tools Appl 78(14):19905–19916
Benbelkacem S, Atmani B (2019) Random forests for diabetes diagnosis. In 2019 International Conference on Computer and Information Sciences (ICCIS). IEEE. pp 1-4
Wang J, Shi L (2020) Prediction of medical expenditures of diagnosed diabetics and the assessment of its related factors using a random forest model, MEPS 2000–2015. Int J Qual Health Care 32(2):99–112
Wang X, Zhai M, Ren Z, Ren H, Li M, Quan D, … Qiu L (2021) Exploratory study on classification of diabetes mellitus through a combined Random Forest Classifier. BMC Med Inform Decis Mak 21(1):1–14
Ooka T, Johno H, Nakamoto K, Yoda Y, Yokomichi H, Yamagata Z (2021) Random forest approach for determining risk prediction and predictive factors of type 2 diabetes: large-scale health check-up data in Japan. BMJ Nutrition, Prevention & Health 4(1):140
Padmaja P, Vikkurty S, Siddiqui NI, Dasari P, Ambica B, Rao VV, … Rudraraju VR (2008) Characteristic evaluation of diabetes data using clustering techniques. IJCSNS 8(11):244
Khanna S, Agarwal S (2013) An Integrated Approach towards the prediction of Likelihood of Diabetes. In 2013 International Conference on Machine Intelligence and Research Advancement. IEEE. pp 294-298
Paul R, Hoque ASML (2010) Clustering medical data to predict the likelihood of diseases. In 2010 fifth international conference on digital information management (ICDIM). IEEE. pp 44-49
Al Hazemi F, Youn CH, Al-Rubeaan KA (2011) Grid-based interactive diabetes system. In 2011 IEEE First International Conference on Healthcare Informatics, Imaging and Systems Biology. IEEE. pp 258-263
Antonelli D, Baralis E, Bruno G, Cerquitelli T, Chiusano S, Mahoto N (2013) Analysis of diabetic patients through their examination history. Expert Syst Appl 40(11):4672–4678
Al-Hazemi F (2014) Grid-based Workflow System for Chronic Disease Study. Life Sci J 11(7):1–3
Jeong S, Youn CH, Kim YW, Shim SO (2014) Temporal progress model of metabolic syndrome for clinical decision support system. IRBM 35(6):310–320
Kim E, Oh W, Pieczkiewicz DS, Castro MR, Caraballo PJ, Simon GJ (2014) Divisive hierarchical clustering towards identifying clinically significant pre-diabetes subpopulations. In AMIA Annual Symposium Proceedings, vol. 2014. American Medical Informatics Association. p 1815
Vijayarani DS, Jothi MP (2014) Hierarchical and partitioning clustering algorithms for detecting outliers in data streams. International Journal of Advanced Research in Computer and Communication Engineering, ISSN, pp 2278–1021
Sanakal R, Jayakumari T (2014) Prognosis of diabetes using data mining approach-fuzzy C means clustering and support vector machine. Int J Comput Trends Technol 11(2):94–98
Flynt A, Daepp MI (2015) Diet-related chronic disease in the northeastern United States: a model-based clustering approach. Int J Health Geogr 14(1):1–14
Barale MS, Shirke DT (2016) Cascaded modeling for PIMA Indian diabetes data. Int J Comput Appl 139(11):1–4
Bhatia K, Syal R (2017) Predictive analysis using hybrid clustering in diabetes diagnosis. In 2017 Recent Developments in Control, Automation & Power Engineering (RDCAPE). IEEE. pp 447-452
Cheruku R, Edla DR, Kuppili V (2017) Diabetes classification using radial basis function network by combining cluster validity index and bat optimization with novel fitness function. Int J Comput Intell Syst 10(1):247–265
Ahlqvist E, Storm P, Käräjämäki A, Martinell M, Dorkhan M, Carlsson A, … Groop L (2018) Novel subgroups of adult-onset diabetes and their association with outcomes: a data-driven cluster analysis of six variables. Lancet Diabetes Endocrinol 6(5):361–369
Rani S, Kautish S (2018) Association Clustering and Time Series Based Data Mining in Continuous Data for Diabetes Prediction. In 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS). IEEE. pp 1209-1214
Derevitskii IV, Kovalchuk SV (2019) Analysis course of the disease of type 2 diabetes patients using Markov chains and clustering methods. Procedia Comput Sci 156:114–122
Lasek P, Mei Z (2019) Clustering and visualization of a high-dimensional diabetes dataset. Procedia Comput Sci 159:2179–2188
Raihan M, Islam MT, Farzana F, Raju MGM, Mondal HS (2019) An Empirical Study to Predict Diabetes Mellitus using K-Means and Hierarchical Clustering Techniques. In 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT). IEEE. pp 1-6
Nguyen HT, Phan NYK, Luong HH, Cao NH, Huynh HX (2020) Binning approach based on classical clustering for type 2 diabetes diagnosis. Int J Adv Comput Sci Appl 11(3)
Syafaah L, Azizah DF, Sofiani IR, Lestandy M, Faruq A (2020) Self-Monitoring and Detection of Diabetes with art Toilet based on Image Processing and K-Means Technique. In 2020 IEEE International Conference on Automatic Control and Intelligent Systems (I2CACIS). IEEE. pp 87-91
Anwar S, Alqarni A, Alafnan A, Alamri A, Mathew S, Ricciardi E, Mathew S, Alamri A, Alafnan A, Alqarni A, Anwar S (2021) Cluster identification of diabetic risk factors among Saudi population. J Pharma Res Int 3(8)45–58
Takahashi K, Uchiyama H, Yanagisawa S, Kamae I (2006) The logistic regression and ROC analysis of group-based screening for predicting diabetes incidence in four years. Kobe J Med Sci 52(6):171
Sparacino G, Zanderigo F, Corazza S, Maran A, Facchinetti A, Cobelli C (2007) Glucose concentration can be predicted ahead in time from continuous glucose monitoring sensor time-series. IEEE Trans Biomed Eng 54(5):931–937
Eren-Oruklu M, Cinar A, Quinn L, Smith D (2009) Estimation of future glucose concentrations with subject-specific recursive linear models. Diabetes Technol Ther 11(4):243–253
Eren-Oruklu M, Cinar A, Quinn L, Smith D (2009) Adaptive control strategy for regulation of blood glucose levels in patients with type 1 diabetes. J Process Control 19(8):1333–1346
Gani A, Gribok AV, Rajaraman S, Ward WK, Reifman J (2008) Predicting subcutaneous glucose concentration in humans: data-driven glucose modeling. IEEE Trans Biomed Eng 56(2):246–254
Estrada GC, Kirchsteiger H, del Re L, Renard E (2010) Innovative approach for online prediction of blood glucose profile in type 1 diabetes patients. In Proceedings of the 2010 American Control Conference. IEEE. pp 2015-2020
Gani A, Gribok AV, Lu Y, Ward WK, Vigersky RA, Reifman J (2009) Universal glucose models for predicting subcutaneous glucose concentration in humans. IEEE Trans Inf Technol Biomed 14(1):157–165
Lu Y, Rajaraman S, Ward WK, Vigersky RA, Reifman J (2011) Predicting human subcutaneous glucose concentration in real time: a universal data-driven approach. In 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE. pp 7945-7948
Zhao C, Dassau E, Zisser HC, Jovanovič L, Doyle FJ III, Seborg DE (2014) Online prediction of subcutaneous glucose concentration for type 1 diabetes using empirical models and frequency-band separation. AICHE J 60(2):574–584
Georga EI, Protopappas VC, Ardigo D, Marina M, Zavaroni I, Polyzos D, Fotiadis DI (2012) Multivariate prediction of subcutaneous glucose concentration in type 1 diabetes patients based on support vector regression. IEEE J Biomed Health Inform 17(1):71–81
Bayrak ES, Turksoy K, Cinar A, Quinn L, Littlejohn E, Rollins D (2013) Hypoglycemia early alarm systems based on recursive autoregressive partial least squares models. J Diabetes Sci Technol 7(1):206–214
Yu C, Zhao C (2014) Rapid model identification for online glucose prediction of new subjects with type 1 diabetes using model migration method. IFAC Proc Volumes 47(3):2094–2099
Zhao C, Yu C (2015) Rapid model identification for online subcutaneous glucose concentration prediction for new subjects with type I diabetes. IEEE Trans Biomed Eng 62(5):1333–1344
Paul SK, Samanta M (2015) Predicting upcoming glucose levels in patients with type 1 diabetes using a generalized autoregressive conditional heteroscedasticity modelling approach. Int J Stat Med Res 4(2):188–198
Bagherzadeh-Khiabani F, Ramezankhani A, Azizi F, Hadaegh F, Steyerberg EW, Khalili D (2016) A tutorial on variable selection for clinical prediction models: feature selection methods in data mining could improve the results. J Clin Epidemiol 71:76–85
Agarwal V, Podchiyska T, Banda JM, Goel V, Leung TI, Minty EP, … Shah NH (2016) Learning statistical models of phenotypes using noisy labeled training data. J Am Med Inform Assoc 23(6):1166–1173
Lee BJ, Kim JY (2016) Identification of type 2 diabetes risk factors using phenotypes consisting of anthropometry and triglycerides based on machine learning. IEEE J Biomed Health Inform 20(1):39–46
Rahimloo P, Jafarian A (2016) Prediction of diabetes by using artificial neural network, logistic regression statistical model and combination of them. Bull Soc R Sci Liège 85:1148–1164
Rau HH, Hsu CY, Lin YA, Atique S, Fuad A, Wei LM, Hsu MH (2016) Development of a web-based liver cancer prediction model for type II diabetes patients by using an artificial neural network. Comput Methods Prog Biomed 125:58–65
Usman S, Reaz MBI, Ali MAM (2016) Risk prediction of having increased arterial stiffness among diabetic patients using logistic regression. In 2016 IEEE EMBS Conference on Biomedical Engineering and Sciences (IECBES). IEEE. pp 699-701
Zhao LP, Bolouri H, Zhao M, Geraghty DE, Lernmark Å, Better Diabetes Diagnosis Study Group (2016) An object-oriented regression for building disease predictive models with multiallelic HLA genes. Genet Epidemiol 40(4):315–332
Bajestani NS, Kamyad AV, Esfahani EN, Zare A (2018) Prediction of retinopathy in diabetic patients using type-2 fuzzy regression model. Eur J Oper Res 264(3):859–869
Hassan M, Butt MA, Baba MZ (2017) Logistic regression versus neural networks: the best accuracy in prediction of diabetes disease. Asi J Comp Sci Tech 6:33–42
Zheng T, Xie W, Xu L, He X, Zhang Y, You M, … Chen Y (2017) A machine learning-based framework to identify type 2 diabetes through electronic health records. Int J Med Inform 97:120–127
Wu H, Yang S, Huang Z, He J, Wang X (2018) Type 2 diabetes mellitus prediction model based on data mining. Inform Med Unlocked 10:100–107
Qiu S, Li J, Chen B, Wang P, Gao X (2019) An improved prediction method for diabetes based on a feature-based least angle regression algorithm. In Proceedings of the 3rd International Conference on Machine Learning and Soft Computing. pp 232-238
Yao L, Zhong Y, Wu J, Zhang G, Chen L, Guan P, … Liu L (2019) Multivariable logistic regression and back propagation artificial neural network to predict diabetic retinopathy. Diabetes Metab Syndr Obes 12:1943
Alshamlan H, Taleb HB, Al Sahow A (2020) A Gene Prediction Function for Type 2 Diabetes Mellitus using Logistic Regression. In 2020 11th International Conference on Information and Communication Systems (ICICS). IEEE. pp 1-4
Kopitar L, Kocbek P, Cilar L, Sheikh A, Stiglic G (2020) Early detection of type 2 diabetes mellitus using machine learning-based prediction models. Sci Rep 10(1):1–12
Hsu W, Lee ML, Liu B, Ling TW (2000) Exploration mining in diabetic patients databases: findings and conclusions. In Proceedings of the sixth ACM SIGKDD international conference on Knowledge discovery and data mining. pp 430-436
Stilou S, Bamidis PD, Maglaveras N, Pappas C (2001) Mining association rules from clinical databases: an intelligent diagnostic process in healthcare. Stud Health Technol Inform 2:1399–1403
Zorman M, Masuda G, Kokol P, Yamamoto R, Stiglic B (2002) Mining diabetes database with decision trees and association rules. In Proceedings of 15th IEEE Symposium on Computer-Based Medical Systems (CBMS 2002). IEEE. pp 134-139
Duru N (2005) An application of apriori algorithm on a diabetic database. In International Conference on Knowledge-Based and Intelligent Information and Engineering Systems (pp. 398-404). Springer, Berlin, Heidelberg
Mao W and Mao J (2009) The application of apriori-gen algorithm in the association study in type 2 diabetes. In 2009 3rd International Conference on Bioinformatics and Biomedical Engineering. IEEE 1-4
Patil B, Joshi R, Toshniwal D (2010) Association rule for classification of type -2 diabetic patients. In 2010 Second International Conference on Machine Learning and Computing. IEEE 330-334
Patil BM, Joshi RC, Toshniwal D (2011) Classification of type-2 diabetic patients by using Apriori and predictive Apriori. Int J Comput Vis Robotics 2(3):254–265
Kasemthaweesab P and Kurutach W (2012) Association analysis of diabetes mellitus (DM) with complication states based on association rules. In 2012 7th IEEE Conference on Industrial Electronics and Applications (ICIEA). IEEE 1453-1457
Kim HS, Shin AM, Kim MK, Kim YN (2012) Comorbidity study on type 2 diabetes mellitus using data mining. Korean J Int Med 27(2):197
Simon GJ, Schrom J, Castro MR, Li PW and Caraballo P J (2013) Survival association rule mining towards type 2 diabetes risk assessment. In AMIA annual symposium proceedings. Am Med Inform Assoc 2013:1293
Schrom JR, Caraballo PJ, Castro MR and Simon GJ (2013) Quantifying the effect of statin use in pre-diabetic phenotypes discovered through association rule mining. In AMIA Annual Symposium Proceedings. Am Med Inform Assoc 2013:1249
Lakshmi KS and Kumar GS (2014) Association rule extraction from medical transcripts of diabetic patients. In The Fifth International Conference on the Applications of Digital Information and Web Technologies (ICADIWT 2014). IEEE 201-206
Karthikeyan T, Vembandasamy K (2015) A novel algorithm to diagnosis type II diabetes mellitus based on association rule mining using MPSO-LSSVM with outlier detection method. Indian J Sci Technol 8(S8):310–320
Ramezankhani A, Pournik O, Shahrabi J, Azizi F, Hadaegh F (2015) An application of association rule mining to extract risk pattern for type 2 diabetes using tehran lipid and glucose study database. Int J Endocrinol Metabol 13(2)
Simon GJ, Caraballo PJ, Therneau TM, Cha SS, Castro MR, Li PW (2013) Extending association rule summarization techniques to assess risk of diabetes mellitus. IEEE Trans Knowl Data Eng 27(1):130–141
Kamalesh MD, Prasanna KH, Bharathi B, Dhanalakshmi R and Canessane RA (2016) Predicting the risk of diabetes mellitus to subpopulations using association rule mining. In proceedings of the international conference on soft computing systems (pp. 59-65). Springer, New Delhi
Alam TM, Iqbal MA, Ali Y, Wahab A, Ijaz S, Baig TI, … Abbas Z (2019) A model for early prediction of diabetes. Inform Med Unlocked 16:100204
Lu PH, Keng JL, Tsai FM, Lu PH, Kuo CY (2021) An apriori algorithm-based association rule analysis to identify acupoint combinations for treating diabetic gastroparesis. Evid-Based Complement Altern Med 2021(1):6649331
Cheng Y, Wang F, Zhang P and Hu J (2016). Risk prediction with electronic health records: A deep learning approach. In Proceedings of the 2016 SIAM International Conference on Data Mining. Soc Industrial App Mathematics 432-440
Pratt H, Coenen F, Broadbent DM, Harding SP, Zheng Y (2016) Convolutional neural networks for diabetic retinopathy. Procedia Comput Sci 90:200–205
Shi X, Hu Y, Zhang Y, Li W, Hao Y, Alelaiwi A, … Hossain MS (2016) Multiple disease risk assessment with uniform model based on medical clinical notes. IEEE Access 4:7074–7083
Zhu Z, Yin C, Qian B, Cheng Y, Wei J and Wang F (2016) Measuring patient similarities via a deep architecture with medical concept embedding. In 2016 IEEE 16th International Conference on Data Mining (ICDM). IEEE 749-758
Lekha S, Suchetha M (2017) Real-time non-invasive detection and classification of diabetes using modified convolution neural network. IEEE J Biomed Health Inform 22(5):1630–1636
Mohebbi A, Aradóttir TB, Johansen AR, Bengtsson H, Fraccaro M and Mørup M (2017) A deep learning approach to adherence detection for type 2 diabetics. In 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE 2896-2899
Kwasigroch A, Jarzembinski B and Grochowski M (2018) Deep CNN based decision support system for detection and assessing the stage of diabetic retinopathy. In 2018 International Interdisciplinary PhD Workshop (IIPhDW). IEEE 111-116
Swapna G, Kp S, Vinayakumar R (2018) Automated detection of diabetes using CNN and CNN-LSTM network and heart rate signals. Procedia Comput Sci 132:1253–1262
Swapna G, Vinayakumar R, Soman KP (2018) Diabetes detection using deep learning algorithms. ICT Express 4(4):243–246
Butt MM, Latif G, Iskandar DA, Alghazo J, Khan AH (2019) Multi-channel Convolutions Neural Network Based Diabetic Retinopathy Detection from Fundus Images. Procedia Comput Sci 163:283–291
Khan SH, Abbas Z and Rizvi SD (2019) Classification of diabetic retinopathy images based on customised CNN architecture. In 2019 Amity International Conference on Artificial Intelligence (AICAI). IEEE 244-248
Sun Y (2019) The neural network of one-dimensional convolution-an example of the diagnosis of diabetic retinopathy. IEEE Access 7:69657–69666
Raj MAH, Al Mamun M and Faruk MF (2020) CNN Based Diabetic Retinopathy Status Prediction Using Fundus Images. In 2020 IEEE Region 10 Symposium (TENSYMP). IEEE 190-193
Rahman M, Islam D, Mukti RJ, Saha I (2020) A deep learning approach based on convolutional LSTM for detecting diabetes. Comput Biol Chem 88:107329
Ismail WN, Hassan MM, Alsalamah HA, Fortino G (2020) CNN-based health model for regular health factors analysis in Internet-of-medical things environment. IEEE Access 8:52541–52549
Islam MT, Al-Absi HR, Ruagh EA, Alam T (2021) DiaNet: A deep learning based architecture to diagnose diabetes using retinal images only. IEEE Access 9:15686–15695
Allam F, Nossai Z, Gomma H, Ibrahim I, Abdelsalam M (2011) A recurrent neural network approach for predicting glucose concentration in type-1 diabetic patients, In Engineering Applications of Neural Networks (pp. 254-259). Springer, Berlin, Heidelberg
Chu J, Dong W, He K, Duan H, Huang Z (2018) Using neural attention networks to detect adverse medical events from electronic health records. J Biomed Inform 87:118–130
Wang WW, Li H, Cui L, Hong X and Yan Z (2018) Predicting clinical visits using recurrent neural networks and demographic information. In 2018 IEEE 22nd International Conference on Computer Supported Cooperative Work in Design ((CSCWD)). IEEE 353-358
Wu S, Liu S, Sohn S, Moon S, Wi CI, Juhn Y, Liu H (2018) Modeling asynchronous event sequences with RNNs. J Biomed Inform 83:167–177
Dong Y, Wen R, Zhang K and Zhang L (2019) A Novel RNN-Based Blood Glucose Prediction Approach Using Population and Individual Characteristics. In 2019 IEEE 7th International Conference on Bioinformatics and Computational Biology (ICBCB). IEEE 145-149
Dong Y, Wen R, Li Z, Zhang K and Zhang L (2019) Clu-RNN: a new RNN based approach to diabetic blood glucose prediction. In 2019 IEEE 7th International Conference on Bioinformatics and Computational Biology (ICBCB). IEEE 50-55
Jang JS, Lee MJ, Lee TR (2019) Development of T2DM Prediction Model Using RNN. J Digit Converg 17(8):249–255
Munoz-Organero M (2020) Deep Physiological Model for Blood Glucose Prediction in T1DM Patients. Sensors 20(14):3896
Zhou H, Myrzashova R, Zheng R (2020) Diabetes prediction model based on an enhanced deep neural network. EURASIP J Wirel Commun Netw 2020(1):1–13
Zhu T, Li K, Chen J, Herrero P, Georgiou P (2020) Dilated recurrent neural networks for glucose forecasting in type 1 diabetes. J Healthc Inform Res 4(3):308–324
Rabby MF, Tu Y, Hossen MI, Lee I, Maida AS, Hei X (2021) Stacked LSTM based deep recurrent neural network with kalman smoothing for blood glucose prediction. BMC Med Inform Decis Mak 21(1):1–15
Martinsson J, Schliep A, Eliasson B, Meijner C, Persson S and Mogren O (2018) Automatic blood glucose prediction with confidence using recurrent neural networks. In KHD@ IJCAI
Chen J, Li K, Herrero P, Zhu T and Georgiou P (2018) Dilated Recurrent Neural Network for Short-time Prediction of Glucose Concentration. In KHD@ IJCAI (pp. 69-73)
Jaafar SFB and Ali DM (2005) Diabetes mellitus forecast using artificial neural network (ANN). In 2005 Asian Conference on Sensors and the International Conference on New Techniques in Pharmaceutical and Biomedical Research. IEEE 135-139
Mougiakakou SG, Prountzou A, Iliopoulou D, Nikita KS, Vazeou A and Bartsocas CS (2006) Neural network based glucose-insulin metabolism models for children with type 1 diabetes. In 2006 International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE 3545-3548
Dey R, Bajpai V, Gandhi G and Dey B (2008) Application of artificial neural network (ANN) technique for diagnosing diabetes mellitus. In 2008 IEEE Region 10 and the Third international Conference on Industrial and Information Systems. IEEE 1-4
Pappada SM, Cameron BD, Rosman PM (2008) Development of a neural network for prediction of glucose concentration in type 1 diabetes patients. J Diabetes Sci Technol 2(5):792–801
Zainuddin Z, Pauline O, Ardil C (2009) A neural network approach in predicting the blood glucose level for diabetic patients. Int J Comput Intell 5(1):72–79
Pérez-Gandía C, Facchinetti A, Sparacino G, Cobelli C, Gómez EJ, Rigla M, … Hernando ME (2010) Artificial neural network algorithm for online glucose prediction from continuous glucose monitoring. Diabetes Technol Ther 12(1):81–88
Pappada SM, Borst MJ, Cameron BD, Bourey RE, Lather JD, Shipp D, … Papadimos TJ (2010) Development of a neural network model for predicting glucose levels in a surgical critical care setting. Patient Safety Surg 4(1):1–5
Chakraborty M and Tudu B (2010) Comparison of ANN models to predict LDL level in Diabetes Mellitus type 2. In 2010 International Conference on Systems in Medicine and Biology. IEEE 392-396
Allam F, Nossair Z, Gomma H, Ibrahim I and Abd-el Salam M (2011) Prediction of subcutaneous glucose concentration for type-1 diabetic patients using a feed forward neural network. In The 2011 International Conference on Computer Engineering and Systems. IEEE 129-133
Pappada SM, Cameron BD, Rosman PM, Bourey RE, Papadimos TJ, Olorunto W, Borst MJ (2011) Neural network-based real-time prediction of glucose in patients with insulin-dependent diabetes. Diabetes Technol Ther 13(2):135–141
Robertson G, Lehmann ED, Sandham W, Hamilton D (2011) Blood glucose prediction using artificial neural networks trained with the AIDA diabetes simulator: a proof-of-concept pilot study. J Electr Comput Eng 2011(1):681786
Ali JB, Hamdi T, Fnaiech N, Di Costanzo V, Fnaiech F, Ginoux JM (2018) Continuous blood glucose level prediction of Type 1 Diabetes based on Artificial Neural Network. Biocybern Biomed Eng 38(4):828–840
Kathiroli R, RajaKumari R and Gokulprasanth P (2018) Diagnosis Of Diabetes Using Cascade Correlation And Artificial Neural Network. In 2018 Tenth International Conference on Advanced Computing (ICoAC). IEEE 299-306
Senturk Z (2020) Artificial Neural Networks based decision support system for the detection of diabetic retinopathy. Sakarya Univ Fen Bilim Enst Derg 24(2):424–431
Sun Q, Jankovic MV, Bally L and Mougiakakou SG (2018) Predicting blood glucose with an LSTM and Bi-LSTM based deep neural network. In 2018 14th Symposium on Neural Networks and Applications (NEUREL). IEEE 1-5
Farías AFS, Mendizabal A, González-Garrido AA, Romo-Vázquez R and Morales A (2018) Long Short-Term Memory Neural Networks for Identifying Type 1 Diabetes Patients with Functional Magnetic Resonance Imaging. In 2018 IEEE Latin American Conference on Computational Intelligence (LA-CCI). IEEE 1-4
Bahadur EH, Masum AKM, Barua A, Alam MGR, Chowdhury MAUZ and Alam MR (2019) LSTM Based Approach for Diabetic Symptomatic Activity Recognition Using Smartphone Sensors. In 2019 22nd International Conference on Computer and Information Technology (ICCIT). IEEE 1-6
De Bois M, El Yacoubi MA and Ammi M (2019) Prediction-coherent LSTM-based recurrent neural network for safer glucose predictions in diabetic people. In International Conference on Neural Information Processing (pp. 510-521). Springer, Cham
De Bois M, El Yacoubi MA and Ammi M (2019) Study of short-term personalized glucose predictive models on type-1 diabetic children. In 2019 International Joint Conference on Neural Networks (IJCNN). IEEE 1-8
Massaro A, Maritati V, Giannone D, Convertini D, Galiano A (2019) LSTM DSS automatism and dataset optimization for diabetes prediction. Appl Sci 9(17):3532
Padmapritha T (2019) Prediction of Blood Glucose Level by using an LSTM based Recurrent Neural networks. In 2019 IEEE International Conference on Clean Energy and Energy Efficient Electronics Circuit for Sustainable Development (INCCES). IEEE 1-4
Carrillo-Moreno J, Pérez-Gandía C, Sendra-Arranz R, García-Sáez G, Hernando ME, Gutiérrez A (2020) Long short-term memory neural network for glucose prediction. Neural Comput Applic 33:4191–4203
Amalia R, Bustamam A, Sarwinda D (2021) Detection and description generation of diabetic retinopathy using convolutional neural network and long short-term memory. J Phys Conf Ser 1722(1):012010
El Idrissi T and Idri A (2020) Deep Learning for Blood Glucose Prediction: CNN vs LSTM. In International Conference on Computational Science and Its Applications (pp. 379-393). Springer, Cham
El Idriss T, Idri A, Abnane I and Bakkoury Z (2019) Predicting blood glucose using an LSTM neural network. In 2019 Federated Conference on Computer Science and Information Systems (FedCSIS). IEEE 35-41
Wang W, Tong M, Yu M (2020) Blood Glucose Prediction With VMD and LSTM Optimized by Improved Particle Swarm Optimization. IEEE Access 8:217908–217916
Beaulieu-Jones BK, Moore JH and POOLED RESOURCE OPEN-ACCESS ALS CLINICAL TRIALS CONSORTIUM (2017) Missing data imputation in the electronic health record using deeply learned autoencoders. In Pacific Symposium on Biocomputing 2017 (pp. 207-218)
Hwang U, Choi S, Lee HB and Yoon S (2018) Adversarial training for disease prediction from electronic health records with missing data. arXiv preprint arXiv:1711.04126
Babu SB, Suneetha A, Babu GC, Kumar YJN, Karuna G (2018) Medical disease prediction using grey wolf optimization and auto encoder based recurrent neural network. Period Eng Nat Sci 6(1):229–240
Kannadasan K, Edla DR, Kuppili V (2019) Type 2 diabetes data classification using stacked autoencoders in deep neural networks. Clin Epidemiol Glob Health 7(4):530–535
Kumar VB, Vijayalakshmi K and Padmavathamma M (2019) A hybrid data mining approach for diabetes prediction and classification. In 2019 World Congress on Engineering and Computer Science, WCECS (Vol. 22, pp. 298-303)
Sahoo AK, Pradhan C and Das H (2020) Performance evaluation of different machine learning methods and deep-learning based convolutional neural network for health decision making. In Nature inspired computing for data science (pp. 201-212). Springer, Cham
Zhang Q, Zhou J and Zhang B (2020) A noninvasive method to detect diabetes mellitus and lung cancer using the stacked sparse autoencoder. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE 1409-1413
García-Ordás MT, Benavides C, Benítez-Andrades JA, Alaiz-Moretón H, García-Rodríguez I (2021) Diabetes detection using deep learning techniques with oversampling and feature augmentation. Comput Methods Prog Biomed 202:105968
Kayaer K and Yildirim T (2003) Medical diagnosis on Pima Indian diabetes using general regression neural networks. In Proceedings of the international conference on artificial neural networks and neural information processing (ICANN/ICONIP) (Vol. 181, p. 184).
Ergün U, Barýþçý N, Ozan AT, Serhatlýoðlu S, Oğur E, Hardalaç F, Güler İ (2004) Classification of MCA stenosis in diabetes by MLP and RBF neural network. J Med Syst 28(5):475–487
Quchani SA, Tahami E (2007) Comparison of MLP and Elman neural network for blood glucose level prediction in type 1 diabetics, In 3rd Kuala Lumpur International Conference on Biomedical Engineering 2006 (pp. 54-58). Springer, Berlin, Heidelberg
Bhatkar AP and Kharat GU (2015) Detection of diabetic retinopathy in retinal images using MLP classifier. In 2015 IEEE international symposium on nanoelectronic and information systems. IEEE 331-335
Ambilwade RP and Manza RR (2016) Prognosis of diabetes using fuzzy inference system and multilayer perceptron. In 2016 2nd International Conference on Contemporary Computing and Informatics (IC3I). IEEE 248-252
Choubey DK, Paul S (2016) GA_MLP NN: a hybrid intelligent system for diabetes disease diagnosis. Int J Intell Syst Appl 8(1):49
Alfian G, Syafrudin M, Ijaz MF, Syaekhoni MA, Fitriyani NL, Rhee J (2018) A personalized healthcare monitoring system for diabetic patients by utilizing BLE-based sensors and real-time data processing. Sensors 18(7):2183
Mohapatra SK, Swain JK, Mohanty MN (2019) Detection of diabetes using multilayer perceptron, In International conference on intelligent computing and applications (pp. 109-116). Springer, Singapore
Bani-Salameh H, Alkhatib SM, Abdalla M, Banat R, Zyod H, Alkhatib AJ (2020) Prediction of diabetes and hypertension using multi-layer perceptron neural networks. Int J Model Simul Sci Comput 12(02):2150012
Güldoğan E, Zeynep TUNÇ, Ayça ACET, Çolak C (2020) Performance Evaluation of Different Artificial Neural Network Models in the Classification of Type 2 Diabetes Mellitus. J Cogn Syst 5(1):23–32
Mishra S, Tripathy HK, Mallick PK, Bhoi AK, Barsocchi P (2020) EAGA-MLP—An Enhanced and Adaptive Hybrid Classification Model for Diabetes Diagnosis. Sensors 20(14):4036
Om KS, Kim HC, Min BG, Shin CS, Lee HK (1998) Statistical RBF Network with Applications to an Expert System for Characterizing Diabetes Mellitus. J Electr Eng Inf Sci 3(3):355–365
Nabney IT (2004) Efficient training of RBF networks for classification. Int J Neural Syst 14(03):201–208
Venkatesan P, Anitha S (2006) Application of a radial basis function neural network for diagnosis of diabetes mellitus. Curr Sci 91(9):1195–1199
Sa’di S, Maleki A, Hashemi R, Panbechi Z, Chalabi K (2015) Comparison of data mining algorithms in the diagnosis of type II diabetes. Int J Comput Sci Appl 5(5):1–12
Ashiquzzaman A, Tushar AK, Islam MR, Shon D, Im K, Park JH, … and Kim J (2018) Reduction of overfitting in diabetes prediction using deep learning neural network. In IT convergence and security 2017 (pp. 35-43). Springer, Singapore
Chetoui M, Akhloufi MA and Kardouchi M (2018) Diabetic retinopathy detection using machine learning and texture features. In 2018 IEEE Canadian Conference on Electrical & Computer Engineering (CCECE). IEEE 1-4
Adegoke V, Chen D, Banissi E (2019) Improving prediction accuracy of breast cancer survivability and diabetes diagnosis via RBF networks trained with EKF models. Int J Comput Inf Syst Ind Manag 11:19–19
Hosseini H, Bardsiri AK (2019) Improving Diagnosis Accuracy of Diabetic Disease Using Radial Basis Function Network and Fuzzy Clustering. Front Health Inform 8(1):24
Kamble VV, Kokate RD (2020) Automated diabetic retinopathy detection using radial basis function. Procedia Comput Sci 167:799–808
Dwivedi AK (2018) Analysis of computational intelligence techniques for diabetes mellitus prediction. Neural Comput Applic 30(12):3837–3845
Thyde DN, Mohebbi A, Bengtsson H, Jensen ML, Mørup M (2021) Machine learning-based adherence detection of type 2 diabetes patients on once-daily basal insulin injections. J Diabetes Sci Technol 15(1):98–108
Pham T, Tran T, Phung D and Venkatesh S (2016) Deepcare: a deep dynamic memory model for predictive medicine. In Pacific-Asia conference on knowledge discovery and data mining (pp. 30-41). Springer, Cham
Choi E, Bahadori MT, Schuetz A, Stewart WF and Sun J (2016) Doctor ai: Predicting clinical events via recurrent neural networks. In Machine learning for healthcare conference. PMLR 301-318
Miotto R, Li L, Kidd BA, Dudley JT (2016) Deep patient: an unsupervised representation to predict the future of patients from the electronic health records. Sci Rep 6(1):1–10
Miotto R, Li L and Dudley JT (2016) Deep learning to predict patient future diseases from the electronic health records. In European Conference on Information Retrieval (pp. 768-774). Springer, Cham
Liang Z, Zhang G, Huang JX and Hu QV (2014) Deep learning for healthcare decision making with EMRs. In 2014 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE 556-559
Lipton ZC, Kale DC, Elkan C and Wetzel R (2015) Learning to diagnose with LSTM recurrent neural networks. arXiv preprint arXiv:1511.03677
Che Z, Kale D, Li W, Bahadori MT and Liu Y (2015) Deep computational phenotyping. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 507-516)
Lasko TA, Denny JC, Levy MA (2013) Computational phenotype discovery using unsupervised feature learning over noisy, sparse, and irregular clinical data. PLoS One 8(6):e66341
Razavian N, Marcus J and Sontag D (2016) Multi-task prediction of disease onsets from longitudinal laboratory tests. In Machine learning for healthcare conference. PMLR 73-100
Choi Y, Chiu CYI, Sontag D (2016) Learning low-dimensional representations of medical concepts. AMIA Summits Transl Sci Proc 2016:41
Tran T, Nguyen TD, Phung D, Venkatesh S (2015) Learning vector representation of medical objects via EMR-driven nonnegative restricted Boltzmann machines (eNRBM). J Biomed Inform 54:96–105
Dernoncourt F, Lee JY, Uzuner O, Szolovits P (2017) De-identification of patient notes with recurrent neural networks. J Am Med Inform Assoc 24(3):596–606
Nguyen P, Tran T, Wickramasinghe N et al (2017) Deepr: a Convolutional Net for Medical Records. IEEE J Biomed Health Inform 21:22–30
Kumar Dewangan A, Agrawal P (2015) Classification of diabetes mellitus using machine learning techniques. Int J Eng Appl Sci 2(5):257905
Deperlioğlu, O, Köse, U. (2018). Diagnosis of Diabetes by Using Deep Neural Network. 2018 2nd International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT).IEEE.
Katsuki T, Ono M, Koseki A, Kudo M, Haida K, Kuroda J, … and Suzuki A (2018) Risk Prediction of Diabetic Nephropathy via Interpretable Feature Extraction from EHR Using Convolutional Autoencoder. In MIE (pp. 106-110)
Makino M, Yoshimoto R, Ono M, Itoko T, Katsuki T, Koseki A, … Suzuki A (2019) Artificial intelligence predicts the progression of diabetic kidney disease using big data machine learning. Sci Rep 9(1):1–9
Deepthi K, Jereesh AS (2020) An ensemble approach for CircRNA-disease association prediction based on autoencoder and deep neural network. Gene 762:145040
Tran D, Nguyen H, Tran B, La Vecchia C, Luu HN, Nguyen T (2021) Fast and precise single-cell data analysis using a hierarchical autoencoder. Nat Commun 12(1):1–10
Li K, Daniels J, Liu C, Herrero P, Georgiou P (2019) Convolutional recurrent neural networks for glucose prediction. IEEE J Biomed Health Inform 24(2):603–613
Sistla S (2022) Predicting Diabetes using SVM Implemented by Machine Learning. International Journal of Soft Computing and Engineering 12(2):2231–2307
Li J, Ding J, Zhi DU, Gu K, Wang H (2022) Identification of type 2 diabetes based on a ten-gene biomarker prediction model constructed using a support vector machine algorithm. BioMed Res Int 2022(1):1230761
Rastogi R, Bansal M (2023) Diabetes prediction model using data mining techniques. Meas Sens 25:100605
Aslan MF, Sabanci K (2023) A novel proposal for deep learning-based diabetes prediction: Converting clinical data to image data. Diagnostics 13(4):796
Ahamed BS, Arya MS, Nancy VAO (2022) Prediction of type-2 diabetes mellitus disease using machine learning classifiers and techniques. Front Comput Sci 4:835242
Özge ŞEN, Keser SB, Keskin K (2023) Early stage diabetes prediction using decision tree-based ensemble learning model. Int Adv Res Eng J 7(1):62–71
Suyanto S, Meliana S, Wahyuningrum T, Khomsah S (2022) A new nearest neighbor-based framework for diabetes detection. Expert Syst Appl 199:116857
Prasad BS, Gupta S, Borah N, Dineshkumar R, Lautre HK, Mouleswararao B (2023) Predicting diabetes with multivariate analysis an innovative KNN-based classifier approach. Prev Med 174:107619
Khanam JJ, Foo SY (2021) A comparison of machine learning algorithms for diabetes prediction. Ict Express 7(4):432–439
Hasan MK, Saeed RA, Alsuhibany SA, Abdel-Khalek S (2022) An empirical model to predict the diabetic positive using stacked ensemble approach. Front Public Health 9:792124
Okikiola FM, Adewale OS, Obe OO (2023) A diabetes prediction classifier model using naive bayes algorithm. Fudma J Sci 7(1):253–260
Mondal S, Banik A, Roy S, Das J, Banerjee S and Navin H (2022) Random Forest Based Diabetic Prediction Model on Highly Unbalanced Dataset. In 2022 IEEE 2nd Mysore Sub Section International Conference (MysuruCon). IEEE 1-6
Gündoğdu S (2023) Efficient prediction of early-stage diabetes using XGBoost classifier with random forest feature selection technique. Multimed Tools Appl 82(22):34163–34181
Hassan MM, Mollick S, Yasmin F (2022) An unsupervised cluster-based feature grouping model for early diabetes detection. Healthc Analyt 2:100112
Alghamdi T (2023) Prediction of Diabetes Complications Using Computational Intelligence Techniques. Appl Sci 13(5):3030
Khafaga DS, Alharbi AH, Mohamed I, Hosny KM (2022) An integrated classification and association rule technique for early-stage diabetes risk prediction. Healthcare 10(10):2070
Madan P, Singh V, Chaudhari V, Albagory Y, Dumka A, Singh R, … AlGhamdi AS (2022) An optimization-based diabetes prediction model using CNN and Bi-directional LSTM in real-time environment. Appl Sci 12(8):3989
Aslan MF, Sabanci K (2023) A novel proposal for deep learning-based diabetes prediction: Converting clinical data to image data. Diagnostics 13(4):796
Srinivasu PN, Shafi J, Krishna TB, Sujatha CN, Praveen SP, Ijaz MF (2022) Using recurrent neural networks for predicting type-2 diabetes from genomic and tabular data. Diagnostics 12(12):3067
Kiruthiga G, Shakkeera L, Asha A, Dhiyanesh B, Saraswathi P, Murali M (2023) Deep Learning-Based Continuous Glucose Monitoring with Diabetic Prediction Using Deep Spectral Recurrent Neural Network. In: International Conference on Information, Communication and Computing Technology. Springer Nature Singapore, Singapore, pp 485–497
Bukhari MM, Alkhamees BF, Hussain S, Gumaei A, Assiri A, Ullah SS (2021) An improved artificial neural network model for effective diabetes prediction. Complexity 2021:1–10
Prakash EP, Srihari K, Karthik S, Kamal MV, Dileep P, Bharath Reddy S, Mukunthan MA, Somasundaram K, Jaikumar R, Gayathri N, Sahile K (2022) Implementation of artificial neural network to predict diabetes with high-quality health system. Comput Intell Neurosci 2022(1):1174173
Al Sadi K, Balachandran W (2023) Prediction Model of Type 2 Diabetes Mellitus for Oman Prediabetes Patients Using Artificial Neural Network and Six Machine Learning Classifiers. Appl Sci 13(4):2344
Alex SA, Jhanjhi NZ, Humayun M, Ibrahim AO, Abulfaraj AW (2022) Deep LSTM Model for Diabetes Prediction with Class Balancing by SMOTE. Electronics 11(17):2737
Prendin F, Pavan J, Cappon G, Del Favero S, Sparacino G, Facchinetti A (2023) The importance of interpreting machine learning models for blood glucose prediction in diabetes: an analysis using SHAP. Sci Rep 13(1):16865
Bani-Salameh H, Alkhatib SM, Abdalla M, Al-Hami MT, Banat R, Zyod H, Alkhatib AJ (2021) Prediction of diabetes and hypertension using multi-layer perceptron neural networks. Int J Model Simul Sci Comput 12(02):2150012
Sivasankari SS, Surendiran J, Yuvaraj N, Ramkumar M, Ravi CN and Vidhya RG (2022)Classification of diabetes using multilayer perceptron. In 2022 IEEE International Conference on Distributed Computing and Electrical Circuits and Electronics (ICDCECE). IEEE 1-5
Ali R, Hussain J, Lee SW (2023) Multilayer perceptron-based self-care early prediction of children with disabilities. Digital Health 9:20552076231184054
Bodapati JD (2022) Stacked convolutional auto-encoder representations with spatial attention for efficient diabetic retinopathy diagnosis. Multimed Tools Appl 81(22):32033–32056
Ismael HA, Al-A’araji NH, Shukur BK (2023) Enhanced the prediction approach of diabetes using an autoencoder with regularization and deep neural network. Period Eng Nat Sci 10(6):156–167
Rashmi K, Rao NK, Bala MM, Lahari M, Fathima N, Prudhvi V (2021) Prediction of diabetes mellitus using rbf neural model and genetic algorithm. Turkish Journal of Physiotherapy and Rehabilitation 32:3
Sivaraman M and Sumitha J (2023) An efficiency of DCKSVM and HRBFNN techniques for diabetic prediction. In AIP Conference Proceedings (Vol. 2831, No. 1). AIP Publishing
Zhang C, Hu C, Wu T, Zhu L, Liu X (2022) Achieving efficient and privacy-preserving neural network training and prediction in cloud environments. IEEE Trans Dependable Secure Comput 20(5):4245–4257
Zhang C, Zhu L, Xu C, Lu R (2018) PPDP: An efficient and privacy-preserving disease prediction scheme in cloud-based e-Healthcare system. Futur Gener Comput Syst 79:16–25
Lei D, Liang J, Zhang C, Liu X, He D, Zhu L, Guo S (2023) Publicly verifiable and secure SVM classification for cloud-based health monitoring services. IEEE Internet Things J
Mathew TE (2019) A comparative study of the performance of different Support Vector machine Kernels in Breast Cancer Diagnosis. Int J Inf Comput Sci 6(6):432–441
Podgorelec V, Kokol P, Stiglic B, Rozman I (2002) Decision trees: an overview and their use in medicine. J Med Syst 26:445–463
Dey L, Chakraborty S, Biswas A, Bose B and Tiwari S (2016) Sentiment analysis of review datasets using naive bayes and k-nn classifier. arXiv preprint arXiv:1610.09982
Khamis HS (2014) Application of k-Nearest Neighbour classification in medical data mining in the context of kenya. In Scientific Conference Proceedings 2022(1):5416722
Breiman L (2001) Random forests. Mach Learn 45:5–32
Cutler DR, Edwards TC Jr, Beard KH, Cutler A, Hess KT, Gibson J, Lawler JJ (2007) Random forests for classification in ecology. Ecology 88(11):2783–2792
Montgomery DC, Peck EA, Vining GG (2021) Introduction to linear regression analysis. John Wiley & Sons
Nathiya G, Punitha SC, Punithavalli M (2010) An analytical study on behavior of clusters using k means, em and k* means algorithm. arXiv preprint arXiv:1004.1743.
Agrawal R, Imieliński T, Swami A (1993) Mining association rules between sets of items in large databases. In Proceedings of the 1993 ACM SIGMOD international conference on Management of data. pp 207-216
Zhu X, Ghahramani Z, Lafferty JD (2003) Semi-supervised learning using gaussian fields and harmonic functions. In Proceedings of the 20th International conference on Machine learning (ICML-03). pp 912-919
Mnih V, Badia AP, Mirza M, Graves A, Lillicrap T, Harley T, … Kavukcuoglu K (2016) Asynchronous methods for deep reinforcement learning. In International conference on machine learning. PMLR. pp 1928-1937
Eiben AE, Smith JE (2015) Introduction to evolutionary computing. Springer-Verlag, Berlin Heidelberg
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. pp 770-778
Madhiarasan M, Louzazni M (2022) Analysis of artificial neural network: architecture, types, and forecasting applications. J Electr Comput Eng 2022(1):5416722
Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, … Farhan L (2021) Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J Big Data 8:1–74
Sherstinsky A (2020) Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys D: Nonlinear Phenom 404:132306
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
There is no conflict of interest in the current research.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Modak, S.K.S., Jha, V.K. Machine and deep learning techniques for the prediction of diabetics: a review. Multimed Tools Appl (2024). https://doi.org/10.1007/s11042-024-19766-9
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11042-024-19766-9