1 Introduction

Nowadays, there is a progressive growth in healthcare applications based on the Internet of Medical Things (IoMT) and several IoT healthcare sensors are integrated to support various healthcare applications [1]. The future reachability and connectivity enhance the IoT thereby generating massive data. The large volumes of data streams during processing are emitted by billions of interconnected IoT devices [2]. The processing services and on-demand scalable storage are offered with cloud computing [3,4,5]. In the healthcare industry, accurate data collection is one of the multiple problems which prevent healthcare providers from offering healthcare services in a real-time context [6].

Recently, the major public health concern is dengue because of its severity in which it harms the central nervous system and the spinal cord. Superior health monitoring systems [2] are critical for the population, regardless of geographic location. Dengue is a viral disease spread by the mosquitos which are infected by the dengue virus and the mosquitos mainly emerge in the tropical and sub-tropical climates worldwide [7]. The disease can attack a man a total of four times based on the four dengue virus serotypes. In most, Asian and Latin American countries, the people affected by dengue severely are prone to death. These people need to be monitored by medical professionals continuously since there exists no specific treatment for dengue. If the dengue-affected people are identified in the early stage, then they can be provided with proper medical access and the fatality rate will be decreased by 1% [8]. A total of 100–400 million dengue infections are identified each year and the disease can be controlled only when there are appropriate precaution measures in place.

Compared to the previous healthcare systems, the major threat and challenging task are to identify the dengue-affected areas and this can be done by deploying the smart sensors developed with the advancements of IoT technology. In real-time applications, the powerful health monitoring system is made with the growth of technology and the advent of pervasive computing. Dengue disease transmission is more prevalent in recent times as per the World Health Organization (WHO) [44]. The main drawbacks associated with dengue disease prediction are the low level of reporting, the complexity associated with disease monitoring, and misclassification errors. The patient health status can be determined via regular monitoring which can be easily achieved via hybridizing different architectures such as IoT, cloud-fog computing, etc. The integrated architectures also offer reliable reporting regarding the patient status and the artificial intelligence technique can be incorporated to yield effective disease diagnosis.

The scarcity of the reverse transcription-polymerase chain reaction (RT-PCR) kits also raises the need for an efficient clinical dengue disease diagnosis procedure. The emergence of different machine learning and deep learning techniques in dengue prediction has helped physicians make quick and meaningful decisions. The different neural network architectures such as handcrafted feature-based neural networks, a cascaded ensemble of Convolutional Neural networks (CNN) [36], Support Vector Machine (SVM) [37], and pre-trained CNN architectures have been used for developing a medical decision support system. These models are capable of predicting different diseases in various organs. These models offer rapid and accurate decisions and can also be applied to other disease predictions.

Because of accurate data analysis, perception of efficiency, large storage capacity, and extreme processing power, the Ternary IoT-Fog-Cloud (TIFC) model gained wide acceptance. The healthcare application is implemented in the TIFC model to monitor the remote site, and improve life quality and emergency decision-making [9, 10]. Several healthcare agencies adopt the TIFC technology to accomplish the targets like minimal delay, output delivery, and smart healthcare monitoring. The current TIFC framework is presented by monitoring new developments in fog-cloud knowledge. In the early stages, the smartphone creates the alert message to the end-user and our proposed TIFC detects the diseases at the initial stage. The major contributions of this study are described as follows:

  • The fog-based model is designed to offer a remote diagnosis of dengue fever concerning the ambient environmental conditions and patient health symptoms.

  • A hybrid CNN-TLSTM with ATLBO algorithm is developed for dengue fever detection and classification which analyzes the disease based on the features extracted from the IoMT sensors. The main aim of this technique is to simplify the dengue disease diagnosis process.

  • The proposed model also aims to monitor the symptoms of dengue-affected patients in remote regions and prevent the disease outbreak by alerting the nearby healthcare centers and people in the locality.

  • The IoMT sensors are initially placed in the regions of infected individuals which offer real-time healthcare provisioning and immediately diagnoses and generate the alert message to the patients.

  • The performance of this model is evaluated in a real-time dataset using different performance metrics such as training time, absolute error, mean absolute percentage error, etc.

The rest of the article is arranged as: Sect. 2 presents the related works concerning the particular domain. The hybrid Convolutional Neural Network-based Tanh Long Short Term Memory (CNN-TLSTM) with Adaptive Teaching Learning Based Optimization (ATLBO) algorithm is formulated in Sect. 3 followed by the proposed work in Sect. 4. The experimental results are discussed in Sect. 5 and the article is concluded in Sect. 6.

2 Related Works

Verma et al. [11] proposed an IoT-aware student-centric stress monitoring (I-SCSM) model to predict the student stress index. At the fog layer, the physiological readings collected from medical sensors with the stress event are classified as normal or abnormal through Bayesian Belief Network (BBN). Based on different stress-oriented parameters at the cloud layer, the time-enriched dataset sequence analyzes abnormal temporal structural data. Further, they formed a multi-level Temporal Dynamic Bayesian Network to calculate the stress index of students. The accuracy and utility of BBN are superior to conducting investigations on both cloud and fog layers. But, it takes larger storage space and unsecured communication.

For a large-scale IoT-based healthcare application, He et al. [12] offered proactive personalized services via fog-cloud computing. In data stream processing, an advantage of complex event processing (CEP) was leveraged. For personalized service to reduce resource waste and accelerate response time, they proposed a hierarchical fog-cloud computing CEP structure. Depending upon this structure, the prototype system called FogCepCare was implemented and it demonstrated higher performance results than other conventional techniques. This model is a large-scale- IoT-based healthcare application with a complex structure.

For healthcare, Gupta et al. [13] suggested the energy-efficient fog-cloud-based structure (EEFCS). Usually, the energy consumption increased because of the huge workload on cloud data centers. This architecture does not include any security protocols due to reduce the system complexity. For hospitals and medical enterprises, the partitioning and scheduling of deep neural network (DNN) based applications via IoT-assisted mobile fog cloud were suggested by Lakhan et al. [14]. This design has better energy consumption and accuracy results. Nevertheless, they only considered the complex mobility features with a dynamic environment and there was a higher delay as well as service cost.

IoT-fog-based cloud architecture for smart systems was introduced by Kallel et al. [15] for COVID‐19 monitoring. The IoT-aware business process modeling was enabled via the extension of the business process model and notation (BPMN). Further, the experimental analysis showed the enhanced performances in case of entire system reliability, computational time, and guaranteed data integrity. But, the overall performance is degraded due to the system production time improvements, reliability, and security. Based on Fog-cloud environments, a one-Dimensional CNN (1D-CNN) approach was established by Cheikhrouhou et al. [16] for ECG arrhythmia analysis. Over the Fog infrastructure, deploy the inference module of the 1D-CNN. While applying the Grid Search algorithm, the MIT-BIH Arrhythmia database with the proposed model demonstrated better performances of F1-measure and accurate results with minimum response time but it met a few limitations such as no latency-sensitive health application, additional delay, and no security model.

Lakhan et al. [38] presented a Deep Neural Networks Energy Cost-Efficient Partitioning and Task Scheduling (DNNECTS) algorithm for cloud-based healthcare partitioning applications. They are mainly focusing on the overhead associated with heavyweight profiling and offloading. The expensive healthcare services are other issues mainly faced by hospitals, and to overcome this issue, different authors used blockchain-based techniques [39, 42], fog networks [40], bio-inspired robotics [43], etc. A critical healthcare management module [41] was developed by integrating different models such as fog and edge computing to improve the healthcare application performance. The ECG signal monitoring, blood pressure control, and heartbeat of the patients in remote locations are analyzed by cloud computing. The edge to the cloud is completely monitored here using the multi-agent system. The main advantage offered by this module is the network usage minimization of 79% and network delay minimization of 65%.

Even though there is different research conducted for dengue disease prediction, the accuracy rate still needs to be enhanced creating room for improvement. This is mainly because when dengue is not p[redicted in time, it may impact the bone marrow which is the main platelet producing center of the body. Hence this paper presents a novel dengue prediction model to offer real-time disease prediction services to the patients in the remote areas.

3 Formulation of Hybrid CNN-TLSTM with ATLBO Algorithm

In this section, we formulate the Hybrid Convolutional Neural Network-based Tanh Long Short Term Memory (CNN-TLSTM) with Adaptive Teaching Learning Optimization (ATLBO) algorithm, which is explained as follows:

3.1 CNN-TLSTM Model

This section constructs the Convolutional Neural Network-based Tanh Long Short Term Memory (CNN-TLSTM) model. Figure 1 describes the CNN-TLSTM architecture. The input and output layers with TLSTM layer and CNN layers are present in this model [17]. The relevant data input for classification is the features from the convolution layer which is reduced by the pooling layer. Next to feature extraction, a time series prediction on the input data is performed by the TLSTM layer thereby computing the final output at the output layer.

Fig. 1
figure 1

General architecture of CNN-LSTM

3.1.1 Convolutional Neural Network (CNN)

The classification and prediction tasks are accurately performed using the CNN model. Further, CNN tackles the issues in feature extraction based on artificial neural networks. The predicted value accuracy is affected based on the feature extracted by the CNN and the core of CNN is the convolutional layer [18]. After the convolutional layer, perform the convolution via the pooling layer. The overfitting, as well as network parameters, are reduced. In the overall CNN, the fully connected layer is considered as the classifier.

3.1.2 Tanh Long Short Term Memory (TLSTM)

The LSTM model structure is improved using the new model called Tanh Long Short Term Memory (TLSTM). From overfitting, the input gate in the LSTM output value is prevented by designing TLSTM. Next to the LSTM input gate, the 1-tanh function is introduced and it retains the significant features of the input data. The sigmoid function activates the LSTM input gate and discards the output values close to zero. The output value close to one is transferred and preserved. The correlation among the time series data is captured by transforming data into several distinct intervals. Figure 2 delineates the architecture of TLSTM.

Fig. 2
figure 2

Architecture of TLSTM

The following equation explains the formulas of TLSTM.

$$I_{t} = \sigma \left( {w_{i} \cdot \left[ {H_{t - 1} ,\,y_{t} } \right] + a_{i} } \right)$$
(1)
$$O_{t} = \sigma \left( {w_{O} \cdot \left[ {H_{t - 1} ,\,y_{t} } \right] + a_{O} } \right)$$
(2)
$$F_{t} = \sigma \left( {w_{F} \left[ {H_{t - 1} ,\,y_{t} } \right] + a_{F} } \right)$$
(3)
$$t_{t} = 1 - \tanh (I_{t} )$$
(4)
$$\mathop {D_{t} }\limits^{\sim} = \tanh \left( {w_{d} \cdot \left[ {H_{t - 1} ,\,y_{t} } \right] + a_{d} } \right)$$
(5)
$$D_{t} = F_{t} * D_{t - 1} + t_{t} * \mathop {D_{t} }\limits^{\sim}$$
(6)
$$H_{t} = O_{t} * \tanh \left( {D_{t} } \right)$$
(7)

From the above equations, the input gate weights, forget gate bias, and forget gate weights are \(w_{i}\), \(a_{F}\) and \(w_{F}\). The bias output gate, output gate weights, and candidate memory cell bias are \(a_{O}\), \(w_{O}\) and \(a_{d}\). Where, \(w_{d}\) and \(a_{I}\) are the candidate cell weight and the input gate bias. Obtain the relevant values based on the previous TLSTM unit in which the candidate memory cell, forget gate, input, and output gate are entered via the input data of the current moment. After that, the output value of the input gate is transformed via the 1-tanh function. Finally, calculate the present memory cell state, previous TLSTM memory cell state, candidate memory cell state, transformed output value of the input gate, and candidate memory cell state. According to the output gate value and the current memory cell state, calculate the output value of TLSTM.

3.2 Adaptive Teaching–Learning-Based Optimization (ATLBO) Algorithm

The superior student considers one student in a realistic class. The student has an active learning ability and good self-learning capacity. Here, we discussed the adaptive teaching–learning-based optimization (ATLBO) algorithm regarding the actual ‘teaching–learning’ situation that reveals the higher convergence speed and better solution quality when compared to the TLBO algorithm. The teaching and learning phases are two major stages of the ATLBO algorithm [19]. The following section explains the ATLBO algorithm.

3.2.1 Teaching Stage

The influence effects of the teacher on learners inspire the conventional TLBO algorithm. In different classes, consider two teachers teaching subjects. Teacher T1 demonstrates optimal results while teacher T1 performs better than teacher T2. In the instructor phase of the TLBO [20], the mark is extremely important for enhancing individual marks. When the fitness value of jth students falls below the mean mark for the minimal optimization issue, consider them superior. Equation (8) updates the obtained knowledge from self-study and best individuals.

$$\begin{gathered} Y_{new,j} = \left( {Y_{old,j} + \left( {random - 0.5} \right) \times 2 \times \left( {Y_{mean} - Y_{old,j} } \right)} \right) \times \lambda_{1} + D \times \lambda_{2} , \hfill \\ \,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,,\,\,\,\,\,\,\,if\,\,\,\,\,F\left( {Y_{old,j} } \right) > F\left( {Y_{mean} } \right) \hfill \\ \end{gathered}$$
(8)
$$Y_{new,j} = Y_{old,j} \times M + \left( {Y_{best} - Y_{old,j} } \right) \times random,\,\,\,\,\,\,\,if\,\,\,\,\,F\left( {Y_{old,j} } \right) < F\left( {Y_{mean} } \right)$$
(9)

From this, the mean fitness value \(F\left( {Y_{mean} } \right)\) is less than the old fitness value \(F\left( {Y_{old} } \right)\). At the starting stage, the student receives knowledge from his teacher. The mean mark and the inertia weights are \(Y_{mean}\) and \(M\). In the initial steps, the student’s knowledge is improved via teachers. Equation (10) shows the updating mechanism.

$$M = \lambda_{start} - \left( {\lambda_{start} - \lambda_{end} } \right) \times \frac{iteration}{{iteration_{\max } }}$$
(10)
$$\lambda_{1} = \sin \left( {\frac{\pi }{2} \times \frac{iteration}{{iteration_{\max } }}} \right)$$
(11)
$$\lambda_{2} = \cos \left( {\frac{\pi }{2} \times \frac{iteration}{{iteration_{\max } }}} \right)$$
(12)

The maximum number of iteration is \(teration_{\max }\) with the current iteration \(\left( {iteration} \right)\). Linearly descend the inertia weight from \(\lambda_{start}\) to \(\lambda_{end}\). At the initial steps, the search spaces are explored by allowing the ATLBO algorithm to adjust the inertia weight process. The convergence speed is accelerated by introducing \(\lambda_{1}\) and \(\lambda_{2}\). At the latter steps, the population diversity is increased to avoid trapping into the local optimum. The student’s knowledge is improved with the important role. The convergence speed and the solution qualities are enhanced via two inertia weights regarding the analysis.

3.2.2 Learning Stage

Store all learners’ fitness values in ascending order next to the teaching phase. Consider the top half of the students in the first class and the rest of the students in the second class. The superior students regard the first group members. Based on Eqs. (13) and (14), updates the initial class and second class results.

$$\begin{gathered} If\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,F\left( {Y_{old,j} } \right) > F\left( {Y_{neighbour} } \right) \hfill \\ \,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,Y_{new,j} = Y_{old,j} + \left( {Y_{neighbour} - Y_{old,j} } \right) \times \cos \left( {\frac{\pi }{2} \times \frac{iteration}{{iteration_{\max } }}} \right)\,\, \hfill \\ Else \hfill \\ \,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,\,Y_{new,j} = Y_{old,j} + \left( {random - 0.5} \right) \times 2 \times \left( {Y_{ul} - Y_{ll} } \right) \hfill \\ End \hfill \\ \end{gathered}$$
(13)
$$Y_{new,j} = Y_{old,j} + \left( {Y_{best} - Y_{old,j} } \right) \times \cos \left( {\frac{\pi }{2} \times \frac{iteration}{{iteration_{\max } }}} \right)$$
(14)

The ATLBO algorithm pseudocode clearly explains the implementation procedures as discussed in Algorithm 1.

figure a

3.3 Hybrid CNN-TLSTM with ATLBO Algorithm

The structure of hybrid CNN-TLSTM with the ATLBO algorithm is delineated in Fig. 3. A set of m layers is the convolutional layer. The feature extraction on various time periods is performed by the convolution kernel size. Set n × n convolution kernel size to maximize the previous data. Before the pooling layer, batch normalization is added to enhance the efficiency of training. The global feature extraction is performed by compressing data [21]. The prediction ability is improved by increasing the number of CNN-TLSTM networks. The ATLBO algorithm prevents overfitting in the dropout layer and the number of neurons in each layer of the CNN-TLSTM model is minimized.

Fig. 3
figure 3

Structure of hybrid CNN-TLSTM with ATLBO algorithm

The below equation expresses the individual fitness function.

$$Fitness = \frac{1}{Error}\left( {1 + \beta \left( {1 - \frac{{m_{Con} }}{{m_{All} }}} \right) + \gamma \left( {1 - \frac{{m_{LSTM} }}{{m_{All} }}} \right)} \right)$$
(15)

Based on Eq. (15), the MAPE error and the fitness function are \(Error\) and \(Fitness\) computed. The influences of LSTM and convolution layer on network performance are \(\gamma\) and \(\beta\). The total number of convolutional kernels in CNN and the number of neurons in the LSTM is \(m_{Con}\) and \(m_{LSTM}\). The major advantage of CNN-TLSTM is to predict the time series data and thereby avoiding the gradient disappearance problem. But it consumes more time and is complete to select the most relevant parameters. The enhancement of the CNN‐TLSTM model is performed using the ATLBO algorithm [22]. The ATLBO algorithm optimizes the CNN and TLSTM modules. Obtain the best individual and updater the better one.

4 Proposed Framework

The vital acquisition element of IoT technology is integrated using several real-time applications. The massive data extraction, health-related data management, processing, and data storage are performed using IoT. The cost-effectiveness, storage capacity, accessibility, and scalability are provided via cloud computing. The remote patient health monitoring system is realized based on the healthcare agencies. The overall block diagram of the proposed method is illustrated in Fig. 4. Several numbers of medical sensors and IoT sensor-based devices provide the data for data acquisition. The infected patients are identified and also the details are forwarded to the fog layer. After that, the data classification is performed using a hybrid CNN-TLSTM with ATLBO algorithm and it generates an alert message. The detailed process of the proposed work is discussed as follows:

Fig. 4
figure 4

Block diagram of the proposed model

The hybrid CNN-TLSTM with the ATLBO algorithm is the one that is accommodated for dengue disease prediction. Based on the patient's historical data stored in the cloud, the proposed model identifies the health state of a person. The results are mainly based on the different disease and environmental parameters such as Soft Bleeding, Muscle Pain, Joint Pain, Skin rash, Fever, Water Site, Carbon Dioxide, Water Site Humidity, and Water Site Temperature. The hybrid CNN-TLSTM architecture mainly depends on three stages: disease monitoring, training, and prediction. The hybrid CNN-TLSTM architecture assigns unique values for each parameter based on their importance. The probabilistic values are mainly determined by the domain experts. The monitoring process mainly happens based on the computational capability of the fog layer and it stores the parameters based on the temporal data.

4.1 Data Acquisition

In real-time from the different sensors, the health-related data is collected using the data acquisition layer [23]. The remote devices transit the acquired information that is responsible for data analysis and it acts as the fog node. Over the wireless network, transmit the data for in-depth analysis and detailed processing. Address the security and privacy concerns of IoT networks in addition to data accumulation. Effective data security is provided with several protocols. All data to and from IoT devices must be encrypted using stronger data security methods such as the Elliptic Curve Cryptography (ECC) algorithm and the Secure Socket Layer (SSL) protocol. Over the Message Query Telemetry Protocol (MQTP) and Hypertext Transfer Protocol (HTTP), these protocols guarantee secure communication [24]. The IoT devices are implemented in a software-level platform with an easy-to-use Application Programming Interface (API). The Amazon Web Services (AWS) architecture allows for increased efficacy and accuracy in data storage. Every data value is subjected to real-time processing at the fog layer. If an intrusion is discovered, the system processes all data values in real-time.

4.2 Hybrid CNN-TLSTM with ATLBO Algorithm for Data Classification and Visualization

It is important to formulate the data into multiple groups after the heterogeneous data is accumulated from various sensor devices. For better classifications, apply the extraction techniques because of data heterogeneity. Disorientation, Coma, Joint pain, Paralysis, Nausea, Vomiting, and High fever are the vital symptoms present in the health data. These kinds of data are collected by using different health-related medical sensors. The data analysis is required to detect health-related vulnerability. The fog layer detects patients using the health state categorization components. The real-time analysis of heterogeneous data is carried out at multiple time slots (△t) to use fog computing nodes [25]. The fog layer detects the patients by using the health state categorization components.

In this study, we used hybrid CNN-TLSTM with the ATLBO algorithm (as discussed in Sect. 3) for dengue disease data prediction and classification. The major task is Spatio-temporal detection in the fog layer. An event in time data value is stored on a primary basis. The event manifestation is assigned to the Spatio-Temporal Granulation [26]. The proposed hybrid CNN-TLSTM with the ATLBO model classifies the identical data instances into two classes which denotes the presence or absence of dengue illness based on the dengue accumulated features. While taking the dataset partition, the cost function of hybrid CNN-TLSTM with ATLBO reduces the error [27]. Based on the classification results, 0 means the absence of dengue, and 1 is for the presence of dengue. Finally, the patient's result is sent to their mobile devices as an alert message.

5 Experimental Results and Discussion

This section evaluates the performance of the proposed model using state-of-art comparisons with different evaluation criteria for its performance. Table 1 explains the hardware and software used for implementation.

Table 1 Hardware and software used for implementation

The optimal parameter setting with its ranges of hybrid CNN-LSTM with ATLBO algorithm is tabulated in Table 2.

Table 2 Optimal parameter settings

5.1 Dataset Explanation

The health attributes based on dengue fever are explained in Table 3. The environmental attributes like temperature, carbon dioxide, humidity, and mosquito-dense sites are captured and deployed in a total of 1240 sensors in different locations [28]. To generate environmental attributes random values, the send script program these sensors [29]. The attribute value, Sensor ID, and sensor location are present in the environmental dataset. From this, the symptoms of Yes are referred to as Y and No are denoted as N.

Table 3 Health attributes description

During simulation, different parameters were altered for tuning. The health-related parameters of the patients are collected via the IoT devices which are interconnected with different devices for healthcare data transmission. The application also provides a form for the patients to enter if they feel any abnormalities such as soft bleeding, muscle pain, joint pain, skin rash, severe abdominal pain, vomiting, pain behind the eyes, severe headache, etc. With these details, the patient's health conditions are sent to edge computing for processing. The physical movements of the patients are also analyzed by their wearable devices. These data gathered are then preprocessed by the Hybrid model.

The missing values are rectified via the imputation and the TLSTM technique. The noise is minimized to improve the detection accuracy and the data is mainly analyzed in terms of rows and columns. To minimize the complexity associated with the diagnosis process, the normalization takes place in the 0 to 1 range for multiple data distribution. The data is obtained from a total of 1500 patients in the rural areas via different sensing gadgets such as smartwatches, mobile phones, etc. Here 70% of data is taken for training and the remaining 30% is taken for testing. For each parameter, their associated threshold value is computed based on their symptoms. There are a total of 6 environmental-related symptoms and 10 health-related symptoms with a label yes or no. We generate the synthetic dataset by allocating the probability value to the health and environmental-related attributes. The probabilities generated for the different dengue symptoms are presented in Table 4.

Table 4 Probabilities generated for different dataset symptoms

5.2 Performance Metrics

The performance measures like accuracy, precision, recall, and F-score values are delineated in the following equations [30], which are essential to evaluate the performance of the proposed model.

$$Accuracy = \frac{Tn + Tp}{{Tn + Tp + Fn + Fp}}$$
(16)
$${\text{Re}} call = \frac{Tp}{{Tp + Fn}}$$
(17)
$$\Pr ecision = \frac{Tp}{{Tp + Fp}}$$
(18)
$$F - measure = \frac{2Tp}{{2 \times \left( {Fp + Fn + Tp} \right)}}$$
(19)

where the true positive and true negative classes are represented as \(Tp\) and \(Fn\) with the false positive and false negative classes are represented as \(Fp\) and \(Fn\).

Equation (20) delineates the mean square error values.

$$MSE = \frac{1}{m}\sum\limits_{j = 1}^{m} {\left( {x_{j} - {\hat x_j} } \right)}$$
(20)

The true and predicted values are \(x_{j}\) and \({\hat{x}_{j} }\). If the prediction capability of the model is high means the MSE is closer to zero. Equation (21) explains the mean absolute percentage error (MAPE),

$$MAPE = \frac{100\% }{m}\sum\limits_{j = 1}^{m} {\left| {\frac{{\hat{x}_{j} - x_{j} }}{{x_{j} }}} \right|}$$
(21)

The true and predicted values are \(x_{j}\) and \({\hat{x}_{j} }\). The limit \(\left[ {0,\,\infty } \right]\) is the range of MAPE. The better classification accuracy is attained with the smaller value of MAPE.

5.3 Performance Analysis

The performance of accuracy based on the number of epochs is shown in Fig. 5. Figure 6 describes the performance of loss based on the number of epochs [31, 32]. These graphs showed the effectiveness of the proposed method. During dengue fever prediction, the proposed hybrid CNN-TLSTM with the ATLBO model reveals higher training and testing accuracies with lower training and testing loss values. Hence, the hybrid CNN-TLSTM with the ATLBO method is well suitable for dengue fever prediction.

Fig. 5
figure 5

Performance of accuracy based on the number of epochs

Fig. 6
figure 6

Performance of loss based on the number of epochs

Table 5 evaluates the training and testing time based on different health records. The training and testing time varies based on a varying number of health records from 200 to 1000 [33,34,35]. Based on 1000 health records, we have obtained 0.034 s of training time and 0.024 s of testing time respectively. The state-of-art results based on the health records are plotted in Fig. 7 in which both training and testing times are plotted in Fig. 7a and b. The methods like I-SCSM, EEFCS, DNN, and proposed hybrid CNN-TLSTM with ATLBO models with varying times are taken. The proposed method takes lower training and testing time than other existing methods like I-SCSM, EEFCS, and DNN respectively.

Table 5 Evaluation of training and testing time
Fig. 7
figure 7

State-of-art results based on the health records, a Training time, and b Testing time

The comparative analysis of accuracy with respect to a varying number of health records is plotted in Fig. 8. For this experiment, the number of health records varied from 200 to 1000 with respect to state-of-art techniques like I-SCSM, EEFCS, DNN, and the proposed model. From this investigation, the overall accuracy based on dengue fever records with the proposed hybrid CNN-TLSTM with ATLBO method offers superior accuracies when compared to the existing studies like I-SCSM, EEFCS, and DNN.

Fig. 8
figure 8

State-of-art accuracy based on varying number of health records

The state-of-art comparison of mean absolute percentage error is shown in Fig. 9. This investigation is performed using state-of-art techniques like multilayer perceptron layer (MLP), Recurrent Neural Network (RNN), LSTM, CNN, CNN-TLTM, and the proposed method. From this investigation, the MAPE value of MLP is 0.0412, RNN is 0.2892, CNN is 0.3092, LSTM is 0.243, CNN-TLSTM is 0.2012 and the proposed method is 0.1856. From this investigation, the proposed technique has fewer MAPE values than all other methods such as MLP, CNN, LSTM, RNN, CNN, and CNN-LSTM.

Fig. 9
figure 9

State-of-art comparison of mean absolute percentage error

The state-of-art comparison of mean square error is plotted in Fig. 10. The state-of-art techniques like CNN, LSTM, RNN, MLP, CNN-TLTM, and proposed hybrid CNN-TLSTM with ATLBO method. The methods like LSTM, RNN, CNN, MLP, CNN-LSTM, and proposed methods demonstrated the mean square error values of 0.00047, 0.00084, 0.00090, 0.00125, 0.00042, and 0.00036 respectively. However, the proposed method demonstrated lower mean square error values than the other previous methods.

Fig. 10
figure 10

State-of-art comparison of mean square error

The state-of-art result of classification performance is plotted in Fig. 11. This graph plots the accuracy, precision, recall, and F-score values of the proposed method. During dengue fever classification, the overall classification performance is analyzed using various state-of-art techniques such as I-SCSM [11], EEFCS [13], DNN [14], and 1D-CNN [16], the proposed hybrid CNN-TLSTM with ATLBO method. From this examination, we have attained 96.9% accuracy, 95.7% precision, 96.8% recall, and 96.2% F-measure values. Among all the existing techniques, the proposed hybrid CNN-TLSTM with the ATLBO method provided higher classification results.

Fig. 11
figure 11

State-of-art result of classification performance

The performance analysis of reliability is plotted in Fig. 12. Additionally, the proposed method is tested for reliability in order to quantify the general outcomes obtained. The classification model is tested and changed for confirmation. When compared to other models, the introduced model achieved superior performances, according to the implementation findings. In comparison to the other techniques of I-SCSM [11], EEFCS [13], DNN [14], and 1D-CNN [16], the proposed model is believed to have superior reliability based on the final result. In comparison to previous state-of-the-art edge techniques, execution prediction analysis shows that the suggested model is extremely reliable.

Fig. 12
figure 12

Performance analysis of reliability

To evaluate the performance of the results one step further, the proposed methodology is compared with different techniques such as Cascading ensemble of CNN [36]. Multimodel classification [37], Deep Reinforcement Learning (DRL) [42] in terms of accuracy, latency, and time complexity. The results obtained are shown in Table 6 for a total of 500 patient records. The table is self-explanatory. The use of fog computing minimizes the latency of this approach and the ATLBO optimized hybrid CNN-TLSTM model selects the accurate parameters (geographical and disease) for disease diagnosis. The results show that the proposed model offers improved accuracy, low latency, and minimal time complexity.

Table 6 Performance analysis

The ATLBO which tunes the hybrid CNN-TLSTM architecture minimizes the training and testing time and also shows improvements in terms of different parameters. The incorporation of the ATLBO algorithm minimizes the local optima problem and helps to discover efficient solutions. In this way, the hybrid CNN-TLSTM architecture assigns appropriate disease labels for effective decision-making. After the report is obtained from the hybrid CNN-LSTM technique it is transmitted to the fog-cloud server. The fog-cloud server sends the results to the healthcare organizations which takes effective decisions based on the report. The main limitation of this technique is the absence of the information regarding the Complete blood count which helps to identify the severe blood loss associated with dengue.

6 Conclusion

This article presented Hybrid CNN-TLSTM with ATLBO algorithm for Dengue disease identification and prevention. While compared to the LSTM, RNN, CNN, MLP, and CNN-LSTM methods, the proposed model outperforms lower error values in terms of both MSE and MAPE but the overall accuracy is higher. The training and testing time of 0.034 and 0.024 s are obtained based on the varying number of health records. Several simulations are carried out for validation in which the experimental outputs are contrasted to state-of-art techniques like I-SCSM, EEFCS, DNN, and 1D-CNN. While comparing several statistical parameters, the simulation results concluded that the proposed method is more effective and better. The performance is evaluated in terms of different metrics such as accuracy, recall, precision, and F-score. The proposed ATLBO optimized hybrid CNN-TLSTM offers accuracy, precision, recall, and F-measure values of 96.9%, 95.7%, 96.8%, and 96.2% respectively. Due to the usage of the fog based computing strategy, the proposed model offers a lower latency of 0.075 which is relatively lower than the Cascading ensemble of CNN (0.071), Multimodel classification (0.054), DRL (0.079), 1DCNN (0.086), and DNN (0.093). The proposed model's efficiency is observed by the results obtained for different statistical parameters. In this work, we have created a healthcare model to meet the needs of dengue disease diagnosis irrespective of gender, age, or race. In the future, we plan to incorporate more disease-related parameters and also improve the security of the patient's credentials stored.