1 Introduction

Corona-genomic viruses that are member of the Coronaviridae species have viruses consisting of large strands of RNA sequences. The RNA is sized about 27–32 kb engulfed with polyadenylated. In each cluster, viruses pigeon-hole their way into host cells genomic sequence. This so-called ‘Corona’ virus got first identified in animals such as mice, horses and many reptiles as well as cows, bullocks etc. that led to severe infections some of which gastric, respiratory tract ones. Predominant infections associated with this virus have been ‘respiratory’, ‘gastric’ however, ‘hepatic’ or ‘neurological diseases’ recorded.

Till date as per various studies reported the structure has been described to be as engulfed non-segmented positively oriented RNA-genomic viral classifying into the category Corona-viridae on the bases of characteristics. There have been a lot of research into the RNA genome of this virus. One of the key outcomes involve that its effect on the host RNA is based on individual-to-individual response which means coronavirus infections could be mild categorized as Beta coronaviruses.

Studied multifarious diseases regarding various characteristics such as usages, action-mechanisms and various others towards treatment of pandemic [1]. Exploration of statistical simulations to study the dynamics of HIV, estimation of river water quality through artificial intelligence, COVID-19 pandemic, complexity analysis for China region modelled to study in detail [2,3,4,5,6]. Studied human corona via glycoproteins in receptor-binding sites [7]. Explained time series transformations required in analysis [8]. Modelled the phases of transmission in which novel coronavirus gets transferred [9]. Discussed a dynamic model indicating transfer through hands, foots, mouths etc. [10]. Analyzed the role of ‘Chloroquine’ during SARS-CoV-2 [11]. Discussed an epidemiology model for HCV-infections [12]. Randomness via fuzzy random set theory modelled for Plasma disruption in fusion reactors [13]. Detailed explanation for corona-theorem with the applicability into spectral-problems [14]. Discussed the newest variant, Omicron [15]. Explored the COVID-19 herd immunity [16]. Studied respiratory-tract in human beings for novel virus isolation [17]. Explained in detail the coronavirus related concerns [18]. Explained the AIDS mechanism so as to study viral diseases [19]. Studied as well as explained the second and third waves of coronavirus [20].

Discussed the effects of coronavirus in lower respiratory-tracts observed in infants [21]. Detailed study of Corona to understand the pneumonia infection in low immunity-patients [22]. Discussed ‘Corona-theorem’ for ‘countably-many functions’ [23]. Pathogenesis of COVID-19 epidemiology in Saudi discussed with the help of nonlinear time series [24]. WHO reported ‘novel-coronavirus’. Survey on ‘MERS-CoV’ updates [25, 26]. Studied the respiratory syndrome during human coronavirus via molecule analysis [27]. WHO stated updates on Coronavirus around January, 2020 along with the discussion about the ‘pneumonia-outbreak’ originating from bats developing into ‘corona-virus’. Discussed important antibiotics that would hinder the entrance of ‘Ebola virus’, ‘MERS-CoV’, ‘SARS-CoV’ in human-beings as well as other living beings. WHO described ‘novel coronavirus’ updated reports [28,29,30,31].

None of the authors designed the wavelet conjuncted neuronal network fuzzified inferences’ layered hybrid model for countrywise daily data COVID-19 genome transmission. In this study, forecasts of the datewise data of confirmed cases. In our forecasting models, we have made an attempt of artificial intelligence (AI) based learned algorithms, such as WNNFIL, WNNFIL–LSSVR, WNNFIL–MARS. Forecasts via validated data for longer time spans have been evaluated on the basis of error performances through RMSE, MAE, \( \hbox {R}^{2} \) .

2 Chaotic analysis

2.1 Lyapunov characteristic exponent (LCEs)

Alexander M. Lyapunov, a Russian engineer first discussed the concept of characteristic exponents hence the name Lyapunov Characteristic Exponents (LCEs) became widely known. It is used to measure the stability of the variables considered under the system. Discussed the degree of strangeness, i.e., randomness that led to strange attractors’ formation. Such exponents have been observed to be the direct indicators as well as quantifiers for determining randomness. The number of Lyapunov exponents depend upon the dimension of the phase diagram. The LCEs \( < 0 \) indicates nearby initial conditionalities converging, i.e., delta-errors decreasing through delta-time increment. However, whenever the LCEs are positive, the infinitesimally nearby initial conditions diverging exponentially, i.e., errors increasing rapidly with respect to time increments. Such a phenomena is known as the sensitivity towards initial conditions famously known as Chaos. LCEs are considered as the standardizations for eigenvalues computed in various states, such as steady-state, limit-cycles or randomness. Determination of these characteristics for non-linear systems involves numerically integrating underlying differential equalities over the variation in parameters and initial conditions.

Steps to compute Lyapunov Exponents:

  1. 1.

    Consider \( x_{n\, }= \quad f_{n}(x_{0}) \) ; \( y_{n\, }= \quad f_{n}(y_{0}) \) then, the nth iteration of orbits for \( x_{0\, }\)and \(y_{0\, } \) under f are: \( \vert x_{0} -{\, }y_{0} \vert \ll 1{\, },\vert x_{n} -{\, }y_{n} \vert \ll 1, \)

  2. 2.

    Then, the divergence at that \( n^{th} \) iteration in case of 1-dimension setup is calculated as the summation of

    $$\begin{aligned} \vert x_{n} -y_{n} \vert \,\approx \left( {\prod \limits _{t=0}^{n-1} {\vert f'(t)\vert } \,\vert x_{0} -y_{0} \vert \vert } \right) \,\,\,\,\,\,\, \end{aligned}$$

    having \( \vert x_{0} -{\, }y_{0} \vert<<1{\, },\vert x_{n} -{\, }y_{n} \vert<<1 \) ; \( x_{n} ={\, }f_{n} (x_{0} ),{\,}y_{n} ={\, }f_{n} (y_{0} ) \) are \( n^{th}\,\) iterations of orbits of \( x_{0} \) and \( y_{0} \) under f.

  3. 3.

    Next, exponential rate of divergence, i.e., \( log\vert f'\left( x \right) \vert \,\, \) for two localized initial conditions formulating into

    $$\begin{aligned}&\lambda (x_{0} )=\lim \limits _{n\rightarrow \infty } \frac{1}{n}\log \left( {\prod \limits _{t=0}^{n-1} {\vert f'(x_{t} )\vert } } \right) \,\,\,\,\,\,\,\, \\&\text { where } \prod \limits _{t-0}^{n-1} {\vert \hbox {f}'(\hbox {x}_{t} )\vert \approx \hbox {e}^{\lambda (x_{0} )n}} \text { for }, n\gg 1. \end{aligned}$$
  4. 4.

    Hence, \( \lambda (x_{0} ) \) can be defined as LCEs for the trajectory of \( x_{0} \).

  5. 5.

    Thus, two trajectories in phasespace having initial-separation, \( \delta x_{0} \) diverge \( \vert \delta \hbox {x(t)}\vert \approx \hbox {e}^{\lambda t}\vert \delta \hbox {x}(0)\vert \,\, \) where \( \lambda >0 \) denotes LCE.

  6. 6.

    Consider \( \lambda _{1} ,\lambda _{2} ,\ldots ,\lambda _{n} \) be the eigen-values for the linearized-equation \( \frac{{\text {d}}u}{{\text {d}}t}=A\left( {u^{*}} \right) \) s.t. \( m_{1} \left( t \right) =e^{\lambda t} \) and \( {\mathop {\lambda }\limits ^{\sim }}{i} =\lim \nolimits _{t\rightarrow \infty } \frac{1}{t}\ln \left| {e^{\lambda _{i} t}} \right| =\hbox {Re}\left[ {\lambda _{i} } \right] \) .

Thus, LCEs can be taken as equivalent to real parts of calculated eigenvalues at all of the fixed points. Whenever these exponents become less than zero, the localized initial conditions become convergent depicting (small) delta-errors are decreasing. The subset formed is also known as attracting set. Therefore, attractor can be defined as an attracting set that consists of a dense orbit. Whereas, whenever the LCEs become positive, the initial conditions become divergent, i.e., (small) delta-errors are increasing. Behavior of attractors as per LCE values being classified into

  • Equilibria: \( 0>\lambda _{1} \ge \lambda _{2} \cdots \ge \lambda _{n} \) ;

  • Periodic limit-cycling: \( \lambda _{1} =0,\,\,0>\lambda _{2} \ge \lambda _{3} \ldots \ge \lambda _{n} \) ;

  • k -periodic limit-cycling: \( \lambda _{1} =\lambda _{2} =\cdots \lambda _{k} =0,\,\,0>\lambda _{K+1} \ge \lambda _{K+2} \cdots \ge \lambda _{K+n} \) ;

  • Strange-randomness, i.e., chaotic: \( \lambda _{1} >0,\sum \nolimits _{i=1}^n {\lambda _{i} } <0 \) ;

  • Hyper-randomness, i.e., hyper-chaotic: \( \lambda _{1}>0,\lambda _{2} >0,\sum \nolimits _{i=1}^n {\lambda _{i} } <0 \) .

2.2 Rescaled R/S analysis

As proposed by H.E. Hurst, Rescaled-range analysis (R/S analysis) involves the understanding of long-term persistence occurring in the time successions. It begins via splitting of timeseries, \( s_{i} \) assuming non-overlapping sections \( \phi \) having magnitude, p resulting in \( K_{p} {\, }={\,}int\left( {K/p} \right) \) segments altogether. Next step, this integrated data is calculated in every section as

$$\begin{aligned} X_{\phi } ({\mathrm{i}})= & {} {\mathop {\sum }\limits _{t=1}^i} {\left( {\mathrm{s}}_{\phi p+t} -\left\langle {s_{\phi p+t} } \right\rangle _{\mathrm{p}} \right) } \\= & {} {\mathop {\sum }\limits _{t=1}^i} {\mathrm{s}_{\phi p+t} } -\frac{i}{p}{\mathop {\sum }\limits _{t=1}^p} {\mathrm{s}_{\phi p+t} } ;\\&\phi =0,\ldots K_{p} -1 \end{aligned}$$

Now, piecewise constants’ trend and various other averaging can be handled through subtraction. Next step involves recording ranges through maximum and minimum values and computing standard deviations in each section considered above denoted as

$$\begin{aligned} \begin{array}{l} R_{\phi } (p)={\max }_{i=1}^{p} X_{\phi } (\mathrm{i})-{\min }_{i=1}^{p} X_{\phi } (\mathrm{i}), \\ S_{\phi } (p)=\sqrt{\frac{1}{p}\sum \limits _{i=1}^p {X_{\phi }^{2} (\mathrm{i})} } \\ p(k)=Ck^{-\alpha };\,\, \\ C-\text {constant};\,\,\,p(k)-ACF\,\text {with}\,\text {lag}\,`{\textit{k}}{\text {'}} \\ {H=1-}\frac{\alpha }{2} \\ \end{array} \end{aligned}$$

Lastly, as the name suggests range needs to be rescaled over all sections to capture the fluctuation in function form, \(\hbox {F}_{{\mathrm{R}}/{\mathrm{S}}} \) as

$$\begin{aligned} F_{R/S} (p)=\frac{1}{K_{p} }\sum \limits _{\phi =0}^{K_{p} -1} {\frac{R_{\phi } (p)}{S_{\phi } (p)}} \sim p^{H_{e} },\,\,\text {for}\,\,{p>>1} \end{aligned}$$

having \( H_{e} \)—Hurst exponent as desired.

In addition, \( H_{e} \) calculated from slope of the regression fitline plotted for the log–log curve of the variable with respect to time.

Remark 3

From Remarks 1 and 2, \(H_{e} \) can be calculated in terms of multi-fractal & spectral analysis as: \( 2H_{e}\approx 1+\eta =2-\varpi \) .

Remark 4

Clearly, whenever \( 0<\varpi <1 \), then \(0.5< H_{e}<1\). In general, this relation may not hold for all multifractal analysis.

As and when the data is stationary, \( H_{e} \) using rescaled R/S. Thus, \( H_{e}< 1/2\) specify long-term anti-correlation among data, and \( H_{e}>1/2\) shows positive correlations. In addition, power-law correlations decay sooner than 1/p, taking \(\hbox {H} = 1/2\) for larger p values.

3 Machine learning-based forecast models

Variable considered with respect to time successions’ corpora that needs to be put together for real-time response variables’ approximation towards nCOVID-19. Furthermore, it is taken as mentioned by the designated authorities that these are the daily-confirmed cases nearly from last week of January for the study of variations in time-periods in accordance with various data sets for simulations. Everyday active cases data for United States and India. For this forecast model case study, datewise data of Confirmed cases recorded in India & USA is taken as available on the designated authorities’ reports and bulletins.

The unprecedented and unforeseen nCov-19 spread had its own course world over. Both the countries studied in this article had their own different timelines for the spread of outbreak. The data sets of USA seemed to have greater variability as compared to that of India. Therefore, both the countries’ confirmed cases became important towards the study of this pandemic. Uncertainty associated with the available data invites a model that can forecast the spread of an infectious disease. To understand the dynamics of the disease including its mode of transmission, we need to have intelligent forecasting models with respect to realistic human data (confirmed, deaths and recovered cases). Accordingly, we have developed innovative forecasting models for the country-based datewise data of confirmed cases.

3.1 Discrete waveform signal analysis (DWS)

Theorem

Sequence consisting of embedded-approximation-subsets to be taken as

$$\begin{aligned} 0\leftarrow ....V_{j-i} \subset V_{j} \subset V_{j+1} ....\rightarrow R \end{aligned}$$

having \( f(x)\in V_{j} - f(2x)\in V_{j+1} \)

$$\begin{aligned} \begin{array}{l} f(t)\in V_{0} \Leftrightarrow f(t-k)\in V_{0} ,k\in Z \\ \varphi (t-k)_{k\in Z} \;\text { forming an orthonormal basis } \\ \end{array} \end{aligned}$$

plus, the sequence of orthogonal complements, details’ subspaces:

$$\begin{aligned}&W_{n} \text { s.t. } V_{n+1} =V_{n} \oplus W_{n} .\\&\varphi - \text { scaling function, i.e., a low-pass sieve.} \end{aligned}$$

Remark

Basis, \(V _{\{j\}}\) to be understood as

$$\begin{aligned} \varphi _{j,k} (t)=2^{j/2}\varphi (2^{j}t-k),k\in Z \end{aligned}$$

For non-stationary time-dependent series, detrended fractal investigation of time successions can be easily implemented by considering wavelet coefficients of the type \( X_{\phi } (i) \) . In this case, the convolution corresponds to the basic arithmetic such as addition/subtraction for mean values of \( X_{\phi } (i) \) within sections having magnitude, p .

The wavelet coefficients \( W_{\varphi } \left( {\kappa \hbox {,p}} \right) \) are based on both time position \( \kappa \) & scaling, p . The waveform transforms required for the data signals, \( d_{s} \) is calculated through the following summation for discrete waveform signals of the time series, \( s_{i}{, i} = 1,\ldots , K\):

$$\begin{aligned} W_{\varphi } \left( {\kappa \hbox {,p}} \right)= & {} \frac{1}{p}\int _{-\infty }^\infty {d_{s} } (t)\varphi [(\mathrm{t}-\kappa )/p]dt\,\,,\,(continuous) \\= & {} \frac{1}{p}\sum \limits _{i=1}^K {d_{s} (\mathrm{i})} \varphi [(\mathrm{t}-\kappa )/p]\,\,,\,\,\,(\text {discrete}) \\ \end{aligned}$$

In addition, the local frequency decomposition of the signal is described by the time resolution apt for the considered frequency \( f = \textit{ 1/p}\).

3.2 Least squares support vector regression (LSSVR)

LS-SVRs handle higher complicacies as compared to SVR as described by Suykens. Objective-function does not change much as compared to that of the classical one. Difference arises when \( \varepsilon \) -based loss function replaces the classical squared-loss function. It has been observed that regression through this approach annihilates noise alongwith monitoring of computational-labour.

Mathematically, it can be understood as

$$\begin{aligned} {\phi _{2} (\omega ,h,\in )=\gamma A_{W} +\zeta A_{D} } \end{aligned}$$

with

$$\begin{aligned} A_{W}= & {} \frac{1}{2}\omega ^{T}\omega \,\,\text { and }\,\,\, \\ A_{D}= & {} \frac{1}{2}\mathop \sum \limits _{i=1}^{K} \in _{i}^{2} \\= & {} \frac{1}{2}\mathop \sum \limits _{i=1}^{K} \left( {y_{i} -\left( {\omega ^{T}\sigma \left( {x_{i} } \right) +h} \right) } \right) ^{2} \end{aligned}$$

where \( \gamma \,,\,\zeta \) - hyper-parameters tuning with respect to SSE

Hence, the LSSVR regressor resultant that can be achieved from the Lagrangian-function as followed:

$$\begin{aligned} \left\{ {\begin{array}{lllll} L_{m} (\omega ,h,\in ,\alpha )=\phi _{2} (\omega ,\in )\\ \quad -\sum \limits _{i=1}^K {\alpha _{i} } \left\{ {\left[ {\omega ^{T}\sigma \left( {x_{i} } \right) +h} \right] +\in _{i} -y_{i} } \right\} \\ =\frac{1}{2}\omega ^{T}\omega +\frac{\mu }{2}\sum \limits _{i=1}^K {\in _{i}^{2} }\\ \quad -\sum \limits _{i=1}^K {\psi _{i} } \left\{ \left[ {\omega ^{T}\sigma \left( {x_{i} } \right) +h} \right] +\in _{i} -y_{i} \right\} \\ \end{array}} \right\} \end{aligned}$$

where \( \psi _{i} \in \mathrm{R} \) as Lagrange multipliers.

3.3 Wavelet neuronal network fuzzified inferences’ layered least-squares support vector regression (WNNFIL–LSSVR)

Beginning with the NNFIL model as follows, Fig. 1:

Fig. 1
figure 1

Working of the NNFIL conjuncted model

The following theorem and lemma for the conjuncted model are proposed as:

Theorem

The simulation of the responses obtained after training through DWTS into approximations and details trained through NNFIL.

Lemma

The results are then transferred towards least squares support vector regressors’ model having ‘Gaussian function’ as kernel forming an important part for the formulation of optimization-based solution.

Thus, the flowchart in Fig. 2 clearly puts forward every step involved in the proposed model.

Fig. 2
figure 2

Flowchart of WNNFIL–LSSVR hybrid model for COVID-19

3.4 Multivariate adaptive regression spline (MARS)

Conceptually, MARS is the process of creating intricate pattern of correlations among the multi-variables known as the predictor-response variables not considering or taking into account the already known knowledge of relations that need to be hypothesized.

Thus, MARS build this model of single and multi-response data then validated and tested using separate sets. Finally, then the model can be utilized for estimation. The generalized-form MARS given as in the equation below:

$$\begin{aligned}&y=f(x)=\beta _{0}+{\mathop {\sum }\limits _{m=1}^{M}}\beta _{m}H_{km}(x_{v(k,m)})\\&y \text { - output parameter;}\\&\beta _{0}\text { - constant};\\&\hbox {M} \text { - no. of basis-function;}\\&H_{km}(x_{v(k,m)}) - m^{th} \text { basis function;}\\&\beta _{m}\text { - Corresponding coefficient of } H_{km}(x_{v(k,m)}).\\&x_{v(k,m)}\text { - predictor in the } k^{th} \text { of } m^{th} \text { product}. \end{aligned}$$

The basis-function, H is defined as

$$\begin{aligned} H_{km}={\mathop {\prod }\limits _{k=1}^{k}}h_{km} \end{aligned}$$

For the order \( K = 1\), model becomes additive, while in case of \(K = 2\), model becomes pairwise interactive

3.5 Wavelet neuronal network fuzzified inferences’ layered multivariate adaptive regression spline (WNNFIL–MARS)

The following theorem and lemma for the conjuncted model are proposed as:

Theorem

Simulation of the resultant values achieved from DWT decomposition into cAs and cDs trained through NNFIL.

Lemma

Thus, results are then transferred towards MARS technique having each feature, f mined leading to the formation of Adaptive Regression Splines.

Thus, the flowchart of Fig. 3 demonstrates every step involved in this proposed model.

Fig. 3
figure 3

Flowchart of WNNFIL–MARS hybrid model for COVID-19

Fig. 4
figure 4

Linear regression fit for WNNFIL trained model for USA

Fig. 5
figure 5

Linear regression fit for WNNFIL–LSSVR trained model for USA

Fig. 6
figure 6

Linear regression fit for WNNFIL–MARS trained model for USA

Fig. 7
figure 7

Lyapunov exponent values for Confirmed cases in USA

Fig. 8
figure 8

Wavelet denoising of confirmed cases for USA

Fig. 9
figure 9

WNNFIL trained model for confirmed USA

Fig. 10
figure 10

WNNFIL tested model for confirmed USA

Fig. 11
figure 11

WNNFIL–LSSVR model for confirmed USA

Fig. 12
figure 12

WNNFIL–MARS model for confirmed USA

Fig. 13
figure 13

Linear regression fit for WNNFIL trained model for India

Fig. 14
figure 14

Linear regression fit for WNNFIL–LSSVR trained model for India

Fig. 15
figure 15

Linear regression fit for WNNFIL–MARS trained model for India

Fig. 16
figure 16

Lyapunov exponent values for confirmed cases in India

Fig. 17
figure 17

Wavelet decomposition of daily data of confirmed cases for India

Fig. 18
figure 18

WNNFIL trained model for confirmed India

Fig. 19
figure 19

WNNFIL tested model for confirmed India

Fig. 20
figure 20

WNNFIL–LSSVR model for confirmed India

Fig. 21
figure 21

WNNFIL–MARS model for confirmed India

4 Performance measures

The estimation provided from the conjuncted models is evaluated on the basis of following performances, such as ‘root mean squared error (RMSE)’, ‘mean absolute error (MAE), ‘coefficient of determination \((R^{2})'\):

  1. (i)
    $$\begin{aligned} {\text {MSE}}=\frac{\sum \nolimits _{i=1}^n {(y_{i} -\hat{{y}}_{i} )^{2}} }{n} \end{aligned}$$

    having \( y_{i} \) is the recorded data; \( \hat{{y}}_{i} \) is the predicted assessment; n is the number of days in prediction.

    The square root for MSE is referred as ‘root mean squared error’:

  2. (ii)
    $$\begin{aligned} {\text {RMSE}}=\sqrt{\text {MSE}} \end{aligned}$$
  3. (iii)
    $$\begin{aligned} R^{2}=1-\frac{{\text {SSE}}}{{\text {SST}}} \end{aligned}$$

    having \( {\text {SSE}}=\sum \nolimits _{i=1}^n {(y_{i} -\hat{{y}}_{i} )^{2}} \); \( {\text {SST}}=\sum \nolimits _{i=1}^n {(y_{i} -\bar{{y}}_{i} )^{2}} \),

  4. (iv)
    $$\begin{aligned} {\text {MAE}}=\frac{1}{N}\left( {\sum \limits _{i=1}^N {\vert y_{i} -\hat{{y}}_{i} \vert } } \right) , \end{aligned}$$

    having \( y_{i} \) is the actual datavalues, \( \hat{{y}}_{i} \) is the predicted assessment, n is the number of days in prediction.

5 Discussion of results

The results for the confirmed cases arising in USA and India have been simulated provided by designated sources. Figure 4 shows linear fitting of WNNFIL trained model for USA data set. Model trained values increase at \(60^{\mathrm{th}} \) day and goes upto \(80^{\mathrm{th}} \) day after which it almost remains stabilized for the rest of the days. Whereas linear regression plots a straight increasing line leaving out all the data values. Figure 5 depicts linear regression of WNNFIL–LSSVR trained model for USA data set. Model trained values remain stabilized till 60 days the increases and goes upto \(80^{\mathrm{th}} \) day after which it almost remains stabilized for the rest of the days. Whereas linear regression plots a straight increasing line leaving out all the data values, Fig. 6 illustrates linear regression of WNNFIL–MARS trained model for USA data set. Figure 7 shows that Lyapunov exponents for the confirmed cases in USA. Whenever Lyapunov exponents can be seen \(<0\), it showed that the nearby initial conditions (ICs) converging to one another leading towards small epsilon-errors decreasing with respect to time. For any of the Lyapunov characteristics \(>0\) (\( + \)ve), it showed infinitesimally nearby initial conditions(ICs) diverging from one another exponentially fast implying that the errors in initial conditions would rise with respect to time. As and when trajectories indicate diverging with respect to time being known as the sensitive dependence on ICs coined as Chaos. Figure 8 clearly tracks the wavelet decomposition into \( A_{3},\, {D}_{1},\,{D}_{2},\, {D}_{3}\) to feed as responses into trained values. It filtered the noise that extracted characteristics trained and tested to forecast the predictor variables. Figure 9 shows the actual values in comparison to trained forecasts obtained from wavelet decomposition values fed into WNNFIL trained model for the confirmed cases in USA. Similarly, Fig. 10 demonstrates the actual values in comparison to tested forecasts obtained from WNNFIL tested model for the confirmed cases. The predicted values based upon the confirmed cases data observed in comparison to the actual ones through WNNFIL–LSSVR using RBF kernel in Fig. 11. Similarly, estimated values upon the confirmed cases observed in comparison to the actual ones through WNNFIL–MARS model as can be seen in Fig. 12.

For India, confirmed cases have been analyzed. Figure 13 shows linear fitting of WNNFIL trained model for India confirmed data set. Model trained values remain almost in the neighbourhood of zero upto 70 days then increase for rest of the days. Whereas linear regression plots a straight increasing line leaving out almost all the data values. Figure 14 depicts linear regression of WNNFIL–LSSVR trained model for India. Model trained values remain almost stabilized upto 120 days then increases for the rest of the period. Whereas linear regression plots a straight increasing line leaving out almost all the data values. Figure 15 illustrates linear regression of WNNFIL–MARS trained model for India. Model trained values remain almost stabilized for almost 120 days then increases nearly exponentially. Figure 16 shows that Lyapunov exponents for the confirmed cases in India. Whenever Lyapunov exponents can be seen \(< 0\)(-ve), it showed that the nearby initial conditions(ICs) converging to one another leading towards small epsilon-errors decreasing with respect to time. For any of the Lyapunov characteristics \(> 0\) ( \( + \) ve), it showed infinitesimally nearby initial conditions(ICs) diverging from one another exponentially fast implying that the errors in initial conditions would rise with respect to time. As and when trajectories indicate diverging with respect to time being known as the sensitive dependence on ICs coined as Chaos. Thus, these values indicated towards the randomness in the system. Figure 17 clearly tracks the wavelet decomposition into \( A_{3} \) , D \( _{1} \) , D \( _{2} \) , D \( _{3} \) to feed as responses into trained values. It shows the decomposition of signal into approximations and details at various levels. It filtered the noise that extracted characteristics trained & tested to forecast the predictor variables. Figure 18 shows the actual values in comparison to trained forecasts obtained from wavelet decomposition values fed into WNNFIL trained model for the confirmed cases in USA. Similarly, Fig. 19 demonstrates the actual values in comparison to tested forecasts obtained from WNNFIL tested model for the confirmed cases. The two trajectories are entirely different as the testing under this model is unable to capture the peaks of future forecasts. The predicted values of the confirmed cases data observed in comparison to the actual ones through WNNFIL–LSSVR using RBF kernel in Fig. 20. Similarly, estimated values up to of the confirmed cases observed in comparison to the actual ones through WNNFIL–MARS intelligent hybrid model in Fig. 21. Table 1 tabulates the statistical characteristics, such as LCEs, Hurst Exponents (HE), R/S value, entropy, fractal dimension (FD), predictability index (PI) and the behaviour of two data sets according to PI. The behaviour observed as anti-persistence based on all the statistical parameters. Determination of these characteristics for non-linear systems involves numerically integrating underlying differential equalities over the variation in parameters and initial conditions. Table 2 shows performance measures that are RMSE, MSE, MAE, \(\hbox {R}^{2} \) for training data models which are WNNFIL, WNNFIL–LSSVR, WNNFIL–MARS intelligent hybrid for confirmed cases of India and USA. WNNFIL–MARS model performs best across all errors and fits best in case of India, i.e., 0.987 and comparable to WNNFIL–LSSVR in case of USA, i.e., 0.995. Table 3 encapsulates various studies carried out towards COVID-19 spread.

Table 1 Various statistical characteristics for the two data types
Table 2 Performance measures of training models of India & USA daily Confirmed cases data
Table 3 Different studies carried out towards COVID-19 spread

6 Conclusions

The study aims to model the features of COVID-19 that can be identified to reduce its spread and develop design strategies such as strictly following social distancing, sanitizing etc. Thus, it is necessary to study ongoing scenario of the pandemic. Modelling outcomes provide an insight on the extent of interventions required to circumvent or to include so as to predict and estimate future trends. In addition, this study concludes that in both the number of confirmed people with corona have seen a gradual dip which indicates contrary to the perception of pandemic causing havoc in one of the most developed countries. Performance error measures of training models have been simulated through Lyapunov, WNNFIL, WNNFIL–LSSVR, WNNFIL–MARS, Hurst and Entropy for both India and USA. For India, WNNFIL–MARS perform better compared to WNNFIL–LSSVR. This can be validated through RMSE, MSE, MAE and \({R}^{2} \) values. Similarly, for USA, WNNFIL–MARS have lower error values as compared to WNNFIL–LSSVR. The goodness of fit, \({R}^{2} \) value has been achieved to be closest to 1, i.e., 0.999 from WNNFIL–MARS for both the countries’ data. This indicates that wavelet neuronal multivariate adaptive regression splines training-learned model fits best among all other prediction models in this study. In the present-day scenario, a rise leading to a second wave of corona-affected confirmed cases in the month of November have been observed in both the countries. This indicates we need to follow extra-precautious behaviour. Thus, the social distancing in the infected parts of different cities and countries might show substantial improvement in real and as well as analytical front. In addition, it may aid to employ similar control strategies and various other measures taken by authorities and citizens against this pandemic. Furthermore, this study helped to understand the rise in transmission and spread through surface contacts or humans coming in contact with each other.