Keywords

1 Introduction

Time series data is very vital for many applications such as economics, medicine, education, social sciences, epidemiology, weather forecasting, physical sciences etc. to derive meaningful insights at different points in time. Conventional statistics methods have several limitations to deal with time series data so specialized methods known as time series analysis requires predominantly in such cases. The simplest and most popular method is linear least square method. Least square method gives the trend line to best fit to a time series data. It exhibits several advantages:

  • It is very simple method to understand and derive the prediction

  • It is to be applicable for all most all applications

  • It gives maximum likelihood solutions if correlate with Markov Conditions.

However, it suffers from several critical limitations:

  • Sensitive towards outliers.

  • Data needs to be normally distributed for better results.

  • It exhibits tendency of outfit data.

Quantile Regressing method by utilizing support vector machine approach is an idyllic approach to deal with the limitations of least square regression methods. It has advantages over least square regression. Table 1 describes the comparison between least square and quantile regression.

Table 1. Comparison of least and quantile regression

Support vector machine in correlation with quantile regression may produce excellent outcomes for time series analysis. The support vector machine has an ability to solve nonlinear regression estimate problems so it is the prominent candidate for time series data analysis. One more significant feature of SVM is that the learning here is analogous to resolve a problem of linear quadratic optimization. Thus, unlike the other traditional stochastic or neural network methods, the solution obtained by applying the SVM method is always unique and globally optimal.

The Sect. 2 of paper will deal with related work in this field. Section 3 will describe QRSVM model in details. Section 4 will discuss about Experiments and Results of the model. At Last, Paper provides the conclusions of research work carried out.

2 Related Work

Statistical Methods predominantly used for time series data analysis. Autoregressive Integrated Moving Average (ARIMA) model is the most prevalent and commonly used for time series data analysis [6]. Notwithstanding, these sort of models depend on the hypothesis that take into an account that time series must be linear and follows a normal distribution of the data. C. Hamzacebi in 2008 [1] proposed a distinction of ARIMA model called as Seasonal ARIMA (SARIMA). The prototypical produced good results for seasonal time series data, however it required to undertake linear form of associated time series data. The limitations of the linear models could be overcome by non-linear stochastic models [5, 19]. However, the implementation of these kind of models is very complex.

Neural Network based time series models have grown as of late and pulled in expanding considerations [8, 9]. The astounding element of ANNs is their inherent capability of non-linear modeling with no presupposition about the statistical distribution monitored by the annotations. The incredible highlights about ANN based models are self-versatile in nature [28]. There is assortment of ANN models exist in the literature. The Multi-Layer Perceptron (MLP) is the most famous and basic model dependent on ANN [2, 4, 13, 22]. MLPs contain different layers of computational components, unified in a feed-forward way [18].MLPs utilize a variety of learning techniques, the conspicuous is back-propagation [16, 20, 29] where the output esteems are related with the exact response to compute the value of some foreordained error-function. The error is then served back through the network. Utilizing this data, the algorithm controls the degree of each linkage so as to decrease the estimation of the error function by some insignificant quantity. An overall strategy for non-linear optimization called gradient descent [21, 23] is applied to regulate the degrees. Time Lagged Neural Network (TLNN) is another variation of Feed Forward way [15, 26]. In TLNN, the input nodes are the time series values at some specific lags. Likewise, there is a constant input term, which may be expediently taken as 1 and this is linked to every neuron in the hidden and output layer. The presentation of this constant input unit circumvents the need of separately introducing a bias term. In 2007, Pang et al. [17] introduced one model dependent on neural network and efficaciously applied to the simulation in the rainfall. Li et al. [14], In 2008, presented hybrid model based on AR * and generalized regression neural network model (GRNN) and that gave respectable results in the setting to the time series data. Chen and Chang in 2009 [3] came out with an Evolutionary Artificial Neural Network model (EANN) to build automatically the architecture and the connections of the weights of the neural network. Khashei and Bijari in 2010 [11] introduced a new hybrid ANN model, utilizing an ARIMA model to discover predictions more precise than the model of neural networks. Wu and Shahidehpour [27] proposed a fusion model based on an adaptive Wavelet Neural Network (AWNN) and time series models, such as the ARMAX and GARCH, to predict the day by day estimation of electricity in the market. In [7] researchers proposed a regression neural network model to anticipate widespread time series, which is a fusion of diverse algorithms for machine learning. Artificial Neural Network based algorithms are overwhelming for time series data analysis however they show various constraints such as: appropriate network structure is attained through trial and error, sometimes mysterious performance of the network, usually require more data to train the model fittingly, computationally complex and affluent. Support Vector Machine [12, 24, 25] is the vigorous machine learning technique for the pattern generation and classification.

The proposed model will use SVM as it isn’t just intended for decent classification yet additionally expected for an improved speculation of the training data. Solutions obtained by SVM is always unique as it depends on linearly constrained quadratic optimization. The model will use the fusion methodology of SVM and Quantile Regression [10]. Quantile Regression methodology permits for comprehension relationships between variables outside of the average of the data, making it valuable in understanding outcomes that are non-normally dispersed and that have nonlinear relationships with predictor variables.

3 Quantile Regression Support Vector Machine (QRSVM) Model

The least square regression model is representing by the Eq. (1).

$$ Y = a + b\;X +\upvarepsilon $$
(1)

Where Y is Dependent Variable whose value is going to be predicted

a is the intercept of Y

b is the slope of line

\(\upvarepsilon \) represents an error and s identically, independently, and normally distributed with mean zero and unknown variance σ2.

Least square regression model attempt to define conditional distribution by utilizing the average of a distribution. Another thing is, it assumes that the error term is same across all values of X in which conditional variable (Y/X) to be assumed a constant variance σ2. When this assumption fail, we must change the LSR algorithm to accommodate conditional mean and scale. The new equation based on conditional scale is:

$$Y=a+b \;X+{e}^{r}\upvarepsilon $$
(2)

Where r is the unknown parameter

$$ {\text{Var}}\,{\text{(Y/X)}} =\upsigma ^{2} e^{r} $$
(3)

In this also, conditional scale for dependent variable y is not vary with independent variable X. In order to realize covariate properties in context to dependent variable Quantile Regressing concept is required.

$$Y={a}^{(p)}+ {b}^{(p)}X+ {\upvarepsilon}^{(p)}$$
(4)

Where p is the probability and it ranges between 0 and 1.

We specify the pth conditional quantile given X with

$${Q}^{\left(p\right) }\left(\frac{Y}{X}\right)= {a}^{(p)}+ {b}^{(p)}X$$
(5)

Least square regression having only one conditional mean while Quantile Regression contains numerous conditional quantiles. In the nonlinear quantile regression, the quantile of the dependent variable Y for a given independent attribute X is assumed to be nonlinearly related to the input vector Xi ∈ Rd and represented by nonlinear mapping function ϕ(…). The new version related to nonlinearity characteristic of quantile function is represented as:

$$ QX = W\theta \,\upphi ({\text{X}}) $$
(6)

Where θ ∈ (0, 1),

\({W}_{\theta }\) is θth regression quantile.

Absolute deviation loss will occur in quantile regression so SVM with quantile regression plays a vital role. The equation of quantile regression with SVM is represented as:

$$ Minimize12\,\left\| {w\uptheta } \right\|2 + C\,i = 1np\theta (Y - W\theta \,\upphi (X)) $$
(7)

for any θ ∈ (0, 1)

Equation (7) is considered as QRSVM Model.

4 Experiments and Results

Experiments of proposed study is carried out by considering weather data of Anand District of Gujarat State, India. The sample data is depicted in the Table 1.

Table 2. Sample weather data of Anand district

Experiment simulation is carried out using R programming language (Figs. 1 and 2).

Fig. 1.
figure 1

Scatter plot temperature vs. rainfall

Fig. 2.
figure 2

Box plot for rainfall data

The box plot of rainfall data indicates that several values are outliers. For the accurate prediction of the time series data, outlier values also play an important role.

According to least square regression model, residuals values and Coefficient statistics, are depicted in Table 2 and Table 3 respectively. They having one value based on central tendency value.

Table 3. Residual values of least square regression
Table 4. Coefficient values of least square regression

Quantile Regression SVM model generate coefficient based on quantile value. For the similar data, the coefficient of the QRSVM Model is depicted in Table 4 (Table 5, Figs. 3 and 4).

Table 5. Coefficient values of QRSVM model
Fig. 3.
figure 3

Scatter plot quantile vs. intercept value in QRSVM

Fig. 4.
figure 4

Scatter plot quantile vs. X in QRSVM

From the results we can see that Least Square regression distributes Intercept and X values on central tendency where as QRSVM model distributes them with multiple values based on the values of percentile so we can understand and explore the insights with multiple dimensions. The other thing is if outlier exist in the data then central tendency value might be compromised whereas this situation will not affect in QRSVM Model.

5 Conclusions

In this proposed work, it is concluded that Least Square Regression Model exhibits several limitations such as it attempts to define conditional distribution by utilizing only the average of a distribution. Another thing is, it assumes that the error term is same across all values of X in which conditional variable (Y/X) to be assumed a constant variance σ2. In order to realize covariate properties in context to dependent variable, Quantile Regressing concept is required. Based on the experiments of time series data of weather, the paper concluded that QRSVM model distributes Intercept and X values in multiple values in order to understand and interpret them effectively. Paper also concluded that the results of LSR model might be compromised to deal to with outliers whereas this situation does not exist in QRSVM model.