Abstract
For time series forecasting which have data sets coming from an unstable and nonlinear system such as the stock market. Support Vector Regression (SVR) appears to be an efficient tool which has been widely used in recent years. It is also reported to have a higher accuracy and generalization ability than other traditional methods. The SVR method deals with the nonlinear problem by mapping the input feature space into a high dimensional space so that it becomes a linear problem. Kernel function is one of the crucial components in SVR algorithm as it is used to calculate the inner product between vectors in the mapped high dimensional space. The kernel function of Radial Basis Function (RBF), which is based on the Euclidean distance, is the most commonly used kernel function in SVR. However, the SVR algorithm may neglect the effect of correlation among the features when processing the training data in time series forecasting problems due to the limitation of Euclidean distance. In this chapter, a Mabalanobis distance RBF kernel is introduced. It is well known that when we need to calculate similarity between two vectors (samples), the use of Mahalanobis distance can take into account the correlation among the features. Thus, the SVR with Mahalanobis distance kernel function may follow the behavior of the data sets better so that it can give more accurate result. From the comparative investigation, we find that in some circumstances, the Mabalanobis distance RBF kernel based SVR can outperform the Euclidean distance based SVR.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
Keywords
- Support Vector Machine
- Kernel Function
- Radial Basis Function
- Support Vector Regression
- Mahalanobis Distance
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Kim, K.: Financial time series forecasting using support vector machines. Neurocomputing 55(1-2), 307–319 (2003)
Khemchandani, R., Chandra, S.: Regularized least squares fuzzy support vector regression for financial time series forecasting. Exp. Sys. With Appl. 36(1), 132–138 (2009)
Huang, W., Nakamori, Y., Wang, S.: Forecasting stock market movement direction with support vector machine. Comput Operat. Res. 32, 2513–2522 (2005)
Radhika, Y., Shashi, M.: Atmospheric temperature prediction using support vector machines. Int. J. Comput. Theor. and Eng. 1(1), 1793–8201 (2009)
Liu, B., Su, H., Huang, W., Chu, J.: Temperature prediction control based on least squares support vector machines. J. Contr. Theor. and Appl. 2(4), 365–370 (2004)
Bhasin, M., Raghava, G.: Prediction of CTL epitopes using QM, SVM and ANN techniques. Vaccine 22(23-24), 3195–3204 (2004)
Hossain, A., Zaman, F., Nasser, M., Islam, M.: Comparison of GARCH, Neural Network and Support vector machine in financial time series prediction. Pat. Rec. and Mach. Intel., 597–602 (2009)
Vapnik, V.N.: The Nature of Statistical Learning Theory. Springer, New York (1995)
Smola, A.J., Schlkopf, B.: A tutorial on support vector regression. Stat. and Comput. 14(3), 199–222 (2004)
Vapnik, V., Golowich, S.E., Smola, A.J.: Support vector method for function approximation, regression estimation, and signal processing. In: Michael, C., Michael, I., Thomas, P. (eds.) Advances in Neural Information Processing Systems 9, pp. 281–287. MIT Press (1996)
Cherkassky, V., Ma, Y.: Practical selection of SVM parameters and noise estimation for SVM regression. Neur. Netw. 17(1), 113–126 (2004)
Tsang, I.W., Kwok, J.T.: Learning the Kernel in Mahalanobis One-Class Support Vector Machines. In: The 2006 IEEE International Joint Conference on Neural Network Proceedings, Vancouver, BC, pp. 1169–1175 (2006)
Zhang, Y., Xie, F., Huang, D., Ji, M.: Support vector classifier based on fuzzy c-means and Mahalanobis distance. J. of Intel. Inf. Sys. 35, 333–345 (2009)
Wang, D., Yeung, D.S., Tsang, E.C.C.: Weighted mahalanobis distance kernels for support vector machines. IEEE Transactions on Neural Networks 18, 1453–1462 (2007)
Mahalanobis, P.C.: On the generalized distance in statistics. In: Proceedings National Institute of Science, India, pp. 49–55 (1936)
Bellman, R., Kalaba, R.: On adaptive control processes. IEEE Trans. on Auto. Contr. 4(2), 1–9 (1959)
Makridakis, S., Wheelwright, S.C., Hyndman, R.J.: Forecasting methods and applications. Wiley-India (2008)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Liu, J.N.K., Hu, Yx. (2013). Support Vector Regression with Kernel Mahalanobis Measure for Financial Forecast. In: Pedrycz, W., Chen, SM. (eds) Time Series Analysis, Modeling and Applications. Intelligent Systems Reference Library, vol 47. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-33439-9_10
Download citation
DOI: https://doi.org/10.1007/978-3-642-33439-9_10
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-33438-2
Online ISBN: 978-3-642-33439-9
eBook Packages: EngineeringEngineering (R0)