Abstract
We propose an approach to embed time series data in a vector space based on the distances obtained from Dynamic Time Warping (DTW), and to classify them in the embedded space. Under the problem setting in which both labeled data and unlabeled data are given beforehand, we consider three embeddings, embedding in a Euclidean space by MDS, embedding in a Pseudo-Euclidean space, and embedding in a Euclidean space by the Laplacian eigenmap technique.
We have found through analysis and experiment that the embedding by the Laplacian eigenmap method leads to the best classification result. Furthermore, the proposed approach with Laplacian eigenmap embedding shows better performance than k-nearest neighbor method.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
Rabiner, L., Juang, B.: Fundamentals of Speech Recognition. Prentice Hall, Englewood Cliffs (1993)
Torgerson, W.S.: Theory and methods of scaling. J. Wiley & Sons, Chichester (1958)
Goldfarb, L.: A new approach to pattern recognition. In: Progress in Pattern Recognition, vol. 2, pp. 241–402. Elsevier Science Publishers B.V., Amsterdam (1985)
Graepel, T., Herbrich, R., Bollmann-Sdorra, P., Obermayer, K.: Classification on pairwise proximity data. In: Advances in Neural Information Processing 11, pp. 438–444 (1999)
Pekalska, E., Paclik, P., Duin, R.P.W.: A generalized kernel approach to dissimilarity-based classification. Journal of Machine Learning Research, Special Issue on Kernel Methods 2(2), 175–211 (2002)
Belkin, M., Niyogi, P.: Laplacian eigenmaps and spectral techniques for embedding and clustering. In: Advances in Neural Information Processing Systems 14, pp. 585–591 (2002)
Belkin, M., Niyogi, P.: Using manifold structure for partially labeled classification. In: Advances in Neural Information Processing Systems 15, pp. 929–936 (2003)
Schölkopf, B., Smola, A.J., Müller, K.R.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation 10, 1299–1319 (1998)
Shimodaira, H., Noma, K., Nakai, M., Sagayama, S.: Dynamic time-alignment kernel in support vector machine. In: Neural Information Processing Systems 14, pp. 921–928 (2002)
Bahlmann, C., Haasdonk, B., Burkhardt, H.: On-line handwriting recognition with support vector machines-a kernel approach. In: Proc. 8th Int. W/S on Frontiers in Handwriting Recognition, pp. 49–54 (2002)
Ham, J., Lee, D.D., Mika, S., Schölkopf, B.: A kernel view of the dimensionality reduction of manifolds. TR-110, Max-Planck-Institut fur biologische Kybernetik, Tubingen (2003)
Corres, C., Vapnik, V.: Support vector networks. Machine Learning 20, 273–297 (1995)
Hettich, S., Bay, S.D.: UCI Repository of KDD Databases (1999), http://kdd.ics.uci.edu/
Seeger, M.: Learning with labeled and unlabeled data. Technical report, Institute for Adaptive and Neural Computation, University of Edinburgh (2001)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hayashi, A., Mizuhara, Y., Suematsu, N. (2005). Embedding Time Series Data for Classification. In: Perner, P., Imiya, A. (eds) Machine Learning and Data Mining in Pattern Recognition. MLDM 2005. Lecture Notes in Computer Science(), vol 3587. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11510888_35
Download citation
DOI: https://doi.org/10.1007/11510888_35
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-26923-6
Online ISBN: 978-3-540-31891-0
eBook Packages: Computer ScienceComputer Science (R0)