Abstract
The k-Nearest-Neighbours (kNN) is a simple but effective method for classification. The major drawbacks with respect to kNN are (1) its low efficiency – being a lazy learning method prohibits it in many applications such as dynamic web mining for a large repository, and (2) its dependency on the selection of a “good value” for k. In this paper, we propose a novel kNN type method for classification that is aimed at overcoming these shortcomings. Our method constructs a kNN model for the data, which replaces the data to serve as the basis of classification. The value of k is automatically determined, is varied for different data, and is optimal in terms of classification accuracy. The construction of the model reduces the dependency on k and makes classification faster. Experiments were carried out on some public datasets collected from the UCI machine learning repository in order to test our method. The experimental results show that the kNN based model compares well with C5.0 and kNN in terms of classification accuracy, but is more efficient than the standard kNN.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Hand, D., Mannila, H., Smyth, P.: Principles of Data Mining. MIT Press, Cambridge (2001)
Wang, H.: Nearest Neighbours without k: A Classification Formalism based on Probability, technical report, Faculty of Informatics, University of Ulster, N.Ireland, UK (2002)
Sebastiani, F.: Machine Learning in Automated Text Categorization. ACM Computing Surveys 34(1), 1–47 (2002)
Wang, H., Duntsch, I., Bell, D.: Data Reduction Based on Hyper Relations. In: Proceedings of KDD 1998, New York, pp. 349–353 (1998)
Hart, P.: The Condensed Nearest Neighbour Rule. IEEE Transactions on Information Theory 14, 515–516 (1968)
Gates, G.: The Reduced Nearest Neighbour Rule. IEEE Transactions on Information Theory 18, 431–433 (1972)
Alpaydin, E.: Voting Over Multiple Condensed Nearest Neoghbors. Artificial Intelligence Review 11, 115–132 (1997); © Kluwer Academic Publishers (1997)
Kubat, M., Jr., M.: Voting Nearest-Neighbour Subclassifiers. In: Proceedings of the 17th International Conference on Machine Learning, ICML 2000, pp. 503–510, Stanford, CA, June 29-July 2 (2000)
Wilson, D.R., Martinez, T.R.: Reduction Techniques for Exemplar-Based Learning Algorithms. Machine learning 38(3), 257–286 (2000)
Mitchell, T.: Machine Learning. MITPress/McGraw-Hill (1997)
Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, UK (1995)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Guo, G., Wang, H., Bell, D., Bi, Y., Greer, K. (2003). KNN Model-Based Approach in Classification. In: Meersman, R., Tari, Z., Schmidt, D.C. (eds) On The Move to Meaningful Internet Systems 2003: CoopIS, DOA, and ODBASE. OTM 2003. Lecture Notes in Computer Science, vol 2888. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-39964-3_62
Download citation
DOI: https://doi.org/10.1007/978-3-540-39964-3_62
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-20498-5
Online ISBN: 978-3-540-39964-3
eBook Packages: Springer Book Archive