Abstract
Re-Sampling methods are some of the different types of approaches proposed to deal with the class-imbalance problem. Although such approaches are very simple, tuning them most effectively is not an easy task. In particular, it is unclear whether oversampling is more effective than undersampling and which oversampling or undersampling rate should be used. This paper presents an experimental study of these questions and concludes that combining different expressions of the resampling approach in a mixture of experts framework is an effective solution to the tuning problem. The proposed combination scheme is evaluated on a subset of the REUTERS-21578 text collection (the 10 top categories) and is shown to be very effective when the data is drastically imbalanced.
We would like to thank Rob Holte and Chris Drummond for their useful comments. This research was funded, in part, by an NSERC grant. The work conducted in this paper was conducted at Dalhousie University.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Domingos, Pedro (1999): Metacost: A general method for making classifiers cost sensitive, Proceedings of the Fifth International Conference on Knowledge Discovery and Data Mining, 155–164.
Estabrooks, Andrew (2000): A Combination Scheme for Inductive Learning from Imbalanced Data Sets, MCS Thesis, Faculty of Computer Science, Dalhousie University.
Hansen, L. K. and Salamon, P. (1990): Neural Network Ensembles, IEEE Transactions on Pattern Analysis and Machine Intelligence, 12(10), 993–1001.
Japkowicz, Nathalie (2000): The Class Imbalance Problem: Significance and Strategies, Proceedings of the 2000 International Conference on Artificial Intelligence (IC-AI’2000), 111–117.
Kubat, Miroslav and Matwin, Stan (1997): Addressing the Curse of Imbalanced Data Sets: One-Sided Sampling, Proceedings of the Fourteenth International Conference on Machine Learning, 179–186.
Lewis, D. and Gale, W. (1994): Training Text Classifiers by Uncertainty Sampling, Proceedings of the Seventh Annual International ACM SIGIR Conference on Research and Development in Information Retrieval.
Shimshoni, Y. and Intrator, N. (1998): Classifying Seismic Signals by Integrating Ensembles of Neural Networks, IEEE Transactions On Signal Processing, Special issue on NN.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Estabrooks, A., Japkowicz, N. (2001). A Mixture-of-Experts Framework for Learning from Imbalanced Data Sets. In: Hoffmann, F., Hand, D.J., Adams, N., Fisher, D., Guimaraes, G. (eds) Advances in Intelligent Data Analysis. IDA 2001. Lecture Notes in Computer Science, vol 2189. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44816-0_4
Download citation
DOI: https://doi.org/10.1007/3-540-44816-0_4
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-42581-6
Online ISBN: 978-3-540-44816-7
eBook Packages: Springer Book Archive