Abstract
In the paper a new method for handling with missing features values in classification is presented. The presented idea is to form an ensemble of one-class classifiers trained on each feature, preselected group of features or to compute from features a dissimilarity representation. Thus when any feature values are missing for a data point to be labeled, the ensemble can still make a reasonable decision based on the remaining classifiers. With the comparison to standard algorithms that handle with the missing features problem it is possible to build an ensemble that can classify test objects with all possible occurrence of missing features without retrain a classifier for each combination of missing features. Additionally, to train such an ensemble a training set does not need to be uncorrupted. The performance of the proposed ensemble is compared with standard methods use with missing features values problem on several UCI datasets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Little, R.J.A., Rubin, D.B.: Statistical analysis with missing data, 2nd edn. Wiley-Interscience, Hoboken (2002) ISBN 0-471-18386-5
Chan, L.S., Dun, O.J.: Alternative approaches to missing values in discriminant analysis. J. Amer. Statist. Assoc. 71, 842–844 (1976)
Dixon, J.K.: Pattern recognition with partly missing data. IEEE Transactions on Sys., Man and Cyber., 617–621 (1979)
Morin, R.L., Raeside, D.E.: A reappraisal of distance-weighted k-nearest neighbor classification for pattern recognition with missing data. IEEE Trans. Syst. Man Cybern. 11, 241–243 (1981)
Little, R.J.A.: Consistent regression methods for discriminant analysis with incomplete data. J. Amer. Statist. Assoc. 73, 319–322 (1978)
Ghahramani, Z., Jordan, M.I.: Supervised learning from incomplete data via an em approach. In: NIPS (1994)
Tax, D.M.J.: One-class classification. PhD thesis, Delft University of Technology (2001)
Ahmad, S., Tresp, V.: Some solutions to the missing feature problem in vision. In: NIPS, pp. 393–400 (1993)
Duda, R.O., Hart, P.E., Stork, D.G.: Pattern classification, 2nd edn. Wiley Interscience, Hoboken (2001) ISBN: 0-471-05669-3
Tax, D.M.J., Duin, R.P.W.: Combining one-class classifiers. In: Kittler, J., Roli, F. (eds.) MCS 2001. LNCS, vol. 2096, pp. 299–308. Springer, Heidelberg (2001)
Pekalska, E., Duin, R.P.W.: Dissimilarity representations allow for building good classifiers. PR Letters 23, 943–956 (2002)
Duin, R.P.W.: On the choice of the smoothing parameters for parzen estimators of probability density functions. IEEE Transactions on Computers (1976)
Pekalska, E., Tax, D.M.J., Duin, R.P.W.: One-class lp classifiers for dissimilarity representations. In: NIPS, pp. 761–768 (2002)
Kittler, J., Hatef, M., Duin, R.P.W.: On combining classifiers. IEEE Transactions on PAMI 20 (1998)
Blake, C.L., Merz, C.J.: (UCI repository of machine learning databases)
Kittler, J.: Feature selection and extraction. In: Handbook of Pattern Recognition and Image Processing, pp. 59–83 (1996)
Ho, T.K.: Data complexity analysis for classifier combination. In: Kittler, J., Roli, F. (eds.) MCS 2001. LNCS, vol. 2096, pp. 53–67. Springer, Heidelberg (2001)
Raudys, S.: Multiple classification systems in the context of feature extraction and selection. In: MCS, pp. 27–41 (2002)
Littlestone, N., Warmuth, M.: Weighted majority algorithm. Information and Computation 108, 212–261 (1994)
Duin, R.P.W.: The combining classifier: to train or not to train. In: ICPR (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Juszczak, P., Duin, R.P.W. (2004). Combining One-Class Classifiers to Classify Missing Data. In: Roli, F., Kittler, J., Windeatt, T. (eds) Multiple Classifier Systems. MCS 2004. Lecture Notes in Computer Science, vol 3077. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-25966-4_9
Download citation
DOI: https://doi.org/10.1007/978-3-540-25966-4_9
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-22144-9
Online ISBN: 978-3-540-25966-4
eBook Packages: Springer Book Archive