Skip to main content

A Study on End-Cut Preference in Least Squares Regression Trees

  • Conference paper
  • First Online:
Progress in Artificial Intelligence (EPIA 2001)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 2258))

Included in the following conference series:

Abstract

Regression trees are models developed to deal with multiple regression data analysis problems. These models fit constants to a set of axes-parallel partitions of the input space defined by the predictor variables. These partitions are described by a hierarchy of logical tests on the input variables of the problem. Several authors have remarked that the preference criteria used to select these tests have a clear preference for what is known as end-cut splits. These splits lead to branches with a few training cases, which is usually considered as counter-intuitive by the domain experts. In this paper we describe an empirical study of the effect of this end-cut preference on a large set of regression domains. The results of this study, carried out for the particular case of least squares regression trees, contradict the prior belief that these type of tests should be avoided. As a consequence of these results, we present a new method to handle these tests that we have empirically shown to have better predictive accuracy than the alternatives that are usually considered in tree-based models.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. J. Bradford and C. Broadley. The effect of instance-space partition on significance. Machine Learning, 42(3):269–286, 2001.

    Article  MATH  Google Scholar 

  2. L. Breiman, J. Friedman, R. Olshen, and C. Stone. Classification and Regression Trees. Statistics/Probability Series. Wadsworth & Brooks/Cole Advanced Books & Software, 1984.

    Google Scholar 

  3. J. Catlett. Megainduction: machine learning on very large databases. PhD thesis, Basser Department of Computer Science, University of Sidney, 1991.

    Google Scholar 

  4. T. Hastie and R. Tibshirani. Generalized Additive Models. Chapman & Hall, 1990.

    Google Scholar 

  5. J. Morgan and R. Messenger. Thaid: a sequential search program forthe analysis of nominal scale dependent variables. Technical report, Ann Arbor: Institute for Social Research, University of Michigan, 1973.

    Google Scholar 

  6. J. Morgan and J. Sonquist. Problems in the analysis of survey data, and a proposal. Journal of American Statistics Society, 58:415–434, 1963.

    MATH  Google Scholar 

  7. J. Quinlan. C4.5: programs for machine learning. Kluwer Academic Publishers, 1993.

    Google Scholar 

  8. L. Torgo. Inductive Learning of Tree-based Regression Models. PhD thesis, Faculty of Sciences, University of Porto, 1999.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Torgo, L. (2001). A Study on End-Cut Preference in Least Squares Regression Trees. In: Brazdil, P., Jorge, A. (eds) Progress in Artificial Intelligence. EPIA 2001. Lecture Notes in Computer Science(), vol 2258. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45329-6_14

Download citation

  • DOI: https://doi.org/10.1007/3-540-45329-6_14

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-43030-8

  • Online ISBN: 978-3-540-45329-1

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics