A Study of Combined Splitting Rules in Regression Trees

  • Published : 2002.04.30

Abstract

Regression trees, a technique in data mining, are constructed by splitting function-a independent variable and its threshold. Lee (2002) considered one-sided purity (OSP) and one-sided extreme (OSE) splitting criteria for finding a interesting node as early as possible. But these methods cannot be crossed each other in the same tree. They are just concentrated on OSP or OSE separately in advance. In this paper, a new splitting method, which is the combination and extension of OSP and OSE, is proposed. By these combined criteria, we can select the nodes by considering both pure and extreme in the same tree. These criteria are not the generalized one of the previous criteria but another option depending on the circumstance.

Keywords

References

  1. Machine Learning v.24 Bagging predictors Breiman,L.
  2. Classification and Regression Trees Breiman, L.;Friedman, J. H.';Olshen, R. A.;Stone, C. J
  3. Journal of Computer and System Science v.55 A Decision-Theoretic Generalization of On-Line Learning and Application to Boosting Freund, Y.;Schapire, R. E.
  4. Clustering Algorithms Hartigan, J. A.
  5. The Korean Communications in Statistics v.8 New Splitting Criteria for Classification Trees Lee, Y-S.
  6. The Korean Journal of Applied Statistics Interesting Node Finding Criteria for Regression Trees Lee, Y-S.
  7. Journal of the American Statistical Association v.58 Problems in the Analysis of Survey Data, and a Proposal Morgan, J. N.;Sonquist, J. A.
  8. C4.5: Programs for Machine Learning Quinlan, J. R.
  9. Model Search and Inference via Tibshirani, R.;Knight, K.