Back to Search
Start Over
DECISION TREES DO NOT GENERALIZE TO NEW VARIATIONS.
- Source :
-
Computational Intelligence . Nov2010, Vol. 26 Issue 4, p449-467. 19p. 3 Diagrams. - Publication Year :
- 2010
-
Abstract
- The family of decision tree learning algorithms is among the most widespread and studied. Motivated by the desire to develop learning algorithms that can generalize when learning highly varying functions such as those presumably needed to achieve artificial intelligence, we study some theoretical limitations of decision trees. We demonstrate formally that they can be seriously hurt by the curse of dimensionality in a sense that is a bit different from other nonparametric statistical methods, but most importantly, that they cannot generalize to variations not seen in the training set. This is because a decision tree creates a partition of the input space and needs at least one example in each of the regions associated with a leaf to make a sensible prediction in that region. A better understanding of the fundamental reasons for this limitation suggests that one should use forests or even deeper architectures instead of trees, which provide a form of distributed representation and can generalize to variations not encountered in the training data. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 08247935
- Volume :
- 26
- Issue :
- 4
- Database :
- Academic Search Index
- Journal :
- Computational Intelligence
- Publication Type :
- Academic Journal
- Accession number :
- 55022427
- Full Text :
- https://doi.org/10.1111/j.1467-8640.2010.00366.x