Back to Search Start Over

DECISION TREES DO NOT GENERALIZE TO NEW VARIATIONS.

Authors :
BENGIO, YOSHUA
DELALLEAU, OLIVIER
SIMARD, CLARENCE
Source :
Computational Intelligence. Nov2010, Vol. 26 Issue 4, p449-467. 19p. 3 Diagrams.
Publication Year :
2010

Abstract

The family of decision tree learning algorithms is among the most widespread and studied. Motivated by the desire to develop learning algorithms that can generalize when learning highly varying functions such as those presumably needed to achieve artificial intelligence, we study some theoretical limitations of decision trees. We demonstrate formally that they can be seriously hurt by the curse of dimensionality in a sense that is a bit different from other nonparametric statistical methods, but most importantly, that they cannot generalize to variations not seen in the training set. This is because a decision tree creates a partition of the input space and needs at least one example in each of the regions associated with a leaf to make a sensible prediction in that region. A better understanding of the fundamental reasons for this limitation suggests that one should use forests or even deeper architectures instead of trees, which provide a form of distributed representation and can generalize to variations not encountered in the training data. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
08247935
Volume :
26
Issue :
4
Database :
Academic Search Index
Journal :
Computational Intelligence
Publication Type :
Academic Journal
Accession number :
55022427
Full Text :
https://doi.org/10.1111/j.1467-8640.2010.00366.x