Back to Search
Start Over
How large should ensembles of classifiers be?
- Source :
- Biblos-e Archivo. Repositorio Institucional de la UAM, instname
- Publication Year :
- 2013
- Publisher :
- Elsevier BV, 2013.
-
Abstract
- This is the author’s version of a work that was accepted for publication in Pattern Recognition. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published in Pattern Recognition 46.5 (2013): 1323 – 1336, DOI: 10.1016/j.patcog.2012.10.021<br />We propose to determine the size of a parallel ensemble by estimating the minimum number of classifiers that are required to obtain stable aggregate predictions. Assuming that majority voting is used, a statistical description of the convergence of the ensemble prediction to its asymptotic (infinite size) limit is given. The analysis of the voting process shows that for most test instances the ensemble prediction stabilizes after only a few classifiers are polled. By contrast, a small but non-negligible fraction of these instances require large numbers of classifier queries to reach stable predictions. Specifically, the fraction of instances whose stable predictions require more than T classifiers for T ≫ 1 has a universal form and is proportional to T−1/2. The ensemble size is determined as the minimum number of classifiers that are needed to estimate the infinite ensemble prediction at an average confidence level , close to one. This approach differs from previous proposals, which are based on determining the size for which the prediction error (not the predictions themselves) stabilizes. In particular, it does not require estimates of the generalization performance of the ensemble, which can be unreliable. It has general validity because it is based solely on the statistical description of the convergence of majority voting to its asymptotic limit. Extensive experiments using representative parallel ensembles (bagging and random forest) illustrate the application of the proposed framework in a wide range of classification problems. These experiments show that the optimal ensemble size is very sensitive to the particular classification problem considered.<br />The authors acknowledge financial support from the Spanish Dirección General de Investigación, project TIN2010-21575-C02-02.
- Subjects :
- Majority rule
media_common.quotation_subject
Ensemble size
Asymptotic ensemble prediction
Ensembles of classifiers
Artificial Intelligence
Bagging
Ensemble learning
Voting
media_common
Mathematics
Informática
Ensemble forecasting
business.industry
Pattern recognition
Random forest
Random subspace method
ComputingMethodologies_PATTERNRECOGNITION
Signal Processing
Computer Vision and Pattern Recognition
Artificial intelligence
business
Algorithm
Software
Cascading classifiers
Subjects
Details
- ISSN :
- 00313203
- Volume :
- 46
- Database :
- OpenAIRE
- Journal :
- Pattern Recognition
- Accession number :
- edsair.doi.dedup.....30e773f08bd2666e238be9777dc58793
- Full Text :
- https://doi.org/10.1016/j.patcog.2012.10.021