Back to Search Start Over

How large should ensembles of classifiers be?

Authors :
Daniel Hernández-Lobato
Alberto Suárez
Gonzalo Martínez-Muñoz
UAM. Departamento de Ingeniería Informática
Aprendizaje Automático (ING EPS-001)
Source :
Biblos-e Archivo. Repositorio Institucional de la UAM, instname
Publication Year :
2013
Publisher :
Elsevier BV, 2013.

Abstract

This is the author’s version of a work that was accepted for publication in Pattern Recognition. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published in Pattern Recognition 46.5 (2013): 1323 – 1336, DOI: 10.1016/j.patcog.2012.10.021<br />We propose to determine the size of a parallel ensemble by estimating the minimum number of classifiers that are required to obtain stable aggregate predictions. Assuming that majority voting is used, a statistical description of the convergence of the ensemble prediction to its asymptotic (infinite size) limit is given. The analysis of the voting process shows that for most test instances the ensemble prediction stabilizes after only a few classifiers are polled. By contrast, a small but non-negligible fraction of these instances require large numbers of classifier queries to reach stable predictions. Specifically, the fraction of instances whose stable predictions require more than T classifiers for T ≫ 1 has a universal form and is proportional to T−1/2. The ensemble size is determined as the minimum number of classifiers that are needed to estimate the infinite ensemble prediction at an average confidence level , close to one. This approach differs from previous proposals, which are based on determining the size for which the prediction error (not the predictions themselves) stabilizes. In particular, it does not require estimates of the generalization performance of the ensemble, which can be unreliable. It has general validity because it is based solely on the statistical description of the convergence of majority voting to its asymptotic limit. Extensive experiments using representative parallel ensembles (bagging and random forest) illustrate the application of the proposed framework in a wide range of classification problems. These experiments show that the optimal ensemble size is very sensitive to the particular classification problem considered.<br />The authors acknowledge financial support from the Spanish Dirección General de Investigación, project TIN2010-21575-C02-02.

Details

ISSN :
00313203
Volume :
46
Database :
OpenAIRE
Journal :
Pattern Recognition
Accession number :
edsair.doi.dedup.....30e773f08bd2666e238be9777dc58793
Full Text :
https://doi.org/10.1016/j.patcog.2012.10.021