Back to Search
Start Over
Cold Posteriors and Aleatoric Uncertainty
- Source :
- ICML workshop on Uncertainty and Robustness in Deep Learning (2020)
- Publication Year :
- 2020
-
Abstract
- Recent work has observed that one can outperform exact inference in Bayesian neural networks by tuning the "temperature" of the posterior on a validation set (the "cold posterior" effect). To help interpret this phenomenon, we argue that commonly used priors in Bayesian neural networks can significantly overestimate the aleatoric uncertainty in the labels on many classification datasets. This problem is particularly pronounced in academic benchmarks like MNIST or CIFAR, for which the quality of the labels is high. For the special case of Gaussian process regression, any positive temperature corresponds to a valid posterior under a modified prior, and tuning this temperature is directly analogous to empirical Bayes. On classification tasks, there is no direct equivalence between modifying the prior and tuning the temperature, however reducing the temperature can lead to models which better reflect our belief that one gains little information by relabeling existing examples in the training set. Therefore although cold posteriors do not always correspond to an exact inference procedure, we believe they may often better reflect our true prior beliefs.<br />Comment: 5 pages, 3 figures
- Subjects :
- Statistics - Machine Learning
Computer Science - Machine Learning
Subjects
Details
- Database :
- arXiv
- Journal :
- ICML workshop on Uncertainty and Robustness in Deep Learning (2020)
- Publication Type :
- Report
- Accession number :
- edsarx.2008.00029
- Document Type :
- Working Paper