Back to Search Start Over

NON-LOCAL PRIORS FOR HIGH-DIMENSIONAL ESTIMATION.

Authors :
Rossell D
Telesca D
Source :
Journal of the American Statistical Association [J Am Stat Assoc] 2017; Vol. 112 (517), pp. 254-265. Date of Electronic Publication: 2017 May 03.
Publication Year :
2017

Abstract

Jointly achieving parsimony and good predictive power in high dimensions is a main challenge in statistics. Non-local priors (NLPs) possess appealing properties for model choice, but their use for estimation has not been studied in detail. We show that for regular models NLP-based Bayesian model averaging (BMA) shrink spurious parameters either at fast polynomial or quasi-exponential rates as the sample size n increases, while non-spurious parameter estimates are not shrunk. We extend some results to linear models with dimension p growing with n . Coupled with our theoretical investigations, we outline the constructive representation of NLPs as mixtures of truncated distributions that enables simple posterior sampling and extending NLPs beyond previous proposals. Our results show notable high-dimensional estimation for linear models with p ≫ n at low computational cost. NLPs provided lower estimation error than benchmark and hyper-g priors, SCAD and LASSO in simulations, and in gene expression data achieved higher cross-validated R <superscript>2</superscript> with less predictors. Remarkably, these results were obtained without pre-screening variables. Our findings contribute to the debate of whether different priors should be used for estimation and model selection, showing that selection priors may actually be desirable for high-dimensional estimation.

Details

Language :
English
ISSN :
0162-1459
Volume :
112
Issue :
517
Database :
MEDLINE
Journal :
Journal of the American Statistical Association
Publication Type :
Academic Journal
Accession number :
29881129
Full Text :
https://doi.org/10.1080/01621459.2015.1130634