Back to Search Start Over

Pessimism for Offline Linear Contextual Bandits using $\ell_p$ Confidence Sets

Authors :
Li, Gene
Ma, Cong
Srebro, Nathan
Publication Year :
2022

Abstract

We present a family $\{\hat{\pi}\}_{p\ge 1}$ of pessimistic learning rules for offline learning of linear contextual bandits, relying on confidence sets with respect to different $\ell_p$ norms, where $\hat{\pi}_2$ corresponds to Bellman-consistent pessimism (BCP), while $\hat{\pi}_\infty$ is a novel generalization of lower confidence bound (LCB) to the linear setting. We show that the novel $\hat{\pi}_\infty$ learning rule is, in a sense, adaptively optimal, as it achieves the minimax performance (up to log factors) against all $\ell_q$-constrained problems, and as such it strictly dominates all other predictors in the family, including $\hat{\pi}_2$.<br />Comment: Accepted to NeurIPS 2022

Details

Language :
English
Database :
OpenAIRE
Accession number :
edsair.doi.dedup.....dcea93669fb13908082cd4cb5e2702c2