Back to Search Start Over

Pitfall of Optimism: Distributional Reinforcement Learning by Randomizing Risk Criterion

Authors :
Cho, Taehyun
Han, Seungyub
Lee, Heesoo
Lee, Kyungjae
Lee, Jungwoo
Publication Year :
2023

Abstract

Distributional reinforcement learning algorithms have attempted to utilize estimated uncertainty for exploration, such as optimism in the face of uncertainty. However, using the estimated variance for optimistic exploration may cause biased data collection and hinder convergence or performance. In this paper, we present a novel distributional reinforcement learning algorithm that selects actions by randomizing risk criterion to avoid one-sided tendency on risk. We provide a perturbed distributional Bellman optimality operator by distorting the risk measure and prove the convergence and optimality of the proposed method with the weaker contraction property. Our theoretical results support that the proposed method does not fall into biased exploration and is guaranteed to converge to an optimal return. Finally, we empirically show that our method outperforms other existing distribution-based algorithms in various environments including Atari 55 games.<br />Comment: NeurIPS 2023

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2310.16546
Document Type :
Working Paper