Back to Search Start Over

Policy Gradient Importance Sampling for Bayesian Inference.

Authors :
El-Laham, Yousef
Bugallo, Monica F.
Source :
IEEE Transactions on Signal Processing; 11/15/2021, p4245-4256, 12p
Publication Year :
2021

Abstract

In this paper, we propose a novel adaptive importance sampling (AIS) algorithm for probabilistic inference. The sampler learns a proposal distribution adaptation strategy by framing AIS as a reinforcement learning problem. Under this structure, the proposal distribution of the sampler is treated as an agent whose state is controlled using a parameterized policy. At each iteration of the algorithm, the agent earns a reward that is related to its contribution to the variance of the AIS estimator of the normalization constant of the target distribution. Policy gradient methods are employed to learn a locally optimal policy that maximizes the expected value of the sum of all rewards. Numerical simulations on two different examples demonstrate promising results for the future application of the proposed method to complex Bayesian models. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
1053587X
Database :
Complementary Index
Journal :
IEEE Transactions on Signal Processing
Publication Type :
Academic Journal
Accession number :
153880579
Full Text :
https://doi.org/10.1109/TSP.2021.3093792