Skip to main navigation Skip to search Skip to main content

Policy Gradient Importance Sampling for Bayesian Inference

  • Stony Brook University

Research output: Contribution to journalArticlepeer-review

11 Scopus citations

Abstract

In this paper, we propose a novel adaptive importance sampling (AIS) algorithm for probabilistic inference. The sampler learns a proposal distribution adaptation strategy by framing AIS as a reinforcement learning problem. Under this structure, the proposal distribution of the sampler is treated as an agent whose state is controlled using a parameterized policy. At each iteration of the algorithm, the agent earns a reward that is related to its contribution to the variance of the AIS estimator of the normalization constant of the target distribution. Policy gradient methods are employed to learn a locally optimal policy that maximizes the expected value of the sum of all rewards. Numerical simulations on two different examples demonstrate promising results for the future application of the proposed method to complex Bayesian models.

Original languageEnglish
Article number9468927
Pages (from-to)4245-4256
Number of pages12
JournalIEEE Transactions on Signal Processing
Volume69
DOIs
StatePublished - 2021

Keywords

  • Bayesian inference
  • Monte Carlo methods
  • importance sampling
  • reinforcement learning
  • stochastic optimization

Fingerprint

Dive into the research topics of 'Policy Gradient Importance Sampling for Bayesian Inference'. Together they form a unique fingerprint.

Cite this