19 results on '"Annealed Importance sampling"'
Search Results
2. On the use of transport and optimal control methods for Monte Carlo simulation
- Author
-
Heng, Jeremy and Doucet, Arnaud
- Subjects
519.5 ,Transport theory--Statistical methods ,Optimal control ,Normalizing constants ,Sequential Monte Carlo ,Approximate dynamic programming ,Monte Carlo ,Mass transport ,Annealed importance sampling ,Reinforcement learning ,Markov chain Monte Carlo - Abstract
This thesis explores ideas from transport theory and optimal control to develop novel Monte Carlo methods to perform efficient statistical computation. The first project considers the problem of constructing a transport map between two given probability measures. In the Bayesian formalism, this approach is natural when one introduces a curve of probability measures connecting the prior to posterior by tempering the likelihood function. The main idea is to move samples from the prior using an ordinary differential equation (ODE), constructed by solving the Liouville partial differential equation (PDE) which governs the time evolution of measures along the curve. In this work, we first study the regularity solutions of Liouville equation should satisfy to guarantee validity of this construction. We place an emphasis on understanding these issues as it explains the difficulties associated with solutions that have been previously reported. After ensuring that the flow transport problem is well-defined, we give a constructive solution. However, this result is only formal as the representation is given in terms of integrals which are intractable. For computational tractability, we proposed a novel approximation of the PDE which yields an ODE whose drift depends on the full conditional distributions of the intermediate distributions. Even when the ODE is time-discretized and the full conditional distributions are approximated numerically, the resulting distribution of mapped samples can be evaluated and used as a proposal within Markov chain Monte Carlo and sequential Monte Carlo (SMC) schemes. We then illustrate experimentally that the resulting algorithm can outperform state-of-the-art SMC methods at a fixed computational complexity. The second project aims to exploit ideas from optimal control to design more efficient SMC methods. The key idea is to control the proposal distribution induced by a time-discretized Langevin dynamics so as to minimize the Kullback-Leibler divergence of the extended target distribution from the proposal. The optimal value functions of the resulting optimal control problem can then be approximated using algorithms developed in the approximate dynamic programming (ADP) literature. We introduce a novel iterative scheme to perform ADP, provide a theoretical analysis of the proposed algorithm and demonstrate that the latter can provide significant gains over state-of-the-art methods at a fixed computational complexity.
- Published
- 2016
3. On the Quantitative Analysis of Sparse RBMs
- Author
-
Zhang, Yanxia, Yang, Lu, Meng, Binghao, Cheng, Hong, Zhang, Yong, Wang, Qian, Zhu, Jiadan, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Chen, Enqing, editor, Gong, Yihong, editor, and Tie, Yun, editor
- Published
- 2016
- Full Text
- View/download PDF
4. Rate Distortion via Deep Learning.
- Author
-
Li, Qing and Chen, Yang
- Subjects
- *
DEEP learning , *VIDEO coding , *BOLTZMANN machine , *PARTITION functions , *SOURCE code - Abstract
We explore the connections between rate distortion/lossy source coding and deep learning models, the Restricted Boltzmann Machines (RBMs) and Deep Belief Networks (DBNs). We show that rate distortion is a function of the RBM log partition function and that RBM/DBN can be used to learn the rate distortion approaching posterior as in the Blahut-Arimoto algorithm. We propose an algorithm for lossy compressing of binary sources. The algorithm consists of two stages, a training stage that learns the posterior with training data of the same class as the source, and a compression/reproduction stage that is comprised of a lossless compression and a lossless reproduction. Theoretical results show that the proposed algorithm achieves the optimum rate distortion function for stationary ergodic sources asymptotically. Numerical experiments show that the proposed algorithm outperforms the reported best results. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
5. Iterative ensemble smoothers in the annealed importance sampling framework.
- Author
-
Stordal, Andreas S. and Elsheikh, Ahmed H.
- Subjects
- *
ITERATIVE methods (Mathematics) , *INVERSE problems , *RELIABILITY in engineering , *MONTE Carlo method , *APPROXIMATION theory - Abstract
Iterative ensemble techniques for solving inverse problems has recently gained a lot of interest in many geophysical communities. This popularity is attributed to the simplicity of implementation, general reliability and the ability to deal with the forward model as a black box without requiring the implementation of analytical gradients. Although several variants exist, we focus on the ensemble smoother with multiple data assimilation. This study highlights the similarity between the ensemble smoother and other existing techniques such as particle flow and annealed importance sampling. It is shown how a sequential Monte Carlo sampler can be used in combination with an annealing process to weight-correct the sampling procedure used in the ensemble smoother. Two different approximations in high dimensions, where the curse of dimensionality is unavoidable, are also presented. The methods proposed are compared with an MCMC run on a synthetic reservoir model. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
6. Evolutionary Sequential Monte Carlo Samplers for Change-Point Models
- Author
-
Arnaud Dufays
- Subjects
bayesian inference ,sequential monte carlo ,annealed importance sampling ,change-point models ,differential evolution ,GARCH models ,Economics as a science ,HB71-74 - Abstract
Sequential Monte Carlo (SMC) methods are widely used for non-linear filtering purposes. However, the SMC scope encompasses wider applications such as estimating static model parameters so much that it is becoming a serious alternative to Markov-Chain Monte-Carlo (MCMC) methods. Not only do SMC algorithms draw posterior distributions of static or dynamic parameters but additionally they provide an estimate of the marginal likelihood. The tempered and time (TNT) algorithm, developed in this paper, combines (off-line) tempered SMC inference with on-line SMC inference for drawing realizations from many sequential posterior distributions without experiencing a particle degeneracy problem. Furthermore, it introduces a new MCMC rejuvenation step that is generic, automated and well-suited for multi-modal distributions. As this update relies on the wide heuristic optimization literature, numerous extensions are readily available. The algorithm is notably appropriate for estimating change-point models. As an example, we compare several change-point GARCH models through their marginal log-likelihoods over time.
- Published
- 2016
- Full Text
- View/download PDF
7. Efficient Methods for Unsupervised Learning of Probabilistic Models
- Author
-
Sohl-Dickstein, Jascha
- Subjects
Computer science ,Statistics ,Physics ,annealed importance sampling ,energy based models ,Hamiltonian Monte Carlo ,natural gradient ,unsupervised learning - Abstract
High dimensional probabilistic models are used for many modern scientific and engineering data analysis tasks. Interpreting neural spike trains, compressing video, identifying features in DNA microarrays, and recognizing particles in high energy physics all rely upon the ability to find and model complex structure in a high dimensional space. Despite their great promise, high dimensional probabilistic models are frequently computationally intractable to work with in practice. In this thesis I develop solutions to overcome this intractability, primarily in the context of energy based models.A common cause of intractability is that model distributions cannot be analytically normalized. Probabilities can only be computed up to a constant, making training exceedingly difficult. To solve this problem I propose `minimum probability flow learning', a variational technique for parameter estimation in such models. The utility of this training technique is demonstrated in the case of an Ising model, a Hopfield auto-associative memory, an independent component analysis model of natural images, and a deep belief network.A second common difficulty in training probabilistic models arises when the parameter space is ill-conditioned. This makes gradient descent optimization slow and impractical, but can be alleviated using the natural gradient. I show here that the natural gradient can be related to signal whitening, and provide specific prescriptions for applying it to learning problems.It is also difficult to evaluate the performance of models that cannot be analytically normalized, providing a particular challenge to hypothesis testing and model comparison. To overcome this, I introduce a method termed `Hamiltonian annealed importance sampling,' which more efficiently estimates the normalization constant of non-analytically-normalizable models. This method is then used to calculate and compare the log likelihoods of several state of the art probabilistic models of natural image patches. Finally, many tasks performed with a trained probabilistic model (for instance, image denoising or inpainting and speech recognition) involve generating samples from the model distribution, which is typically a very computationally expensive process. I introduce a modification to Hamiltonian Monte Carlo sampling that reduces the tendency of sampling trajectories to double back on themselves, and enables statistically independent samples to be generated more rapidly.Taken together, it is my hope that these contributions will help scientists and engineers to build and manipulate probabilistic models.
- Published
- 2012
8. Algorithms for estimating the partition function of restricted Boltzmann machines
- Author
-
Krause, Oswin, Fischer, Asja, Igel, Christian, Krause, Oswin, Fischer, Asja, and Igel, Christian
- Abstract
Accurate estimates of the normalization constants (partition functions) of energy-based probabilistic models (Markov random fields) are highly important, for example, for assessing the performance of models, monitoring training progress, and conducting likelihood ratio tests. Several algorithms for estimating the partition function (in relation to a reference distribution) have been introduced, including Annealed Importance Sampling (AIS) and Bennett's Acceptance Ratio method (BAR). However, their conceptual similarities and differences have not been worked out so far and systematic comparisons of their behavior in practice have been missing. We devise a unifying theoretical framework for these algorithms, which comprises existing variants and suggests new approaches. It is based on a generalized form of Crooks' equality linking the expectation over a distribution of samples generated by a transition operator to the expectation over the distribution induced by the reversed operator. The framework covers different ways of generating samples, such as parallel tempering and path sampling. An empirical comparison revealed the differences between the methods when estimating the partition function of restricted Boltzmann machines and Ising models. In our experiments, BAR using parallel tempering worked well with a small number of bridging distributions, while path sampling based AIS performed best when many bridging distributions were available. Because BAR gave the overall best results, we favor it over AIS. Furthermore, the experiments showed the importance of choosing a proper reference distribution.
- Published
- 2020
9. Estimating the evidence - a review.
- Author
-
Friel, Nial and Wyse, Jason
- Subjects
- *
BAYESIAN analysis , *REGRESSION analysis , *MULTIVARIATE analysis , *FACTOR analysis , *ANALYSIS of variance - Abstract
The model evidence is a vital quantity in the comparison of statistical models under the Bayesian paradigm. This study presents a review of commonly used methods. We outline some guidelines and offer some practical advice. The reviewed methods are compared for two examples; non-nested Gaussian linear regression and covariate subset selection in logistic regression. [ABSTRACT FROM AUTHOR]
- Published
- 2012
- Full Text
- View/download PDF
10. Stochastic Gradient Annealed Importance Sampling for Efficient Online Marginal Likelihood Estimation
- Author
-
Hans C. Eggers, Steve Kroon, and Scott Cameron
- Subjects
Independent and identically distributed random variables ,FOS: Computer and information sciences ,SGHMC ,Schedule ,Computer Science - Machine Learning ,Computer science ,General Physics and Astronomy ,Machine Learning (stat.ML) ,010501 environmental sciences ,nested sampling ,annealed importance sampling ,01 natural sciences ,Stability (probability) ,Article ,Machine Learning (cs.LG) ,010104 statistics & probability ,symbols.namesake ,Factorization ,Statistics - Machine Learning ,Statistics ,0101 mathematics ,Monte Carlo ,Nested sampling algorithm ,0105 earth and related environmental sciences ,evidence ,Markov chain Monte Carlo ,marginal likelihood ,stochastic gradients ,Marginal likelihood ,symbols ,Importance sampling - Abstract
We consider estimating the marginal likelihood in settings with independent and identically distributed (i.i.d.) data. We propose estimating the predictive distributions in a sequential factorization of the marginal likelihood in such settings by using stochastic gradient Markov Chain Monte Carlo techniques. This approach is far more efficient than traditional marginal likelihood estimation techniques such as nested sampling and annealed importance sampling due to its use of mini-batches to approximate the likelihood. Stability of the estimates is provided by an adaptive annealing schedule. The resulting stochastic gradient annealed importance sampling (SGAIS) technique, which is the key contribution of our paper, enables us to estimate the marginal likelihood of a number of models considerably faster than traditional approaches, with no noticeable loss of accuracy. An important benefit of our approach is that the marginal likelihood is calculated in an online fashion as data becomes available, allowing the estimates to be used for applications such as online weighted model combination.
- Published
- 2019
11. Computation of marginal likelihoods with data-dependent support for latent variables.
- Author
-
Heaps, Sarah E., Boys, Richard J., and Farrow, Malcolm
- Subjects
- *
MARGINAL distributions , *MAXIMUM likelihood statistics , *LATENT variables , *MONTE Carlo method , *DATA analysis , *BAYESIAN analysis - Abstract
Abstract: Several Monte Carlo methods have been proposed for computing marginal likelihoods in Bayesian analyses. Some of these involve sampling from a sequence of intermediate distributions between the prior and posterior. A difficulty arises if the support in the posterior distribution is a proper subset of that in the prior distribution. This can happen in problems involving latent variables whose support depends upon the data and can make some methods inefficient and others invalid. The correction required for models of this type is derived and its use is illustrated by finding the marginal likelihoods in two examples. One concerns a model for competing risks. The other involves a zero-inflated over-dispersed Poisson model for counts of centipedes, using latent Gaussian variables to capture spatial dependence. [Copyright &y& Elsevier]
- Published
- 2014
- Full Text
- View/download PDF
12. Exponential family restricted boltzmann machines and annealed importance sampling
- Author
-
Yifeng Li and Xiaodan Zhu
- Subjects
Computer science ,Stochastic process ,Monte Carlo method ,Computer Science::Neural and Evolutionary Computation ,Boltzmann machine ,Sampling (statistics) ,deep learning ,02 engineering and technology ,annealed importance sampling ,03 medical and health sciences ,0302 clinical medicine ,Exponential family ,exponential family ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,generative model ,Algorithm ,030217 neurology & neurosurgery ,Importance sampling ,restricted Boltzmann machine - Abstract
In this paper, we investigate restricted Boltzmann machines (RBMs) from the exponential family perspective, en-abling the visible units to follow any suitable distributions from the exponential family. We derive a unified view to compute the free energy function for exponential family RBMs (exp-RBMs). Based on that, annealed important sampling (AIS) is generalized to the entire exponential family, allowing for estimating the log-partition function and log-likelihood. Our experiments on a document processing task demonstrate that the generalized free energy functions and AIS estimation perform well in helping capture useful knowledge from the data; the estimated log-partition functions are stable. The appropriate instances of exp-RBMs can generate novel and meaningful samples and can be applied to classification tasks., 2018 International Joint Conference on Neural Networks (IJCNN), July 7-13, 2018, Rio de Janeiro, Brazil, Series: Proceedings of International Joint Conference on Neural Networks
- Published
- 2018
- Full Text
- View/download PDF
13. Scalable Monte Carlo inference for state-space models
- Author
-
Yıldırım, Sinan, Andrieu, Christophe, and Doucet, Arnaud
- Subjects
Particle Markov chain Monte Carlo ,State-space models ,Methodology (stat.ME) ,FOS: Computer and information sciences ,Annealed importance sampling ,Sequential Monte Carlo ,Statistics - Methodology ,Statistics::Computation - Abstract
We present an original simulation-based method to estimate likelihood ratios efficiently for general state-space models. Our method relies on a novel use of the conditional Sequential Monte Carlo (cSMC) algorithm introduced in \citet{Andrieu_et_al_2010} and presents several practical advantages over standard approaches. The ratio is estimated using a unique source of randomness instead of estimating separately the two likelihood terms involved. Beyond the benefits in terms of variance reduction one may expect in general from this type of approach, an important point here is that the variance of this estimator decreases as the distance between the likelihood parameters decreases. We show how this can be exploited in the context of Monte Carlo Markov chain (MCMC) algorithms, leading to the development of a new class of exact-approximate MCMC methods to perform Bayesian static parameter inference in state-space models. We show through simulations that, in contrast to the Particle Marginal Metropolis-Hastings (PMMH) algorithm of Andrieu_et_al_2010, the computational effort required by this novel MCMC scheme scales very favourably for large data sets.
- Published
- 2018
14. Controlled Sequential Monte Carlo
- Author
-
Jeremy Heng, George Deligiannidis, Arnaud Doucet, and Adrian N. Bishop
- Subjects
Statistics and Probability ,FOS: Computer and information sciences ,reinforcement learning ,Computational complexity theory ,Statistics & Probability ,Monte Carlo method ,Inference ,annealed importance sampling ,Stability (probability) ,Statistics - Computation ,optimal control ,62M05 ,State space ,Reinforcement learning ,Computation (stat.CO) ,Mathematics ,0102 Applied Mathematics, 0104 Statistics, 1403 Econometrics ,normalizing constants ,Probability distribution ,62M10 ,Statistics, Probability and Uncertainty ,Particle filter ,approximate dynamic programming ,62F12 ,Algorithm ,State space models - Abstract
Sequential Monte Carlo methods, also known as particle methods, are a popular set of techniques for approximating high-dimensional probability distributions and their normalizing constants. These methods have found numerous applications in statistics and related fields; for example, for inference in nonlinear non-Gaussian state space models, and in complex static models. Like many Monte Carlo sampling schemes, they rely on proposal distributions which crucially impact their performance. We introduce here a class of controlled sequential Monte Carlo algorithms, where the proposal distributions are determined by approximating the solution to an associated optimal control problem using an iterative scheme. This method builds upon a number of existing algorithms in econometrics, physics and statistics for inference in state space models, and generalizes these methods so as to accommodate complex static models. We provide a theoretical analysis concerning the fluctuation and stability of this methodology that also provides insight into the properties of related algorithms. We demonstrate significant gains over state-of-the-art methods at a fixed computational complexity on a variety of applications.
- Published
- 2017
15. Structure discovery in Bayesian networks by sampling partial orders
- Subjects
ta113 ,Markov chain Monte Carlo ,Fast zeta transform ,Annealed importance sampling ,Directed acyclic graph ,Linear extension - Published
- 2016
16. Stochastic Gradient Annealed Importance Sampling for Efficient Online Marginal Likelihood Estimation †.
- Author
-
Cameron, Scott A., Eggers, Hans C., and Kroon, Steve
- Subjects
- *
MARKOV chain Monte Carlo , *INDEPENDENT sets , *MONTE Carlo method - Abstract
We consider estimating the marginal likelihood in settings with independent and identically distributed (i.i.d.) data. We propose estimating the predictive distributions in a sequential factorization of the marginal likelihood in such settings by using stochastic gradient Markov Chain Monte Carlo techniques. This approach is far more efficient than traditional marginal likelihood estimation techniques such as nested sampling and annealed importance sampling due to its use of mini-batches to approximate the likelihood. Stability of the estimates is provided by an adaptive annealing schedule. The resulting stochastic gradient annealed importance sampling (SGAIS) technique, which is the key contribution of our paper, enables us to estimate the marginal likelihood of a number of models considerably faster than traditional approaches, with no noticeable loss of accuracy. An important benefit of our approach is that the marginal likelihood is calculated in an online fashion as data becomes available, allowing the estimates to be used for applications such as online weighted model combination. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
17. On the conjugacy of off-line and on-line Sequential Monte Carlo Samplers
- Author
-
Dufays, Arnaud
- Subjects
Multifractal model ,Monte-Carlo-Simulation ,Bayesian inference ,Differential Evolution ,Markov-switching model ,Statistics::Computation ,Volatility models ,Markov-Kette ,Bayes-Statistik ,ddc:330 ,C58 ,Annealed Importance sampling ,C15 ,Sequentialtest ,Sequential Monte Carlo ,C11 ,C22 ,Theorie - Abstract
Sequential Monte Carlo (SMC) methods are widely used for filtering purposes of non-linear economic or financial models. Nevertheless the SMC scope encompasses wider applications such as estimating static model parameters so much that it is becoming a serious alternative to Markov- Chain Monte-Carlo (MCMC) methods. Not only SMC algorithms draw posterior distributions of static or dynamic parameters but additionally provide an estimate of the normalizing constant. The tempered and time (TNT) algorithm, developed in the paper, combines (off-line) tempered SMC inference with on-line SMC inference for estimating many slightly different distributions. The method encompasses the Iterated Batch Importance Sampling (IBIS) algorithm and more generally the Resample Move (RM) algorithm. Besides the number of particles, the TNT algorithm self-adjusts its calibrated parameters and relies on a new MCMC kernel that allows for particle interactions. The algorithm is well suited for efficiently back-testing models. We conclude by comparing in-sample and out-of-sample performances of complex volatility models.
- Published
- 2014
18. On the conjugacy of off-line and online Sequential Monte Carlo Samplers
- Author
-
Banque Nationale Belge - Département de recherche, Dufays, Arnaud, Banque Nationale Belge - Département de recherche, and Dufays, Arnaud
- Abstract
Sequential Monte Carlo (SMC) methods are widely used for non-linear filtering purposes. Nevertheless the SMC scope encompasses wider applications such as estimating static model parameters so much that it is becoming a serious alternative to Markov-Chain Monte-Carlo (MCMC) methods. Not only SMC algorithms draw posterior distributions of static or dynamic parameters but additionally provide an estimate of the normalizing constant. The tempered and time (TNT) algorithm, developed in the paper, combines (off-line) tempered SMC inference with on-line SMC inference for estimating many slightly different distributions. The method encompasses the Iterated Batch Importance Sampling (IBIS) algorithm and more generally the Re-sample Move (RM) algorithm. Besides the number of particles, the TNT algorithm self-adjusts its calibrated parameters and relies on a new MCMC kernel that allows for particle interactions. The algorithm is well suited for efficiently back-testing models. We conclude by comparing in-sample and out-of-sample performances of complex volatility models.
- Published
- 2014
19. Structure Discovery in Bayesian Networks by Sampling Partial Orders
- Author
-
Teppo Mikael Niinimäki, Pekka Parviainen, Mikko Koivisto, University of Helsinki, Department of Computer Science, Aalto-yliopisto, and Aalto University
- Subjects
Markov chain Monte Carlo ,Fast zeta transform ,Annealed importance sampling ,Directed acyclic graph ,Linear extension - Abstract
We present methods based on Metropolis-coupled Markov chain Monte Carlo (MC3) and annealed importance sampling (AIS) for estimating the posterior distribution of Bayesian networks. The methods draw samples from an appropriate distribution of partial orders on the nodes, continued by sampling directed acyclic graphs (DAGs) conditionally on the sampled partial orders. We show that the computations needed for the sampling algorithms are feasible as long as the encountered partial orders have relatively few down-sets. While the algorithms assume suitable modularity properties of the priors, arbitrary priors can be handled by dividing the importance weight of each sampled DAG by the number of topological sorts it has - we give a practical dynamic programming algorithm to compute these numbers. Our empirical results demonstrate that the presented partial-order-based samplers are superior to previous Markov chain Monte Carlo methods, which sample DAGs either directly or via linear orders on the nodes. The results also suggest that the convergence rate of the estimators based on AIS are competitive to those of MC3. Thus AIS is the preferred method, as it enables easier large-scale parallelization and, in addition, supplies good probabilistic lower bound guarantees for the marginal likelihood of the model.
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.