Back to Search Start Over

Sequential monte carlo methods To train neural network models

Authors :
de Freitas JF
M Niranjan M
Gee AH
doucet A
Source :
Neural computation [Neural Comput] 2000 Apr; Vol. 12 (4), pp. 955-93.
Publication Year :
2000

Abstract

We discuss a novel strategy for training neural networks using sequential Monte Carlo algorithms and propose a new hybrid gradient descent sampling importance resampling algorithm (HySIR). In terms of computational time and accuracy, the hybrid SIR is a clear improvement over conventional sequential Monte Carlo techniques. The new algorithm may be viewed as a global optimization strategy that allows us to learn the probability distributions of the network weights and outputs in a sequential framework. It is well suited to applications involving on-line, nonlinear, and nongaussian signal processing. We show how the new algorithm outperforms extended Kalman filter training on several problems. In particular, we address the problem of pricing option contracts, traded in financial markets. In this context, we are able to estimate the one-step-ahead probability density functions of the options prices.

Details

Language :
English
ISSN :
1530-888X
Volume :
12
Issue :
4
Database :
MEDLINE
Journal :
Neural computation
Publication Type :
Academic Journal
Accession number :
10770839
Full Text :
https://doi.org/10.1162/089976600300015664