Sequential Monte Carlo methods to train neural network models

Citation
Jfg. De Freitas et al., Sequential Monte Carlo methods to train neural network models, NEURAL COMP, 12(4), 2000, pp. 955-993
Citations number
66
Categorie Soggetti
Neurosciences & Behavoir","AI Robotics and Automatic Control
Journal title
NEURAL COMPUTATION
ISSN journal
08997667 → ACNP
Volume
12
Issue
4
Year of publication
2000
Pages
955 - 993
Database
ISI
SICI code
0899-7667(200004)12:4<955:SMCMTT>2.0.ZU;2-K
Abstract
We discuss a novel strategy for training neural networks using sequential M onte Carlo algorithms and propose a new hybrid gradient descent/sampling im portance resampling algorithm (HySIR). In terms of computational time and a ccuracy, the hybrid SIR is a clear improvement over conventional sequential Monte Carlo techniques. The new algorithm may be viewed as a global optimi zation strategy that allows us to learn the probability distributions of th e network weights and outputs in a sequential framework. It is well suited to applications involving on-line, nonlinear, and nongaussian signal proces sing. We show how the new algorithm outperforms extended Kalman filter trai ning on several problems. In particular, we address the problem of pricing option contracts, traded in financial markets. In this context, we are able to estimate the one-step-ahead probability density functions of the option s prices.