Abstract
We propose a Bayesian framework for regression problems, which covers areas which are usually dealt with by function approximation. An online learning algorithm is derived which solves regression problems with a Kalman filter. Its solution always improves with increasing model complexity, without the risk of over-fitting. In the infinite dimension limit it approaches the true Bayesian posterior. The issues of prior selection and over-fitting are also discussed, showing that some of the commonly held beliefs are misleading. The practical implementation is summarised. Simulations using 13 popular publicly available data sets are used to demonstrate the method and highlight important issues concerning the choice of priors.
Original language | English |
---|---|
Pages (from-to) | 130-142 |
Number of pages | 13 |
Journal | Neural Computing and Applications |
Volume | 4 |
Issue number | 3 |
DOIs | |
Publication status | Published - Sept 1996 |
Bibliographical note
The original publication is available at www.springerlink.comKeywords
- Bayesian framework
- regression problems
- Kalman filter
- Simulations