New Bayesian work
David MacKay
mackay at hope.caltech.edu
Tue May 21 13:40:57 EDT 1991
Two new papers available
------------------------
The papers that I presented at Snowbird this year are
now available in the neuroprose archives.
The titles:
[1] Bayesian interpolation (14 pages)
[2] A practical Bayesian framework
for backprop networks (11 pages)
The first paper describes and demonstrates recent
developments in Bayesian regularisation and model comparison.
The second applies this framework to backprop. The first paper
is a prerequisite for understanding the second.
Abstracts and instructions for anonymous ftp follow.
If you have problems obtaining the files by ftp, feel
free to contact me.
David MacKay Office: (818) 397 2805
Fax: (818) 792 7402
Email: mackay at hope.caltech.edu
Smail: Caltech 139-74,
Pasadena, CA 91125
Abstracts
---------
Bayesian interpolation
----------------------
Although Bayesian analysis has been in use since Laplace,
the Bayesian method of {\em model--comparison} has only
recently been developed in depth.
In this paper, the Bayesian approach to
regularisation and model--comparison is demonstrated by
studying the inference problem of interpolating noisy data.
The concepts and methods described are quite general and can
be applied to many other problems.
Regularising constants are set by examining their
posterior probability distribution. Alternative regularisers
(priors) and alternative basis sets are objectively compared
by evaluating the {\em evidence} for them. `Occam's razor' is
automatically embodied by this framework.
The way in which Bayes infers the values of regularising
constants and noise levels has an elegant interpretation in terms
of the effective number of parameters determined by the data set.
This framework is due to Gull and Skilling.
A practical Bayesian framework for backprop networks
----------------------------------------------------
A quantitative and practical Bayesian framework is described
for learning of mappings in feedforward networks.
The framework makes possible:
(1) objective comparisons between solutions using
alternative network architectures;
(2) objective stopping rules for deletion of weights;
(3) objective choice of magnitude and type of weight decay terms or
additive regularisers (for penalising large weights, etc.);
(4) a measure of the effective number of well--determined parameters
in a model;
(5) quantified estimates of the error bars on network parameters
and on network output;
(6) objective comparisons with alternative learning and interpolation
models such as splines and radial basis functions.
The Bayesian `evidence' automatically embodies `Occam's razor,'
penalising over--flexible and over--complex architectures.
The Bayesian approach helps detect poor underlying assumptions in
learning models. For learning models well--matched to a problem,
a good correlation between generalisation ability and the Bayesian
evidence is obtained.
Instructions for obtaining copies by ftp from neuroprose:
---------------------------------------------------------
unix> ftp cheops.cis.ohio-state.edu # (or ftp 128.146.8.62)
Name: anonymous
Password: neuron
ftp> cd pub/neuroprose
ftp> binary
ftp> get mackay.bayes-interpolation.ps.Z
ftp> get mackay.bayes-backprop.ps.Z
ftp> quit
unix> [then `uncompress' files and lpr them.]
More information about the Connectionists
mailing list