bayesian methods for back-propagation, update
Wray Buntine
wray at ptolemy.arc.nasa.gov
Mon Dec 23 13:57:28 EST 1991
To appear in December 1991 issue of {\it Complex Systems}.
An early draft appeared in the Neuroprose archive July 1991,
and was available as NASA Ames AI Research Branch
TR FIA-91-22. The new version is considerably improved and
contains new, updated and corrected material.
A limited number of reprints are available. by writing to
Wray Buntine. (Andreas is currently on holidays!) Please
only request one if your library doesn't get Complex Systems
journal.
PS. distributing the first draft via connectionists allowed us
to get all sorts of helpful feedback on the early version!
-------------
Bayesian Back-Propagation
Wray L. Buntine Andreas S. Weigend
wray at ptolemy.arc.nasa.gov andreas at psych.stanford.edu
RIACS \& NASA Ames Research Center Xerox Palo Alto Research Center
Mail Stop 269-2 3333 Coyote Hill Rd.
Moffet Field, CA 94035, USA Palo Alto, CA, 94304, USA
Connectionist feed-forward networks, trained with back-propagation,
can be used both for non-linear regression and for (discrete
one-of-$C$) classification. This paper presents approximate Bayesian
methods to statistical components of back-propagation: choosing a cost
function and penalty term (interpreted as a form of prior
probability), pruning insignificant weights, estimating the
uncertainty of weights, predicting for new patterns
(``out-of-sample''), estimating the uncertainty in the choice of this
prediction (``error bars''), estimating the generalization error,
comparing different network structures, and handling missing values in
the training patterns. These methods extend some heuristic techniques
suggested in the literature, and in most cases require a small
additional factor in computation during back-propagation, or
computation once back-propagation has finished.
More information about the Connectionists
mailing list