NIPS paper in Neuroprose

Anders Krogh krogh at cse.ucsc.edu
Fri Jan 17 15:02:35 EST 1992



The following paper has been placed in the Neuroprose archive

Title:   A Simple Weight Decay Can Improve Generalization
Authors: Anders Krogh and John A. Hertz

Filename: krogh.weight-decay.ps.Z

(To appear in proceedings from NIPS 91)

Abstract:
It has been observed in numerical simulations that a weight decay can
improve generalization in a feed-forward neural network.  This paper
explains why.  It is proven that a weight decay has two effects in a
linear network.  First, it suppresses any irrelevant components of the
weight vector by choosing the smallest vector that solves the learning
problem.  Second, if the size is chosen right, a weight decay can
suppress some of the effects of static noise on the targets, which
improves generalization quite a lot.  It is then shown how to extend
these results to networks with hidden layers and non-linear units.
Finally the theory is confirmed by some numerical simulations using the
data from NetTalk.


----------------------------------------------------------------
FTP INSTRUCTIONS

     unix> ftp archive.cis.ohio-state.edu (or 128.146.8.52)
     Name: anonymous
     Password: anything
     ftp> cd pub/neuroprose
     ftp> binary
     ftp> get krogh.weight-decay.ps.Z
     ftp> bye
     unix> zcat krogh.weight-decay.ps.Z | lpr
                  (or however you uncompress and print postscript)



More information about the Connectionists mailing list