reprint available
john moody
moody-john at CS.YALE.EDU
Mon Aug 17 16:10:51 EDT 1992
Fellow Connectionists:
The following reprint has been placed on Jordan Pollack's neuroprose archive:
LEARNING RATE SCHEDULES FOR FASTER STOCHASTIC GRADIENT SEARCH
Christian Darken*, Joseph Chang+, and John Moody*
Yale Departments of Computer Science* and Statistics+
ABSTRACT
Stochastic gradient descent is a general algorithm that includes LMS,
on-line backpropagation, and adaptive k-means clustering as special cases.
The standard choices of the learning rate $\eta$ (both adaptive and fixed
functions of time) often perform quite poorly. In contrast, our recently
proposed class of ``search then converge'' ($STC$) learning rate schedules
(Darken and Moody, 1990b, 1991) display the theoretically optimal asymptotic
convergence rate and a superior ability to escape from poor local minima.
However, the user is responsible for setting a key parameter. We propose
here a new methodology for creating the first automatically adapting learning
rates that achieve the optimal rate of convergence.
To retrieve it via anonymous ftp, do the following:
% ftp cheops.cis.ohio-state.edu
Connected to cheops.cis.ohio-state.edu.
220 cheops.cis.ohio-state.edu FTP server ready.
Name: anonymous
331 Guest login ok, send ident as password.
Password: your email addressneuron
230 Guest login ok, access restrictions apply.
ftp> binary
200 Type set to I.
ftp> cd pub/neuroprose
250 CWD command successful.
ftp> get darken.learning_rates.ps.Z
200 PORT command successful.
150 Opening ASCII mode data connection for darken.learning_rates.ps.Z (238939 by
tes).
226 Transfer complete.
local: darken.learning_rates.ps.Z remote: darken.learning_rates.ps.Z
239730 bytes received in 11 seconds (22 Kbytes/s)
ftp> quit
221 Goodbye.
% uncompress darken.learning_rates.ps
% lpr -P printer_name darken.learning_rates.ps
Enjoy,
John Moody
-------
More information about the Connectionists
mailing list