Preprint about scaled conjugate gradient available.

Martin Moller fodslett at daimi.aau.dk
Fri Nov 30 13:48:14 EST 1990


*************************************************************
******************** PREPRINT announcement: ****************

    A Scaled Conjugate Gradient Algorithm for Fast Supervised Learning.


		Martin F. Moller
		Computer Science Dept.
		University of Aarhus
		Denmark
		e-mail: fodslett at daimi.aau.dk


Abstract-- 

A supervised learning algorithm (Scaled Conjugate Gradient, SCG)
with superlinear convergence rate is introduced. SCG uses second order 
information from the neural network, but requires only O(N) memory usage.
The performance of SCG is benchmarked against the performance of the standard
backpropagation algorithm (BP) and several recently proposed standard conjugate
gradient algorithms. SCG yields a speed-up at least an order of magnitude 
relative to BP. The speed-up depends on the convergence criterion,i.e., the
bigger demand for reduction in error the bigger the speed-up. SCG is fully
automated including no user dependent parameters and avoids a time consuming 
line search, which other conjugate gradient algorithms use in order to 
determine a good step size.
Incorporating problem dependent structural information in the architecture of 
a neural network often lowers the overall complexity. The smaller the 
complexity of the neural network relative to the problem domain, the bigger the
possibility that the weights space contains long ravines characterized by sharp
curvature. While BP is inefficient on these ravine phenomena, SCG handles them 
effectively.



The paper is available by ftp in the neuroprose directory under the name:

	moller.conjugate-gradient.ps.Z

Any question or comments on this short writing or on the preprint would 
be very much appriciated.


Martin M.





More information about the Connectionists mailing list