Does backprop need the derivative ??

Scott_Fahlman@SEF-PMAX.SLISP.CS.CMU.EDU Scott_Fahlman at SEF-PMAX.SLISP.CS.CMU.EDU
Fri Feb 5 22:55:28 EST 1993


    In his paper, 'An Empirical Study of Learning Speed in Back-Propagation
    Networks', Scott E. Fahlmann shows that with the encoder/decoder problem
    it is possible to replace the derivative of the transfer function by
    a constant. I have been able to reproduce this example. However, for
    several other examples, it was not possible to get the network 
    converged using a constant for the derivative.
    
Interesting.  I just tried this on encoder problems and a couple of other
simple things, and leapt to the conclusion that it was a general
phenomenon.  It seems plausible to me that any "derivative" function that
preserves the sign of the error and doesn't have a "flat spot" (stable
point of 0 derivative) would work OK, but I don't know of anyone who has
made an extensive study of this.

I'd be interested in hearing more about the problems you've encountered and
about any results others send to you.

-- Scott

===========================================================================
Scott E. Fahlman			Internet:  sef+ at cs.cmu.edu
Senior Research Scientist		Phone:     412 268-2575
School of Computer Science              Fax:       412 681-5739
Carnegie Mellon University		Latitude:  40:26:33 N
5000 Forbes Avenue			Longitude: 79:56:48 W
Pittsburgh, PA 15213
===========================================================================



More information about the Connectionists mailing list