Why batch learning is slower
Nathan Intrator
nin at cns.brown.edu
Thu Mar 19 10:59:36 EST 1992
"Why Batch Learning is Slower Than Per-Sample Learning"
Thomas H. Hildebrandt
From the abstract:
"...For either algorithm, convergence is guaranteed as long as no
step exceeds the minimum ideal step size by more than a factor of 2.
By limiting the discussion to a fixed, safe step size, we can compare
the maximum step that can be taken by each algorithm in the worst case."
-------
There is no "FIXED safe step size" for the stochastic version, namely there is
no convergence proof for a fixed learning rate of the stochastic version.
The paper cited by Chung-Ming Kuan and Kurt Hornik does not imply that either.
It is therefore difficult to draw conclusions from this paper.
- Nathan
More information about the Connectionists
mailing list