new paper on Asymptotic Statistical Theory of Overtraining and Cross-Validation
klaus@prosun.first.gmd.de
klaus at prosun.first.gmd.de
Mon Sep 4 05:08:16 EDT 1995
FTP-host: archive.cis.ohio-state.edu
FTP-file: pub/neuroprose/amari.overtraining.ps.Z
The following paper is now available for copying from the Neuroprose
repository: amari.overtraining.ps.Z
amari.overtraining.ps.Z klaus at first.gmd.de
(128151 bytes) 32 pages.
S. Amari, N. Murata, K.-R. M\"uller, M. Finke, H. Yang:
"Asymptotic Statistical Theory of Overtraining and Cross-Validation"
A statistical theory for overtraining is proposed. The analysis
treats general realizable stochastic neural networks, trained with
Kullback-Leibler loss in the asymptotic case of a large number
of training examples.
It is shown that the asymptotic gain in the generalization error is
small if we perform early stopping, even if we have access to the
optimal stopping time. Considering cross-validation stopping we
answer the question: In what ratio the examples should be divided
into training and testing sets in order to obtain the optimum
performance. However, cross-validated early stopping is useless in
the asymptotic region, while it decreases the generalization error
only in the non-asymptotic region. Our large scale simulations done
on a CM5 are in nice agreement with our analytical findings.
(University of Tokyo Technical Report METR 06-95 and submitted to IEEE
Transactions on NN)
Best regards,
Klaus
&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
Dr. Klaus-Robert M\"uller
GMD First (Forschungszentrum Informationstechnik)
Rudower Chaussee 5, 12489 Berlin
Germany
mail: klaus at first.gmd.de
Tel: +49 30 6392 1860
Fax: +49 30 6392 1805
web-page:
http://www.first.gmd.de/persons/Mueller.Klaus-Robert.html
&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&
More information about the Connectionists
mailing list