Paper : Learning theory and algebraic analysis and geometry
Sumio Watanabe
swatanab at pi.titech.ac.jp
Wed Jul 19 01:28:32 EDT 2000
Dear connectionists,
I would like to announce the following paper
accepted for publication in Neural Computation.
This paper firstly clarifies the algebraic geometrical
structure of the hierarchical learning machines.
Questions and comments are welcome.
+++++
Title:
Algebraic analysis for non-identifiable learning machines.
Cite:
http://watanabe-www.pi.titech.ac.jp/~swatanab/appendix.html
ABSTRACT:
This paper clarifies learning efficiency of a non-identifiable
learning machine such as a multi-layer neural network
whose true parameter set is an analytic set with
singular points. By using a concept in algebraic analysis,
we rigorously prove that the free energy or the Bayesian
stochastic complexity is asymptotically equal to
$\lambda_{1}\log n -(m_{1}-1)\log\log n + $constant,
where $\lambda_{1}$ is a rational number, $m_{1}$ is a
natural number, and $n$ is the number of training samples.
Also we show an algorithm to calculate $\lambda_{1}$ and
$m_{1}$ based on the resolution of singularities in algebraic
geometry. In regular models, $2\lambda_{1}$ is equal to
the number of parameters and $m_{1}=1$, whereas in
non-regular models such as mutilayer networks
$2\lambda_{1}$ is not larger than the number of
parameters and $m_{1}\geq 1$. Since the increase of
the free energy is equal to the generalization error, the
non-identifiable learning machines are the better models
than the regular ones if Bayesian or ensemble training is
applied.
+++++
Sincerely,
Dr. Sumio Watanabe, Associate Professor
Advanced Information Processing Division
Precision and Intelligence Laboratory
Tokyo Institute of Technology
E-mail: swatanab at pi.titech.ac.jp
(Fax)+81-45-924-5018
More information about the Connectionists
mailing list