new paper in neuroprose

Wolfgang Maass maass at igi.tu-graz.ac.at
Thu Jul 1 13:19:37 EDT 1993




FTP-host: archive.cis.ohio-state.edu
FTP-filename: /pub/neuroprose/maass.agnostic.ps.Z

The file maass.agnostic.ps.Z is now available for copying
from the Neuroprose repository. This is a 22-page long paper.
Hardcopies are not available.


      AGNOSTIC PAC-LEARNING OF FUNCTIONS ON ANALOG NEURAL NETS  

                               by
                          Wolfgang Maass                    

           Institute for Theoretical Computer Science   
         Technische Universitaet Graz, A-8010 Graz, Austria
                  email: maass at igi.tu-graz.ac.at   
 

Abstract:
This is a revised version of a preprint from May 93. In this new
version we have added a parallelization  of the learning algorithm 
LEARN that runs in polynomial time in ALL relevant parameters.

We consider learning of real-valued functions on analog neural nets
in Haussler's refinement of Valiant's model for probably approximately
correct learning ("PAC-learning"). One commonly refers to this refined
model as "agnostic PAC-learning", since it requires no a-priori
assumptions about the structure of the learning target. The learning 
target need not even be a function, instead it may be any distribution 
of input/output pairs. In particular, arbitrary errors and imprecisions 
are permitted in the training data. Hence the setup of this model is 
well-suited for the analysis of real-world learning problems on neural 
Cnets. 

The goal of the learner in this model is to find a hypothesis whose  
true error (resp. loss) is within an epsilon of the true
error of the best hypothesis from the considered hypothesis class. In 
our application to neural nets this hypothesis class is given by the 
class of all functions that can be computed on some fixed neural net N 
(with arbitrary real weights). 

We prove a positive result about agnostic PAC-learning on an arbitrary
fixed analog neural net N (with arbitrary piecewise polynomial
activation functions). We first construct for such N a somewhat larger
neural net N' (the "learning network"). Then we exhibit a learning 
algorithm LEARN that computes from the training examples an assignment 
of rational numbers to the weights of N' such that with high probability
the true error of the function that is computed by N' with these weights 
is within an epsilon of that of the best hypothesis that is computable 
on N (with arbitrary real weights).  

If the number of gates in N may be viewed as a constant, then the 
computation time of the learning algorithm LEARN is polynomial (in all
other relevant parameters). In addition one can parallelize this
learning algorithm in such a way that its computation time is polynomial 
in ALL relevant parameters, including the number of gates in N .

It should be noted that in contrast to common learning algorithms such
as backwards propagation, this learning algorithm LEARN is guaranteed
to solve its learning task (with high probability).

As a part of this positive learning result we show that the pseudo-
dimension of neural nets with piecewise polynomial activation functions 
can be bounded by a polynomial in the number of weights of the neural 
net. It has previously been shown by Haussler that the pseudo-dimension 
is the appropriate generalization of the VC-dimension for learning real 
valued functions. With the help of our upper bound on the pseudo-
dimension of neural nets with piecewise polynomial activation functions 
we can bound the number of training-examples that are needed  for 
agnostic PAC-learning. In this way one can reduce the task of minimizing
the true error of the neural net to the finite task of minimizing its
apparent error (i.e. its error on the training examples).



More information about the Connectionists mailing list