paper available in neuroprose
arun maskara spec lec cis
arun at hertz.njit.edu
Tue May 19 12:35:01 EDT 1992
Paper availble:
Forced Simple Recurrent Neural Networks and Grammatical Inference
Arun Maskara
New Jersey Institute of Technology
Department of Computer and Information Sciences
University Heights, Newark, NJ 07102
arun at hertz.njit.edu
Andrew Noetzel
The William Paterson College
Department of Computer Science
Wayne, NJ 07470
ABSTRACT
A simple recurrent neural network (SRN) introduced by Elman can be trained to
infer a regular grammar from the positive examples of symbol sequences generated
by the grammar. The network is trained, through the back-propagation of error,
to predict the next symbol in each sequence, as the symbols are presented
successively as inputs to the network. The modes of prediction failure of
the SRN architecture are investigated. The SRN's internal encoding of the
context (the previous symbols of the sequence) is found to be insufficiently
developed when a particular aspect of context is not required for the immediate
prediction at some point in the input sequence, but is required later. It is
shown that this mode of failure can be avoided by using the auto-associative
recurrent network (AARN). The AARN architecture contains additional output
units, which are trained to show the current input and the current context.
The effect of the size of the training set for grammatical inference is also
considered. The SRN has been shown to be effective when trained on an infinite
(very large) set of positive examples. When a finite (small) set of positive
training data is used, the SRN architectures demonstrate a lack of
generalization capability. This problem is solved through a new training
algorithm that uses both positive and negative examples of the sequences.
Simulation results show that when there is restriction on the number of nodes
in the hidden layers, the AARN succeeds in the cases where the SRN fails.
----------------------------------------------------------------------------
This paper will appear in Proceedings of 14th Annual Coggnitive Science
Conference, 1992
It is FTPable from archive.cis.ohio-state.edu
in: pub/neuroprose (Courtesy of Jordan Pollack)
Sorry, no hardcopy available.
FTP procedure:
unix> ftp archive.cis.ohio-state.edu (or 128.146.8.52)
Name: anonymous
Password: (your email address)
ftp> cd pub/neuroprose
ftp> binary
ftp> get maskara.cogsci92.ps.Z
ftp> quit
unix> uncompress maskara.cogsci92.ps.Z
unix> lpr -Pxxx maskara.cogsci92.ps (or however you print postscript)
Arun Maskara
arun at hertz.njit.edu
More information about the Connectionists
mailing list