oh boy, more tech reports...
Michael C. Mozer
mozer%neuron at boulder.Colorado.EDU
Wed Jan 18 16:19:46 EST 1989
Please e-mail requests to "kate at boulder.colorado.edu".
Skeletonization: A Technique for Trimming the Fat
from a Network via Relevance Assessment
Michael C. Mozer
Paul Smolensky
University of Colorado
Department of Computer Science
Tech Report # CU-CS-421-89
This paper proposes a means of using the knowledge in a network to deter-
mine the functionality or _relevance_ of individual units, both for the
purpose of understanding the network's behavior and improving its perfor-
mance. The basic idea is to iteratively train the network to a certain
performance criterion, compute a measure of relevance that identifies which
input or hidden units are most critical to performance, and automatically
trim the least relevant units. This _skeletonization_ technique can be
used to simplify networks by eliminating units that convey redundant infor-
mation; to improve learning performance by first learning with spare hidden
units and then trimming the unnecessary ones away, thereby constraining
generalization; and to understand the behavior of networks in terms of
minimal "rules."
[An abridged version of this TR will appear in NIPS proceedings.]
---------------------------------------------------------------------------
And while I'm at it, some other recent junk, I mean stuff...
A Focused Back-Propagation Algorithm
for Temporal Pattern Recognition
Michael C. Mozer
University of Toronto
Connectionist Research Group
Tech Report # CRG-TR-88-3
Time is at the heart of many pattern recognition tasks, e.g., speech recog-
nition. However, connectionist learning algorithms to date are not well-
suited for dealing with time-varying input patterns. This paper introduces
a specialized connectionist architecture and corresponding specialization
of the back-propagation learning algorithm that operates efficiently on
temporal sequences. The key feature of the architecture is a layer of
self-connected hidden units that integrate their current value with the new
input at each time step to construct a static representation of the tem-
poral input sequence. This architecture avoids two deficiencies found in
other models of sequence recognition: first, it reduces the difficulty of
temporal credit assignment by focusing the back propagated error signal;
second, it eliminates the need for a buffer to hold the input sequence
and/or intermediate activity levels. The latter property is due to the
fact that during the forward (activation) phase, incremental activity
_traces_ can be locally computed that hold all information necessary for
back propagation in time. It is argued that this architecture should scale
better than conventional recurrent architectures with respect to sequence
length. The architecture has been used to implement a temporal version of
Rumelhart and McClelland's verb past-tense model. The hidden units learn
to behave something like Rumelhart and McClelland's "Wickelphones," a rich
and flexible representation of temporal information.
---------------------------------------------------------------------------
A Connectionist Model of Selective Attention in Visual Perception
Michael C. Mozer
University of Toronto
Connectionist Research Group
Tech Report # CRG-TR-88-4
This paper describes a model of selective attention that is part of a con-
nectionist object recognition system called MORSEL. MORSEL is capable of
identifying multiple objects presented simultaneously on its "retina," but
because of capacity limitations, MORSEL requires attention to prevent it
from trying to do too much at once. Attentional selection is performed by
a network of simple computing units that constructs a variable-diameter
"spotlight" on the retina, allowing sensory information within the
spotlight to be preferentially processed. Simulations of the model demon-
strate that attention is more critical for less familiar items and that at-
tention can be used to reduce inter-item crosstalk. The model suggests
four distinct roles of attention in visual information processing, as well
as a novel view of attentional selection that has characteristics of both
early and late selection theories.
More information about the Connectionists
mailing list