NIPS'89 Postconference Workshops

Alex.Waibel@SPEECH2.CS.CMU.EDU Alex.Waibel at SPEECH2.CS.CMU.EDU
Fri Oct 6 21:19:38 EDT 1989


Below are the preliminary program and brief descriptions of the workshop
topics covered during this years NIPS-Postconference Workshops to be
held in Keystone from November 30 through December 2 (right following
the NIPS conference).  Please register for both conference and Workshops
using the general NIPS conference registration forms.  With it, please
indicate which workshop topic below you may be most interested in
attending.  Your preferences are in no way binding or limiting you to any
particular workshop but will help us in allocating suitable meeting rooms
and scheduling workshop sessions in an optimal way.  For your convenience,
you may simply include a copy of the form below with your registration
material marking it for your three most prefered workshop choices in order
of preference (1,2 and 3).  

For registration information (both NIPS conference as well as Postconference
Workshops), please contact the Local Arrangements Chair, Kathie Hibbard,
by sending email to hibbard at boulder.colorado.edu, or by writing to:
	Kathie Hibbard
	NIPS '89
	University of Colorado
	Campus  Box 425
	Boulder, Colorado  80309-0425

For technical questions relating to individual conference workshops, please
contact the individual workshop leaders listed below.  Please feel free to
contact me with any questions you may have about the workshops in general.
See you in Denver/Keystone,

	Alex Waibel
	NIPS Workshop Program Chairman
	School of Computer Science
	Carnegie Mellon University
	Pittsburgh, PA 15213
	412-268-7676, waibel at cs.cmu.edu	


================================================================
  ____________________________________________________________
 !       POST  CONFERENCE  WORKSHOPS  AT  KEYSTONE            !
 !  THURSDAY,  NOVEMBER  30  -  SATURDAY,  DECEMBER  2,  1989 !
 !____________________________________________________________!

   Thursday,  November  30,  1989
       5:00 PM:  Registration and Reception at Keystone

   Friday,  December  1,  1989
       7:30 -  9:30 AM: Small Group Workshops
       4:30 -  6:30 PM: Small Group Workshops
       7:30 - 10:30 PM: Banquet and Plenary Discussion

   Saturday,  December  2,  1989
       7:30 -  9:30 AM: Small Group Workshops
       4:30 -  6:30 PM: Small Group Workshops
       6:30 -  7:15 PM: Plenary Discussion, Summaries
       7:30 - 11:00 PM: Fondue Dinner, MountainTop Restaurant

================================================================


PLEASE MARK YOUR PREFERENCES (1,2,3) AND ENCLOSE WITH REGISTRATION MATERIAL:
-----------------------------------------------------------------------------

______1.	LEARNING THEORY:  STATISTICAL ANALYSIS OR VC DIMENSION?
______2.	STATISTICAL INFERENCE IN NEURAL NETWORK MODELLING
______3.	NEURAL NETWORKS AND GENETIC ALGORITHMS
______4.	VLSI NEURAL NETWORKS: CURRENT MILESTONES AND FUTURE HORIZONS
______5.	APPLICATION OF NEURAL NETWORK PROCESSING TECHNIQUES TO REAL
			WORLD MACHINE VISION PROBLEMS
______6.	IMPLEMENTATIONS OF NEURAL NETWORKS ON DIGITAL, MASSIVELY
			PARALLEL COMPUTERS
______7.	LARGE, FAST, INTEGRATED SYSTEMS BASED ON ASSOCIATIVE MEMORIES
______8.	NEURAL NETWORKS FOR SEQUENTIAL PROCESSING WITH APPLICATIONS IN
			SPEECH RECOGNITION
______9.	LEARNING FROM NEURONS THAT LEARN
______10.	NEURAL NETWORKS AND OPTIMIZATION PROBLEMS
      11.	(withdrawn)
______12.	NETWORK DYNAMICS
______13.	ARE REAL NEURONS HIGHER ORDER NETS?
______14.	NEURAL NETWORK LEARNING: MOVING FROM BLACK ART TO A PRACTICAL
			TECHNOLOGY
______15.	OTHERS ?? __________________________________________________






1.	LEARNING THEORY:  STATISTICAL ANALYSIS OR VC DIMENSION?

		  	     Sara A. Solla
			AT&T Bell Laboratories
	 		 Crawford Corner Road
		       Holmdel, NJ   07733-1988

			Phone:   (201) 949-6057
			E-mail:  solla at homxb.att.com

Recent success at describing the process of learning in layered neural
networks and the resulting generalization ability has emerged from two
different approaches.  Work based on the concept of VC dimension emphasizes
the connection between learning and statistical inference in order to analyze
questions of bias and variance.  The statistical approach uses an ensemble
description to focus on the prediction of network performance for a
specific task.

Participants interested in learning theory are invited to discuss the 
differences and similarities between the two approaches, the mathematical 
relation between them, and their respective range of applicability.  Specific
questions to be discussed include comparison of predictions for required
training set sizes, for the distribution of generalization abilities, for
the probability of obtaining good performance with a training set of
fixed size, and for estimates of problem complexity applicable to the
determination of learning times.



2.	STATISTICAL INFERENCE IN NEURAL NETWORK MODELLING

		  Workshop Chair:  Richard Golden
			Stanford University
	   	       Psychology Department
			Stanford, CA   94305
			  (415) 725-2456
        	E-mail:  golden at psych.stanford.edu

This workshop is designed to show how the theory of statistical inference
is directly applicable to some difficult neural network modelling problems.
The format will be tutorial in nature (85% informal lecture, 15%
discussion).  Topics to be discussed include:  obtaining probability
distributions for neural networks, interpretation and derivation of optimal
learning cost functions, evaluating the generalization performance of
networks, asymptotic sampling distributions of network weights, statistical
mechanics calculation of learning curves in some simple examples,
statistical tests for comparing internal representations and deciding
which input units are relevant to the prediction task.  Dr. Naftali Tishby
(AT&T Bell Labs) and Professor Halbert White (UCSD Economics Department)
are the invited experts.



3.	Title: NEURAL NETWORKS AND GENETIC ALGORITHMS

	Organizers:  Lawrence Davis (Bolt Beranek and Newman, Inc.) 
		     Michael Rudnick (Oregon Graduate Center)

Description: Genetic algorithms have many interesting relationships with
neural networks.  Recently, a number of researchers have investigated some
of these relationships.  This workshop will be the first forum bringing
those researchers together to discuss the current and future directions of
their work.  The workshop will last one day and will have three parts.
First, a tutorial on genetic algorithms will be given, to ground those
unfamiliar with the technology.  Second, seven researchers will summarize
their results.  Finally there will be an open discussion on the topics
raised in the workshop.  We expect that anyone familiar with neural network
technology will be comfortable with the content and level of discussion in
this workshop.




4.		       VLSI NEURAL NETWORKS:
              CURRENT MILESTONES AND FUTURE HORIZONS

        		     Moderators: 

Joshua Alspector     		and	    Daniel B. Schwartz
Bell Communications Research		    GTE Laboratories, Inc.
445 South Street			    40 Sylvan Road
Morristown, NJ   07960-19910		    Waltham, MA   02254
(201) 829-4342				    (617) 466-2414
e-mail:  josh at bellcore.com		    e-mail:  dbs%gte.com at relay.cs.net

This workshop will explore the areas of applicability of neural network
implementations in VLSI.  Several speakers will discuss their present
implementations and speculate about where their work may lead.  Workshop
attendees will then be encouraged to organize working groups to address
several issues which will be raised in connection with the presentations.
Although it is difficult to predict which issues will be selected, some
examples might be:

         1) Analog vs. digital implementations.
         2) Limits to VLSI complexity for neural networks.
         3) Algorithms suitable for VLSI architectures.

The working groups will then report results which will be included in the
workshop summary.



5.	APPLICATION OF NEURAL NETWORK PROCESSING TECHNIQUES TO REAL
		      WORLD MACHINE VISION PROBLEMS


Paul J. Kolodzy        (617) 981-3822     kolodzy at ll.ll.mit.edu
Murali M. Menon        (617) 981-5374


This workshop will discuss the application of neural networks to
vision applications, including image restoration and pattern
recognition.  Participants will be asked to present their specific
application for discussion to highlight the relevant issues.
Examples of such issues include, but are not limited to, the use
of deterministic versus stochastic search procedures for neural
network processing, using networks to extract shape, scale and
texture information for recognition and using network mapping
techniques to increase data separability.  The discussions will
be driven by actual applications with an emphasis on the advantages
of using neural networks at the system level in addition to the
individual processing steps.  The workshop will attempt to cover a
wide breadth of network architectures and invites participation
from researchers in machine vision, neural network modeling,
pattern recognition and biological vision.




6.  		IMPLEMENTATIONS OF NEURAL NETWORKS ON 
		DIGITAL, MASSIVELY PARALLEL COMPUTERS

			Dr. K. Wojtek Przytula
				  and
			    Prof. S.Y. Kung
		 Hughes Research Laboratories, RL 69
			3011 Malibu Cyn. Road
			  Malibu, CA   90265

			Phone:  (213) 317-5892
		E-mail:  wojtek%csfvax at hac2arpa.hac.com

			
Implementations of neural networks span a full spectrum from software
realizations on general-purpose computers to strictly special-purpose hardware
realizations. Implementations on programmable, parallel machines, which are to
be discussed during the workshop, constitute a compromise between the two
extremes. The architectures of programmable parallel machines reflect the
structure of neural network models better than those of sequential machines,
thus resulting in higher processing speed. The programmability provides more
flexibility than is available in specialized hardware implementations and
opens a way for realization of various models on a single machine. The issues
to be discussed include: mapping neural network models onto existing parallel
machines, design of specialized programmable parallel machines for neural
networks, evaluation of performance of parallel machines for neural networks,
uniform characterization of the computational requirements of various neural
network models from the point of view of parallel implementations. 



7.		LARGE, FAST, INTEGRATED SYSTEMS 
		 BASED ON ASSOCIATIVE MEMORIES

		       Michael R. Raugh

	     Director of Learning Systems Division
    Research Institute for Advanced Computer Science (RIACS)
		NASA Ames Research Center, MS 230-5
		    Moffett Field, CA 94035

	 	    e-mail:  raugh at riacs.edu
		    Phone:   (415) 694-4998


This workshop will address issues in the construction of large systems 
that have thousands or even millions of hidden units.  It will present and 
discuss alternatives to backpropagation that allow large systems to learn
rapidly.  Examples from image analysis, weather prediction, and speech
transcription will be discussed.

The focus on backpropagation with its slow learning has kept researchers from
considering such large systems.  Sparse distributed memory and related
associative-memory structures provide an alternative that can learn,
interpolate, and abstract, and can do so rapidly.

The workshop is open to everyone, with special encouragement to those working
in learning, time-dependent networks, and generalization.



8.	NEURAL NETWORKS FOR SEQUENTIAL PROCESSING WITH
	     APPLICATIONS IN SPEECH RECOGNITION

			 Herve Bourlard 

		Philips Research Laboratory Brussels 
		Av. Van Becelaere 2, Box 8
		B-1170 Brussels, Belgium

		Phone: 011-32-2-674-22-74 

		e-mail address:  bourlard at prlb.philips.be 
			    or:  prlb2!bourlard at uunet.uu.net

Speech recognition must contend with the statistical and sequential nature of
the human speech production system.  Hidden Markov Models (HMM) provide
a powerful method to cope with both of these, and their use made a 
breakthrough in speech recognition.  On the other hand, neural networks
have recently been recognized as an alternative tool for pattern recognition 
problems such as speech recognition.  Their main useful properties are their 
discriminative power and their capability to deal with non-explicit knowledge.
However, the sequential aspect remains difficult to handle in connectionist 
models.  If connections are supplied with delays, feedback loops can be added 
providing dynamic and implicit memory.  However, in the framework of 
continuous speech recognition, it is still difficult to use only neural 
networks for the segmentation and recognition of a sentence into a sequence 
of speech units, which is efficiently solved in the HMM approach
by the well known ``Dynamic Time Warping'' algorithm.

This workshop should be the opportunity for reviewing neural network 
architectures which are potentially able to deal with sequential and 
stochastic inputs.  It should also be discussed to which extent the different
architectures can be useful in recognizing isolated units (phonemes, 
words, ...) or continuous speech.  Amongst others, we should consider 
spatiotemporal models, time-delayed neural networks (Waibel, Sejnowsky), 
temporal flow models (Watrous), hidden-to-input (Elman) or output-to-input 
(Jordan) recurrent models, focused back-propagation networks (Mozer) or 
hybrid approaches mixing neural networks and standard sequence matching 
techniques (Sakoe, Bourlard).



9.                LEARNING FROM NEURONS THAT LEARN

                          Moderated by
                         Thomas P. Vogl
          Environmental Research Institute of Michigan
                        1501 Wilson Blvd.
                       Arlington, VA 22209
                      Phone: (703) 528-5250
                    E-mail: TVP%nihcu.bitnet at cunyvm.cuny.edu
                       FAX: (703) 524-3527

In furthering our understanding of artificial and biological neural
systems, the insights that can be gained from the perceptions of
those trained in other disciplines can be particularly fruitful.
Computer scientists, biophysicists, engineers, psychologists,
physicists, and neurobiologists tend to have different perspectives
and conceptions of the mechanisms and components of "neural
networks" and to weigh differently their relative importance. The
insights obvious to practitioners of one of these disciplines are
often far from obvious to those trained in another, and therefore
may be especially relevant to the solutions of ornery problems.

The workshop provides a forum for the interdisciplinary discussion
of biological and artificial networks and neurons and their
behavior.  Informal group  discussion of ongoing research, novel
ideas, approaches, comparisons, and the sharing of insights will
be emphasized.  The specific topics to be considered and the depth
of the analysis/discussion devoted to any topic will be determined
by the interest and enthusiasm of the participants as the
discussion develops.  Participants are encouraged to consider
potential topics in advance, and to present them informally but
succinctly (under five minutes) at the beginning of the workshop.



10.	        NEURAL NETWORKS AND OPTIMIZATION PROBLEMS
	        ----------------------------------------

			Prof. Carsten Peterson
			University of Lund
			Dept. of Theoretical Physics
			Solvegatan 14A
			S-223 62 Lund		
			Sweden
			phone: 011-46-46-109002
			bitnet: THEPCAP%SELDC52

			Workshop description:

The purpose of the workshop is twofold; to establish the present state
of the art and to generate novel ideas. With respect to the former, firm
answers to the following questions should emerge: (1). Does the Hopfield-
Tank approach or variants thereof really work with respect to quality,
reliability, parameter insensitivity and scalability? (2). If this is
the case, how does it compare with other cellular approaches like "elastic
snake" and genetic algorithms?  Novel ideas should focus on new encoding
schemes and new application areas (in particular, scheduling problems).
Also, if time allows, optimization of neural network learning architectures
will be covered.

People interested in participating are encouraged to communicate their
interests and expertise to the chairman via e-mail. This would facilitate
the planning.


12.			Title:  NETWORK DYNAMICS

Chair: 	Richard Rohwer
	Centre for Speech Technology Research
	Edinburgh University
	80, South Bridge
	Edinburgh  EH1 1HN, Scotland

Phone:	(44 or 0) (31) 225-8883 x280

e-mail:	rr%uk.ac.ed.eusip at nsfnet-relay.ac.uk

Summary:

	This workshop will be an attempt to gather and improve our
knowledge about the time dimension of the activation patterns produced
by real and model neural networks.  This broad subject includes the
description, interpretation and design of these temporal patterns.  For
example, methods from dynamical systems theory have been used to
describe the dynamics of network models and real brains.  The design
problem is being approached using dynamical training algorithms. 
Perhaps the most important but least understood problems concern the
cognitive and computational significance of these patterns.  The
workshop aims to summarize the methods and results of researchers from
all relevant disciplines, and to draw on their diverse insights in order
to frame incisive, approachable questions for future research into network
dynamics.


Richard Rohwer                            JANET: rr at uk.ac.ed.eusip
Centre for Speech Technology Research      ARPA: rr%uk.ac.ed.eusip at nsfnet-relay.ac.uk
Edinburgh University                     BITNET: rr at eusip.ed.ac.uk,
80, South Bridge                                 rr%eusip.ed.UKACRL
Edinburgh  EH1 1HN,   Scotland             UUCP: ...!{seismo,decvax,ihnp4}
                                                          !mcvax!ukc!eusip!rr
PHONE:  (44 or 0) (31) 225-8883 x280        FAX: (44 or 0) (31) 226-2730



13.		 ARE REAL NEURONS HIGHER ORDER NETS?

Most existing artificial neural networks have processing elements which 
are computationally much simpler than real neurons. One approach to 
enhancing the computational capacity of artificial neural networks is to 
simply scale up the number of processing elements, but there are limits 
to this. An alternative is to build  modules or subnets and link these 
modules in a larger net. Several groups of investigators have begun 
to analyze the computational abilities of real single neurons in terms of 
equivalent neural nets, in particular higher order nets, in which the 
inputs explicitly interact (eg. sigma-pi units). This workshop would 
introduce participants to the results of these efforts, and examine the 
advantages and problems of applying these complex processors in larger 
networks. 
                                                
Dr. Thomas McKenna
Office of Naval Research
Div. Cognitive and Neural Sciences
Code 1142 Biological Intelligence
800 N. Quincy St.
Arlington, VA 22217-5000

phone:202-696-4503
email: mckenna at nprdc.arpa
       mckenna at nprdc.navy.mil



14.			  NEURAL NETWORK LEARNING:
	      MOVING FROM BLACK ART TO A PRACTICAL TECHNOLOGY


			      Scott E. Fahlman
			 School of Computer Science
			 Carnegie-Mellon University
			    Pittsburgh, PA 15213

			Internet: fahlman at cs.cmu.edu
			   Phone: (412) 268-2575

There are a number of competing algorithms for neural network learning, all
rather new and poorly understood.  Where theory is lacking, a reliable
technology can be built on shared experience, but it usually takes a long
time for this experience to accumulate and propagate through the community.
Currently, each research group has its own bag of tricks and its own body
of folklore about how to attack certain kinds of learning tasks and how to
diagnose the problem when things go wrong.  Even when groups are willing to
share their hard-won experience with others, this can be hard to accomplish.

This workshop will bring together experienced users of back-propagation and
other neural net learning algorithms, along with some interested novices,
to compare views on questions like the following:

I. Which algorithms and variations work best for various classes of
problems?  Can we come up with some diagnostic features that tell us what
techniques to try?  Can we predict how hard a given problem will be?

II. Given a problem, how do we go about choosing the parameters for various
algorithms?  How do we choose what size and shape of network to try?  If
our first attempt fails, are there symptoms that can tell us what to try
next?  

III. What can we do to bring more coherence into this body of folklore, and
facilitate communication of this informal kind of knowledge?  An online
collection of standard benchmarks and public-domain programs is one idea,
already implemented at CMU.  How can we improve this, and what other ideas
do we have?





More information about the Connectionists mailing list