doctoral thesis on automatic structuring available by ftp

Uli Bodenhausen uli at ira.uka.de
Fri Feb 10 11:10:18 EST 1995


The following doctoral thesis is available by ftp. 
Sorry, no hardcopies available.

ftp://archive.cis.ohio-state.edu/pub/neuroprose/Thesis/bodenhausen.thesis.ps.Z


FTP-host: archive.cis.ohio-state.edu
FTP-file: pub/neuroprose/Thesis/bodenhausen.thesis.Z

 -----------------------------------------------------------------------------

        	Automatic Structuring of Neural Networks for
        	  Spatio-Temporal Real-World Applications
	
			    (153 pages)

			 Ulrich Bodenhausen

			  Doctoral Thesis

		   University of Karlsruhe, Germany


      			     Abstract

The successful application of speech recognition (SR) and on-line 
handwriting recognition (OLHR) systems to new domains greatly depend 
on the tuning of a recognizer's architecture to a new task. Architectural 
tuning is especially important if the amount of training data is small 
because the amount of training data limits the number of trainable 
parameters that can be estimated properly using an automatic learning 
algorithm. The number of trainable parameters of a connectionist SR 
or OLHR is dependent on architectural parameters like the width of 
input windows over time, the number of hidden units and the number of 
state units. Each of these architectural parameters provides different 
functionality in the system and can not be optimized independently. 
Manual optimization of these architectural parameters is time-consuming 
and expensive. Automatic optimization algorithms can free the developer 
of SR and OLHR applications from this task. 

In this thesis I develop and evaluate novel methods that allocate 
connectionist resources for spatio-temporal classification problems 
automatically. The methods are evaluated under the following evaluation 
criteria:

- Suitability for small systems (~ 1,000 parameters) as well as for large 
  systems (more than 10,000 parameters): Is the proposed method efficient 
  for various sizes of the system?

- Ease of use for non-expert users: How much knowledge is necessary to 
  adapt the system to a customized application?

- Final performance: Can the automatically optimized system compete with 
  state-of-the-art well engineered systems?

Several algorithms were developed and evaluated in this thesis. The 
Automatic Structure Optimization (ASO) algorithm performed best under the 
above criteria. ASO automatically optimizes 

- the width of the input windows over time which allow the following unit 
  of the neural network to capture a certain amount of temporal context of 
  the input signal;

- the number of hidden units which allow the neural network to learn 
  non-linear classification boundaries;

- the number of states that are used to model segments of the spatio-
  temporal input, such as acoustic segments of speech or strokes of 
  on-line handwriting.

The ASO algorithm uses a constructive approach to find the best 
architecture. Training starts with a neural network of minimum size. 
Resources are added to specifically improve parts of the network which 
are involved in classification errors. ASO was developed on the recognition 
of spoken letters and improved the performance on an independent test set 
from 88.0% to 92.2% over a manually tuned architecture. The performances 
of architectures found by ASO for different domains and databases are also 
compared to architectures optimized manually by other researchers. For 
example, ASO improved the performance on on-line handwritten digits from 
98.5% to 99.5% over a manually optimized architecture. It is also shown 
that ASO can successfully adapt to different sizes of the training database 
and that it can be applied to the recognition of connected spoken letters.

The ASO algorithm is applicable to all classification problems with 
spatio-temporal input. It was tested on speech and on-line handwriting, 
as two instances of such tasks. The approach is new, requires no domain 
specific knowledge by the user and is efficient. It is shown for the 
first time that fully automatic tuning of all relevant architectural
parameters of speech and on-line handwriting recognizers (window widths, 
number of hidden units and states) to the domain and the available 
amount of training data is actually possible with the ASO algorithm 
automatic tuning by ASO is efficient, both in terms of computational 
effort and final performance.


------------------------------------------------------------------------

Instructions for ftp retrieval of this paper are given below. Our university
requires that the title page is in German. The rest of the thesis is English.

FTP INSTRUCTIONS:

unix> ftp archive.cis.ohio-state.edu (or 128.146.8.52)
    Name: anonymous
    Password: <your e-mail address>
    ftp> cd pub/neuroprose/Thesis
    ftp> binary
    ftp> get bodenhausen.thesis.Z
    ftp> quit
unix> uncompress bodenhausen.thesis.Z

Thanks to Jordan Pollack for maintaining this archive. 

Uli Bodenhausen

=======================================================================
Uli Bodenhausen
University of Karlsruhe
Germany
uli at ira.uka.de
=======================================================================


More information about the Connectionists mailing list