Thesis available: Optimal Linear Combinations of Neural Networks

Sherif Hashem vg197 at neutrino.pnl.gov
Mon May 9 20:24:40 EDT 1994


FTP-host: archive.cis.ohio-state.edu
FTP-file: pub/neuroprose/Thesis/hashem.thesis.ps.Z

The file hashem.thesis.ps.Z is now available for copying from the
Neuroprose archive:

      
          OPTIMAL LINEAR COMBINATIONS OF NEURAL NETWORKS

                         Sherif Hashem

                          Ph.D. Thesis
                        Purdue University
      
ABSTRACT: Neural network (NN) based modeling often involves trying multiple
networks with different architectures, learning techniques, and
training parameters in order to achieve ``acceptable'' model accuracy.
Typically, one of the trained networks is chosen as ``best,'' while
the rest are discarded.

In this dissertation, using optimal linear combinations (OLCs) of the 
corresponding outputs of a number of NNs is proposed as an alternative to 
using a single network.  Modeling accuracy is measured by mean squared 
error (MSE) with respect to the distribution of random inputs to the NNs.
Optimality is defined by minimizing the MSE, with the resultant combination
referred to as MSE-OLC.

MSE-OLCs are investigated for four cases: allowing (or not) a constant
term in the combination and requiring (or not) the combination-weights to 
sum to one.  In each case, deriving the MSE-OLC is straightforward and the 
optimal combination-weights are simple, requiring modest matrix 
manipulations.

In practice, the optimal combination-weights need to be estimated from 
observed data: observed inputs, the corresponding true responses, and the
corresponding outputs of each component network.  Given the data, 
estimating the optimal combination-weights is straightforward.

Collinearity among the outputs and/or the approximation errors of the 
component NNs sometimes degrades the generalization ability of the 
estimated MSE-OLC.  To improve generalization in the presence of degrading 
collinearity, six algorithms for selecting subsets of the NNs for the 
MSE-OLC are developed and tested.

Several examples, including a real-world problem and an empirical
study, are discussed.  The examples illustrate the importance of
addressing collinearity and demonstrate significant improvements in model 
accuracy as a result of employing MSE-OLCs supported by the NN selection 
algorithms.

---------------------------

The thesis is 126 Pages (10 preamble + 116 text). 

To obtain a copy of the Postscript file:

  %ftp archive.cis.ohio-state.edu
  >Name: anonymous
  >Password: <Your email address>
  >cd pub/neuroprose/Thesis
  >binary
  >get hashem.thesis.ps.Z
  >quit

Then:
 
  %uncompress hashem.thesis.ps.Z  
            (The size of the uncompressed file is about 1.1Mbyte)
  %lpr -s -P<Printer-name> hashem.thesis.ps

---------------------------

Hard copies may be requested from the School of Industrial Engineering,
1287 Grissom Hall, Purdue University, West Lafayette, IN 47907-1287, USA. 
(Refer to Technical Report SMS 94-4.)


--Sherif Hashem

===================================================================
  Pacific Northwest Laboratory        E-mail: s_hashem at pnl.gov
  906 Battelle Boulevard              Tel. (509) 375-6995
  P.O. Box 999, MSIN K1-87            Fax. (509) 375-6631
  Richland, WA 99352
  USA
===================================================================


More information about the Connectionists mailing list