papers/neuroprose: Unifying MLP and RBFN
Georg Dorffner
georg at ai.univie.ac.at
Mon Sep 20 10:10:37 EDT 1993
FTP-host: archive.cis.ohio-state.edu
FTP-file: pub/neuroprose/dorffner.csfn.ps.Z pub/neuroprose/dorffner.nn-clinical.ps.Z
The files dorffner.csfn.ps.Z and dorffner.nn-clinical.ps.Z are now
available for copying from the Neuroprose repository:
-------------------------------------------------------------------------
dorffner.csfn.ps.Z:
A Unified Framework for MLPs and RBFNs:
Introducing Conic Section Function Networks
Georg Dorffner
Austrian Research Inst. for Artificial Intelligence
and
Dept. of Medical Cybernetics and AI
Univ. of Vienna
georg at ai.univie.ac.at
ABSTRACT: Multilayer Perceptrons (MLP, Werbos 1974, Rumelhart et al.
1986) and Radial Basis Function Networks (RBFN, Broomhead & Lowe 1988,
Moody & Darken 1989) probably are the most widely used neural network
models for practical applications. While the former belong to a group
of ``classical'' neural networks (whose weighted sums are loosely
inspired by biology), the latter have risen only recently from an
analogy to regression theory (Broomhead & Lowe 1988). On first sight,
the two models -- except for being multilayer feedforward networks --
do not seem to have much in common. On second thought, however,
MLPs and RBFNs share a variety of features, worthy of viewing them
in the same context and comparing them to each other with respect to
their properties. Consequently, a few attempts on arriving at a
unified picture of a class of feedforward networks -- with MLPs and
RBFNs as members -- have been undertaken (Robinson et al. 1988,
Maruyama et al. 1992, Dorffner 1992, 1993). Most of these attempts
have centered around the observation that the function of a neural
network unit can be divided into a propagation rule (``net input'')
and an activation or transfer function. The dot product (``weighted
sum'') and the Euclidean distance are special cases of propagation
rules, whereas the sigmoid and Gaussian function are examples for
activation functions. This paper introduces a novel neural network
model based on a more general conic section function as propagation
rule, containing hyperplane (straight line) and hypersphere (circle)
as special cases, thus unifying the net inputs of MLPs and RBFNs with
an easy-to-handle continuum in between. A new learning rule --
complementing the existing methods of gradient descent in weight
space and initialization -- is introduced which enables the network
to make a continuous decision between bounded and unbounded
(infinite half-space) decision regions. The capabilities of CSFNs
are illustrated with several examples and compared with exisiting
approaches. CSFNs are viewed as a further step toward more efficient
and optimal neural network solutions in practical applications.
length: 37 pages.
submitted for publication
-------------------------------------------------------------------------
dorffner.nn-clinical.ps.Z:
On Using Feedforward Neural Networks for Clinical Diagnostic Tasks
Georg Dorffner
Austrian Research Inst. for Artificial Intelligence
and
Dept. of Medical Cybernetics and AI
Univ. of Vienna
georg at ai.univie.ac.at
and
Georld Porenta
Dept. of Cardiology
Clinic for Internal Medicine II
University of Vienna
ABSTRACT: In this paper we present an extensive comparison between
several feedforward neural network types in the context of a clinical
diagnostic task, namely the detection of coronary artery disease
(CAD) using planar thallium-201 dipyridamole stress-redistribution
scintigrams. We introduce results from well-known (e.g. multilayer
perceptrons or MLPs, and radial basis function networks or RBFNs) as
well as novel neural network techniques (e.g. conic section
function networks) which demonstrate promising new routes for future
applications of neural networks in medicine, and elsewhere. In particular
we show that initializations of MLPs and conic section function networks
-- which can learn to behave more like an MLP or more like an RBFN --
can lead to much improved results in rather difficult diagnostic tasks.
Keywords: Feedforward neural networks, neural network
initialization, multilayer perceptrons, radial basis function networks,
conic section function networks; thallium scintigraphy, angiography,
clinical diagnosis and decision making.
length: 21 pages
submitted for publication
-----------------------------------------------------------------------
To obtain a copy:
ftp cheops.cis.ohio-state.edu
login: anonymous
password: <your email address>
cd pub/neuroprose
binary
get dorffner.csfn.ps.Z
AND/OR
get dorffner.nn-clinical.ps.Z
quit
Then at your system
uncompress dorffner.*
to obtain (a) postscript file(s).
Many thanks to Jordan Pollack for the maintenance and support of this
archive.
-----------------------------------------------------------------------
both papers are also available through anonymous ftp from
ftp.ai.univie.ac.at in the directory 'papers' as
oefai-tr-93-23.ps.Z (== dorffner.nn-clinical) and oefai-tr-93-25.ps.Z
(== dorffner.csfn)
Hardcopies are available (only if you don't have access to ftp!) by
sending email to sec at ai.univie.ac.at and asking for technical report
oefai-tr-93-23 or oefai-tr-93-25 (see previous paragraph).
More information about the Connectionists
mailing list