Articles on combining multiple models
Chris Merz
cmerz at saanen.ics.uci.edu
Thu Sep 17 12:48:43 EDT 1998
I am announcing the availability of several articles related to
classification and regression by combining models. The first list
below contains the titles, url's and FEATURES of each article. The
second list contains the abstracts.
Please contact me at cmerz at ics.uci.edu with any questions.
Thanks,
Chris Merz
================== Titles, URL's and *** FEATURES *** =================
1. My dissertation:
"Classification and Regression by Combining Models"
Merz, Christopher J. (1998)
http://www.ics.uci.edu/~cmerz/thesis.ps
*** DESCRIBES TWO ROBUST METHODS FOR COMBINING LEARNED MODELS ***
*** USING TECHNIQUES BASED ON SINGULAR VALUE DECOMPOSITION. ***
*** CONTAINS COMPREHENSIVE BACKGROUND AND SURVEY CHAPTERS. ***
2. Preprints of two accepted Machine Learning Journal articles:
"A Principal Components Approach to Combining Regression Estimates",
Merz, C. J., Pazzani, M. J. (1997)
To appear in the Special Issue of Machine Learning on Integrating
Multiple Learned Models.
http://www.ics.uci.edu/~cmerz/jr.html/mlj.pcr.ps
*** SHOWS HOW PCA MAY BE USED TO SYSTEMATICALLY EXPLORE ***
*** WEIGHT SETS WITH VARYING DEGREES OF REGULARIZATION. ***
"Using Correspondence Analysis to Combine Classifiers",
Merz, C. J. (1997)
To appear in the Special Issue of Machine Learning on Integrating
Multiple Learned Models.
http://www.ics.uci.edu/~cmerz/jr.html/mlj.scann.ps
*** SHOWS THAT THE SCANN METHOD COMBINES BOOSTED MODEL ***
*** SETS BETTER THAN BOOSTING DOES. ***
3. A bibtex file of the references in my dissertation survey:
http://www.ics.uci.edu/~cmerz/bib.html/survey.bib
*** COMPREHENSIVE BIBLIOGRAPHY - MANY ABSTRACTS INCLUDED ***
======================== Abstracts ==========================
1. "Classification and Regression by Combining Models"
Two novel methods for combining predictors are introduced in this
thesis; one for the task of regression, and the other for the task of
classification. The goal of combining the predictions of a set of
models is to form an improved predictor. This dissertation
demonstrates how a combining scheme can rely on the stability of the
consensus opinion and, at the same time, capitalize on the unique
contributions of each model.
An empirical evaluation reveals that the new methods consistently
perform as well or better than existing combining schemes for a
variety of prediction problems. The success of these algorithms is
explained empirically and analytically by demonstrating how they
adhere to a set of theoretical and heuristic guidelines.
A byproduct of the empirical investigation is the evidence that
existing combining methods fail to satisfy one or more of the
guidelines defined. The new combining approaches satisfy these
criteria by relying upon Singular Value Decomposition as a tool for
filtering out the redundancy and noise in the predictions of the learn
models, and for characterizing the areas of the example space where
each model is superior. The SVD-based representation used in the new
combining methods aids in avoiding sensitivity to correlated
predictions without discarding any learned models. Therefore, the
unique contributions of each model can still be discovered and
exploited. An added advantage of the combining algorithms derived in
this dissertation is that they are not limited to models generated by
a single algorithm; they may be applied to model sets generated by a
diverse collection of machine learning and statistical modeling
methods.
The three main contributions of this dissertation are:
1. The introduction of two new combining methods capable of
robustly combining classification and regression estimates, and
applicable to a broad range of model sets.
2. An in-depth analysis revealing how the new methods address the
specific problems encountered in combining multiple learned
models.
3. A detailed account of existing combining methods and an
assessment of where they fall short in the criteria for
combining approaches.
----------------
2. Preprints of two accepted Machine Learning Journal articles:
"A Principal Components Approach to Combining Regression Estimates"
Christopher J. Merz and Michael J. Pazzani
Abstract
The goal of combining the predictions of multiple learned
models is to form an improved estimator. A combining strategy must be
able to robustly handle the inherent correlation, or
multicollinearity, of the learned models while identifying the unique
contributions of each. A progression of existing approaches and their
limitations with respect to these two issues are discussed. A new
approach, PCR*, based on principal components regression is proposed
to address these limitations. An evaluation of the new approach on a
collection of domains reveals that 1) PCR* was the most robust
combining method, 2) correlation could be handled without eliminating
any of the learned models, and 3) the principal components of the
learned models provided a continuum of ``regularized'' weights from
which PCR* could choose.
"Using Correspondence Analysis to Combine Classifiers"
Christopher J. Merz
Abstract
Several effective methods have been developed recently for improving
predictive performance by generating and combining multiple learned
models. The general approach is to create a set of learned models
either by applying an algorithm repeatedly to different versions of
the training data, or by applying different learning algorithms to the
same data. The predictions of the models are then combined according
to a voting scheme. This paper focuses on the task of combining the
predictions of a set of learned models. The method described uses the
strategies of stacking and Correspondence Analysis to model the
relationship between the learning examples and their classification by
a collection of learned models. A nearest neighbor method is then
applied within the resulting representation to classify previously
unseen examples. The new algorithm does not perform worse than, and
frequently performs significantly better than other combining
techniques on a suite of data sets.
----------------
3. The bibtex file contains all of the references in my dissertation,
including the survey. I've managed to paste in the abstracts of
many of the articles. I am willing to update this bibliography
if any authors want to contribute references, abstracts and/or URL's.
More information about the Connectionists
mailing list