Announcement of NIPS Bayesian workshop and associated ftp archive

David MacKay mackay at hope.caltech.edu
Tue Nov 5 13:20:59 EST 1991


One of the two day workshops at Vail this year will be:

   `Developments in Bayesian methods for neural networks'
   ------------------------------------------------------
        David MacKay and Steve Nowlan, organizers

   The first day of this workshop will be 50% tutorial in content, 
reviewing some new ways Bayesian methods may be applied to neural
networks. 
   The rest of the workshop will be devoted to discussions of 
the frontiers and challenges facing Bayesian work in neural networks. 
   Participants are encouraged to obtain preprints by anonymous ftp 
before the workshop. Instructions end this message.  
   Discussion will be moderated by John Bridle. 

Day 1, Morning: Tutorial review.

   0   Introduction to Bayesian data modelling.          David MacKay
   1   E-M, clustering and mixtures.                     Steve Nowlan
   2   Bayesian model comparison and determination of 
        regularization constants - 
        application to regression networks.              David MacKay
   3   The use of mixture decay schemes in backprop 
        networks.                                        Steve Nowlan

Day 1, Evening: Tutorial continued.

   4   The `evidence' framework for classification 
        networks.                                        David MacKay

Day 1, Evening: Frontier Discussion.

Background:
   A:
   In many cases the true Bayesian posterior distribution over a hypothesis
   or parameter space is difficult to obtain analytically. Monte Carlo
   methods may provide a useful and computationally efficient way to estimate
   posterior distributions in such cases.

   B:
   There are many applications where training data is expensive to obtain,
   and it is desirable to select training examples so we can learn as much as
   possible from each one. This session will discuss approaches for selecting 
   the next training point "optimally". The same approaches may also be 
   useful for reducing the size of a large data set by omitting the 
   uninformative data points.    

   A   Monte Carlo clustering                           Radford Neal
   B   Data selection / active query learning           Jurgen Schmidhuber
                                                        David MacKay

Day 2, morning discussion:

   C   Prediction of generalisation

Background:
    The Bayesian approach to model comparison evaluates 
    how PROBABLE alternative models are given the data. 
    In contrast, the real problem is often to estimate 
    HOW WELL EACH MODEL IS EXPECTED TO GENERALISE. 
    In this session we will hear about various approaches 
    to predicting generalisation. 
    It is hoped that the discussion will shed light on the questions:
     -   How does Bayesian model comparison relate to generalisation?
     -   Can we predict generalisation ability of one model assuming
            that the `truth' is in a different specified model class?
     -   Is it possible to predict generalisation ability WITHOUT 
            making implicit assumptions about the properties 
            of the `truth'?
     -   Can we interpret GCV (cross-validation) in terms of prediction
            of generalisation?

   1   Prediction of generalisation with `GPE'  John Moody 
   2   Prediction of generalisation 
         - worst + average case analysis        David Haussler + Michael Kearns
   3   News from the statistical physics front  Sara Solla 


Day 2, Evening discussion:
   (Note: There will probably be time in this final session for continued
        discussion from the other sessions.)

   D   Missing inputs, unlabelled data and discriminative training

Background:
       When training a classifier with a data set D_1 = {x,t},
    a full probability model is one which assigns a
    parameterised probability P(x,t|w). However, many classifiers
    only produce a discriminant P(t|x,w), ie they do not model P(x).
    Furthermore, classifiers of the first type often yield better
    discriminative performance if they are trained as if they were only of
    the second type. This is called `discriminative training'. The problem
    with discriminative training is that it leaves us with no obvious
    way to use UNLABELLED data D_2 = {x}. Such data is usually cheap,
    but how can we integrate it with discriminative training?
       The same problem arises for most regression or classifier
    models when some of the input variables are missing from the input
    vector. What is the right thing to do?

   1   Introduction: the problem of combining unlabelled
           data and discriminative training            Steve Renals
   2   Combining labelled and unlabelled data
           for the modem problem                       Steve Nowlan


Reading up before the workshop
------------------------------
People intending to attend this workshop are encouraged to obtain 
preprints of relevant material before NIPS. A selection of preprints 
are available by anonymous ftp, as follows: 

unix> ftp hope.caltech.edu                (or ftp 131.215.4.231)
Name: anonymous
Password: <your name>
ftp> cd pub/mackay
ftp> get README.NIPS
ftp> quit

Then read the file README.NIPS for further information. 
Problems? Contact David MacKay, mackay at hope.caltech.edu, 
or Steve Nowlan, nowlan at helmholtz.sdsc.edu   
---------------------------------------------------------------------------


More information about the Connectionists mailing list