PhD thesis available

Nick Adams nicka at dai.ed.ac.uk
Tue Jan 8 11:00:00 EST 2002


I am pleased to announce that my PhD thesis entitled

 DYNAMIC TREES: A HIERARCHICAL PROBABILISTIC APPROACH TO IMAGE MODELLING 

is now available at 

  http://www.anc.ed.ac.uk/code/adams/

Matlab and C++ code implementing Gibbs Sampling, Metropolis and the Mean
Field Variational approaches and various EM-style learning algorithms
based upon them for the Dynamic Tree model is also available at

  http://www.anc.ed.ac.uk/code/adams/dt/dt.tgz


----------------------------------------------------------------------

ABSTRACT

This work introduces a new class of image model which we call
Dynamic Trees or DTs. A Dynamic Tree model specifies a prior
over structures of trees, each of which is a forest of one or more
tree-structured belief networks (TSBN). In the literature standard
tree-structured belief network models were found to produce ``blocky''
segmentations when naturally occurring boundaries within an image did
not coincide with those of the subtrees in the rigid fixed structure
of the network. Dynamic Trees have a flexible architecture which
allows the structure to vary to accommodate configurations where the
subtree and image boundaries align, and experimentation with the model
showed significant improvements. They are also hierarchical in nature
allowing a multi-scale representation and are constructed within a
well founded Bayesian framework.

For large models the number of tree configurations quickly becomes
intractable to enumerate over, presenting a problem for exact
inference. Techniques such as Gibbs sampling over trees are considered
and search using simulated annealing finds high posterior probability
trees on synthetic 2-d images generated from the model. However
simulated annealing and sampling techniques are rather
slow. Variational methods are applied to the model in an attempt to
approximate the posterior by a simpler tractable distribution, and the
simplest of these techniques, mean field, found comparable solutions
to simulated annealing in the order of 100 times faster. This
increase in speed goes a long way towards making real-time inference
in the Dynamic Tree viable. Variational methods have the further
advantage that by attempting to model the full posterior distribution
it is possible to gain an indication as to the quality of the
solutions found.

An EM-style update based upon mean field inference is derived and the
learned conditional probability tables (describing state transitions
between a node and its parent) are compared with exact EM on small
tractable fixed architecture models. The mean field approximation by
virtue of its form is biased towards fully factorised solutions which
tends to create degenerate CPTs, but despite this mean field learning
still produces solutions whose log-likelihood rivals exact EM.

Development of algorithms for learning the probabilities of the prior
over tree structures completes the Dynamic Tree picture. After
discussion of the relative merits of certain representations for the
disconnection probabilities and initial investigation on small model
structures the full Dynamic Tree model is applied to a database of
images of outdoor scenes where all of its parameters are learned.
DTs are seen to offer significant improvement in performance over the 
fixed architecture TSBN and in a coding comparison the DT achieves
0.294 bits per pixel (bpp) compression compared to 0.378 bpp for
lossless JPEG on images of 7 colours.




More information about the Connectionists mailing list