Fwd: Thesis Oral -Dougal J. Sutherland - Today, 2pm!
Jeff Schneider
schneide at cs.cmu.edu
Wed Sep 14 13:43:28 EDT 2016
Please come and listen to Dougal's defense!
---------- Forwarded message ---------
From: Debbie Cavlovich <deb at cs.cmu.edu <mailto:deb at cs.cmu.edu>>
Date: Wed, Sep 7, 2016, 9:11 AM
Subject: Thesis Oral -Dougal J. Sutherland - September 14, 2016
To: <cs-friends at cs.cmu.edu <mailto:cs-friends at cs.cmu.edu>>,
<cs-students at cs.cmu.edu <mailto:cs-students at cs.cmu.edu>>,
<cs-visitors at cs.cmu.edu <mailto:cs-visitors at cs.cmu.edu>>,
<cs-advisors-ext at cs.cmu.edu <mailto:cs-advisors-ext at cs.cmu.edu>>, Catherine
Copetas <copetas+ at cs.cmu.edu <mailto:copetas%2B at cs.cmu.edu>>
CALENDAR OF EVENTS
Dougal J. Sutherland
September 14, 2016
2:00 PM - GHC 6501
Thesis Oral
Title: Scalable, Flexible, and Active Learning on Distributions
Abstract:
A wide range of machine learning problems, including astronomical
inference about galaxy clusters, natural image scene classification,
parametric statistical inference, and detection of potentially harmful
sources of radioactive, can be well-modeled as learning a function on
(samples from) distributions. This thesis explores problems in learning
such functions via kernel methods, and applies the framework to yield
state-of-the-art results in several novel settings.
One major challenge with this approach is one of computational
efficiency when learning from large numbers of distributions: the
computation of typical methods scales between quadratically and
cubically, and so they are not amenable to large datasets. As a
solution, we investigate approximate embeddings into Euclidean spaces
such that inner products in the embedding space approximate kernel
values between the source distributions. We provide a greater
understanding of the standard existing tool for doing so on Euclidean
inputs, random Fourier features. We also present a new embedding for a
class of information-theoretic distribution distances, and evaluate it
and existing embeddings on several real-world applications.
The next challenge is that the choice of distance is important for
getting good practical performance, but how to choose a good distance
for a given problem is not obvious. We study this problem in the setting
of two-sample testing, where we attempt to distinguish two distributions
via the maximum mean divergence, and provide a new technique for kernel
choice in these settings, including the use of kernels defined by deep
learning-type models.
In a related problem setting, common to physical observations,
autonomous sensing, and electoral polling, we have the following
challenge: when observing samples is expensive, but we can choose where
we would like to do so, how do we pick where to observe? We give a
method for a closely related problem where we search for instances of
patterns by making point observations.
Throughout, we combine theoretical results with extensive empirical
evaluations to increase our understanding of the methods.
Thesis Committee:
Jeff Schneider, Chair
Barnabás Póczos
Maria-Florina Balcan
Arthur Gretton, University College London
More information about the Autonlab-research
mailing list