Fwd: Reminder - Thesis Proposal - 1/12/18 - Manzil Zaheer - Representation Learning @ Scale
Artur Dubrawski
awd at cs.cmu.edu
Thu Jan 11 18:06:43 EST 2018
if you're free please attend Manzil's presentation
-------- Forwarded Message --------
Subject: Reminder - Thesis Proposal - 1/12/18 - Manzil Zaheer -
Representation Learning @ Scale
Date: Thu, 11 Jan 2018 17:09:00 -0500
From: Diane Stidle <diane+ at cs.cmu.edu>
To: ml-seminar at cs.cmu.edu <ML-SEMINAR at cs.cmu.edu>, Alex Smola
<alex.smola at gmail.com>, mccallum at cs.umass.edu <mccallum at cs.umass.edu>
/Thesis Proposal/
Date: January 12, 2018
Time: 3:00 PM
Place: 8102 GHC
Speaker: Manzil Zaheer
Title: Representation Learning @ Scale
Abstract:
Machine learning techniques are reaching or exceeding human level
performances in tasks like image classification, translation, and
text-to-speech. The success of these machine learning algorithms have
been attributed to highly versatile representations learnt from data
using deep networks or intricately designed Bayesian models*.
*_*Represen*_*tation learning* has also provided hints in neuroscience,
e.g. for understanding how humans might categorize objects. Despite
these instances of success, many open questions remain.
Data come in all shapes and sizes: not just as images or text, but also
as point clouds, sets, graphs, compressed, or even heterogeneous mixture
of these data types. In this thesis, we want to develop
representation learning algorithms for such unconventional data types by
leveraging their structure and establishing new mathematical properties.
Representations learned in this fashion were applied on diversedomains
and found to be competitive with task specific state-of-the-art methods.
Once we have the representations, in various applications its
interpretability is as crucial as its accuracy. Deep models often yield
better accuracy, but require a large number of parameters, often
notwithstanding the simplicity of the underlying data, rendering it
uninterpretable which is highly undesirable in tasks
like user modeling. On the other hand, Bayesian models produce sparse
discrete representations, easily amenable to human interpretation. In
this thesis, we want to explore methods *that are**capable of *
*learning *mixed representations retaining best of both the worlds. Our
experimental evaluations show that the proposed techniques compare
favorably with several state-of-the-art baselines.
Finally, one would want such interpretable representations to be
inferred from large-scale data, however, often there is a mismatch
between our computational resources and the statistical models. In this
thesis, we want to bridge this gap by solutions based on a combination
of modern computational techniques/data structures on one side and
modified statistical inference algorithms on the other. We introduce new
ways to parallelize, reduce look-ups, handle variable state space size,
and escape saddle points. On latent variable models, like latent
Dirichlet allocation (LDA), we find significant gains in performance.
To summarize, in this thesis, we want to explore three major aspects of
representation learning --- diversity: being able to handle different
types of data, interpretability: being accessible to and understandable
by humans, and scalablity: being able to process massive datasets in a
reasonable time and budget.
Thesis Committee:
Barnabas Poczos, Co-Chair
Ruslan Salakhutdinov, Co-Chair
Alexander J Smola (Amazon)
Andrew McCallum (UMass Amherst)
Link to proposal document:
http://manzil.ml/proposal.pdf
--
Diane Stidle
Graduate Programs Manager
Machine Learning Department
Carnegie Mellon University
diane at cs.cmu.edu
412-268-1299
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/autonlab-research/attachments/20180111/ffff3591/attachment.html>
More information about the Autonlab-research
mailing list