<div dir="ltr">Team,<div><br></div><div>If you can, come see Junier becoming a doctor.</div><div><br></div><div>It is today at 10 in Gates 8102.</div><div><br></div><div>Cheers</div><div>Artur</div><div><br></div><div><br></div><div><br></div><div><br><div class="gmail_quote">---------- Forwarded message ----------<br>From: <b class="gmail_sendername">Diane Stidle</b> <span dir="ltr"><<a href="mailto:stidle@andrew.cmu.edu">stidle@andrew.cmu.edu</a>></span><br>Date: Thu, May 24, 2018 at 3:29 PM<br>Subject: Reminder - Thesis Defense - 5/25/18 - Junier Oliva - Distribution and Histogram (DisH) Learning<br>To: "<a href="mailto:ml-seminar@cs.cmu.edu">ml-seminar@cs.cmu.edu</a>" <<a href="mailto:ML-SEMINAR@cs.cmu.edu">ML-SEMINAR@cs.cmu.edu</a>>, Le Song <<a href="mailto:lsong@cc.gatech.edu">lsong@cc.gatech.edu</a>><br><br><br>
<div text="#000000" bgcolor="#FFFFFF">
<p>Thesis Defense</p>
<p>Date: May 25, 2018<br>
Time: 10:00am<br>
Place: 8102 GHC<br>
PhD Candidate: Junier Oliva</p>
<p>Title: Distribution and Histogram (DisH) Learning</p>
<p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:0pt" id="m_-5677237024966799674docs-internal-guid-227d75e8-5075-9006-8a95-b5df73f1ed23">Abstract:<span style="font-size:11pt;font-family:Arial;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre-wrap;white-space:pre-wrap">
Machine learning has made incredible advances in the last couple of decades. Notwithstanding, a lot of this progress has been limited to basic point-estimation tasks. That is, a large bulk of attention has been geared at solving problems that take in a static finite vector and map it to another static finite vector. However, we do not navigate through life in a series of point-estimation problems, mapping x to y. Instead, we find broad patterns and gather a far-sighted understanding of data by considering collections of points like sets, sequences, and distributions. Thus, contrary to what various billionaires, celebrity theoretical physicists, and sci-fi classics would lead you to believe, true machine intelligence is fairly out of reach currently. In order to bridge this gap, we have developed algorithms that understand data at an aggregate, holistic level.
</span></p>
<p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:0pt"><span style="font-size:11pt;font-family:Arial;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre-wrap;white-space:pre-wrap">This thesis pushes machine learning past the realm of operating over static finite vectors, to start reasoning ubiquitously with complex, dynamic collections like sets and sequences. We develop algorithms that consider distributions as functional covariates/responses, and methods that use distributions as internal representations. We consider distributions since they are a straightforward characterization of many natural phenomena and provide a richer description than simple point data by detailing information at an aggregate level. Our approach may be seen as addressing two sides of the same coin: on one side, we use traditional machine learning algorithms adjusted to directly operate on inputs and outputs that are probability functions (and sample sets); on the other side, we develop better estimators for traditional tasks by making use of and adjusting internal distributions.
</span></p>
<p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:0pt"><span style="font-size:11pt;font-family:Arial;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre-wrap;white-space:pre-wrap">We begin by developing algorithms for traditional machine learning tasks for the cases when one’s input (and/or possibly output) is not a finite point, but is instead a distribution, or sample set drawn from a distribution. We develop a scalable nonparametric estimator for regressing a real valued response given an input that is a distribution, a case which we coin distribution to real regression (DRR). Furthermore, we extend this work to the case when both the output response and the input covariate are distributions; a task we call distribution to distribution regression (DDR).
</span></p>
<p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:0pt"><span style="font-size:11pt;font-family:Arial;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre-wrap;white-space:pre-wrap">After, we look to expand the versatility and efficacy of traditional machine learning tasks through novel methods that operate with distributions of features. For example, we show that one may improve the performance of kernel learning tasks by learning a kernel’s spectral distribution in a data-driven fashion using Bayesian nonparametric techniques. Moreover, we study how to perform sequential modeling by looking at summary statistics from past points. Lastly, we also develop methods for high-dimensional density estimation that make use of flexible transformations of variables and autoregressive conditionals.</span></p>
<p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:0pt"><span style="font-size:11pt;font-family:Arial;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre-wrap;white-space:pre-wrap">
</span></p>
<div><br>
Thesis Committee:</div>
<div>Barnabas Poczos (Co-Chair)</div>
<div>Jeff Schneider (Co-Chair)</div>
<div>Ruslan Salakhutdinov</div>
<div>Le Song (Georgia Institute of Technology, <a href="mailto:lsong@cc.gatech.edu" target="_blank">lsong@cc.gatech.edu</a>)</div>
<div><br>
</div>
<div>Link to draft document:<br>
</div>
<a href="https://www.dropbox.com/s/z93s3qanl02fs8l/draft.pdf?dl=0" target="_blank">https://www.dropbox.com/s/<wbr>z93s3qanl02fs8l/draft.pdf?dl=0</a><span class="HOEnZb"><font color="#888888">
<p dir="ltr" style="line-height:1.38;margin-top:0pt;margin-bottom:0pt"><span style="font-size:11pt;font-family:Arial;color:#000000;background-color:transparent;font-weight:400;font-style:normal;font-variant:normal;text-decoration:none;vertical-align:baseline;white-space:pre-wrap;white-space:pre-wrap"></span></p>
<pre class="m_-5677237024966799674moz-signature" cols="72">--
Diane Stidle
Graduate Programs Manager
Machine Learning Department
Carnegie Mellon University
<a class="m_-5677237024966799674moz-txt-link-abbreviated" href="mailto:diane@cs.cmu.edu" target="_blank">diane@cs.cmu.edu</a>
412-268-1299</pre>
</font></span></div>
</div><br></div></div>