<div><span class="" style="font-family:arial,sans-serif;font-size:13px">Reminder</span><span style="font-family:arial,sans-serif;font-size:13px">: today</span><br></div><br><div class="gmail_extra"><div class="gmail_quote">
<blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left-width:1px;border-left-color:rgb(204,204,204);border-left-style:solid;padding-left:1ex"><div style="font-family:arial,sans-serif;font-size:13px">===================================================</div>
<div style="font-family:arial,sans-serif;font-size:13px"><b>CL+NLP Lunch</b> (<a href="http://www.cs.cmu.edu/~nlp-lunch/" target="_blank">http://www.cs.cmu.edu/~nlp-lunch/</a>)</div>
<div style="font-family:arial,sans-serif;font-size:13px"><b>Speaker</b>: Dirk Hovy, University of Southern California</div><div style="font-family:arial,sans-serif;font-size:13px"><b>Date</b>: Tuesday, December 4, 2012</div>
<div style="font-family:arial,sans-serif;font-size:13px"><b>Time</b>: 12:00 noon</div><div style="font-family:arial,sans-serif;font-size:13px"><b>Venue</b>: GHC 6115</div><div style="font-family:arial,sans-serif;font-size:13px">
<br></div><div style="font-family:arial,sans-serif;font-size:13px"><b>Title</b>:<br>Learning Semantic Types and Relations from Syntactic Context<br><br><b>Abstract</b>:<br>Natural Language Processing (NLP) is moving towards incorporating more semantics into its applications, such as Machine Translation (MT) or Question Answering (QA). Most semantic frameworks depend on predefined "external" resources, such as knowledge bases or ontologies. This requires a lot of manual effort, but even then it is impossible to build a complete representation of the world. Instead, we would like to learn a sufficient representation directly from data.</div>
<div style="font-family:arial,sans-serif;font-size:13px"><br>One way to encode meaning is through syntactic and semantic relations between predicates and their arguments. Relations are thus at the core of meaning and information. Recently, several approaches have collected large corpora of syntactically related word chains (e.g., subject, verb, object). However, these chains are extracted at the lexical level and do not generalize well, so their use for semantic interpretation is limited. If we could use these lexical chains as inputs to generalize beyond the word level, we would be able to learn semantic relations and make use of these existing resources.<br>
<br></div><div style="font-family:arial,sans-serif;font-size:13px">In this talk, I will present a method to learn semantic types and relations from raw text.<br>I construct unsupervised models on syntactic dependency arcs, using potential types as latent variables. The resulting models allow for quick domain adaptation and unknown relations, and avoid data sparsity caused by intervening words.</div>
<div style="font-family:arial,sans-serif;font-size:13px">I show improvements over state-of-the-art systems as well as novel approaches to fully exploit the structure contained in the data. My method builds on existing triple stores and does not require any external knowledge bases, manual annotation, or pre-defined predicates or arguments.<br>
<br><b>Bio</b>:<br>Dirk Hovy is a PhD candidate in Natural Language Processing at the University of Southern California (USC), and holds a Masters in linguistics. He is interested in data-driven models of meaning and understanding and has published on unsupervised learning, information extraction, word-sense disambiguation, and temporal relation classification (see <<a href="http://www.dirkhovy.com/portfolio/papers/index.php" target="_blank">http://www.dirkhovy.com/portfolio/papers/index.php</a>>).<br>
His thesis work is concerned with how computers can learn semantic types and relations from raw text, without recurrence to external resources.<br>His other interests include baking, cooking, crossfit, and medieval art and literature.</div>
</blockquote></div><br></div>