[Research] Fwd: Thesis Defense - Ajit Singh - 9/30/09

Ajit Paul Singh ajit at cs.cmu.edu
Sun Sep 27 23:09:44 EDT 2009


I'll be defending this Wednesday at noon. Everyone's welcome to  
attend (NSH 1507).

-- Ajit

Begin forwarded message:

> From: Diane Stidle <diane+ at cs.cmu.edu>
> Date: September 17, 2009 1:52:23 PM CDT
> To: ml-seminar at cs.cmu.edu
> Subject: Thesis Defense - Ajit Singh - 9/30/09
>
> Thesis Defense
>
> Date: 9/30/09
> Time: 12:00pm
> Place: 1507 NSH
>
> PhD Candidate: Ajit Singh
>
> Title: *Efficient Models for Relational Learning
> *
> Abstract:
> Information integration deals with the setting where one has  
> multiple sources of data, each describing different properties of  
> the same set of entities. We are concerned primarily with settings  
> where the properties are pairwise relations between entities, and  
> attributes of entities. We want to predict the value of relations  
> and attributes, but relations between entities violate the basic  
> statistical assumption of exchangeable data points, or entities.  
> Furthermore, we desire models that scale gracefully as the number  
> of entities and relations increase.
>
> Matrices are the simplest form of relational data; and we begin by  
> distilling the literature on low-rank matrix factorization into a  
> small number of modelling choices. We then frame information  
> integration as simultaneously factoring sets of related matrices:  
> i.e., Collective Matrix Factorization. Each entity is described by  
> a small number of parameters, and if an entity is described by more  
> than one matrix, those parameters participate in multiple matrix  
> factorizations. Maximum likelihood estimation of the resulting  
> model involves a large non-convex optimization, which we reduce to  
> cyclically solving convex optimizations over small subsets of the  
> parameters. Each convex subproblem can be solved by Newton-Raphson,  
> which we extend to the setting of stochastic Newton-Raphson.
>
> To address the limitations of maximum likelihood estimation in  
> matrix factorization models, we extend our approach to the  
> hierarchical Bayesian setting. Here, Bayesian estimation involves  
> computing a high-dimensional integral with no analytic form. If we  
> resorted to standard Metropolis-Hastings techniques, slow mixing  
> would limit the scalability of our approach to large sets of  
> entities. We show how to accelerate Metropolis-Hastings by using  
> our efficient solution for maximum likelihood estimation to guide  
> the sampling process.
>
> This thesis rests on two claims, that (i) that Collective Matrix  
> Factorization can effectively integrate different sources of data  
> to improve prediction; and, (ii) that training scales well as the  
> number of entities and observations increase. Two real-world data  
> sets are considered in experimental support of these claims:  
> augmented collaborative filtering and augmented brain imaging. In  
> augmented collaborative filtering, we show that genre information  
> about movies can be used to increase the predictive accuracy of  
> user's ratings. In augmented brain imaging, we show that word co- 
> occurrence information can be used to increase the predictive  
> accuracy of a model of changes in brain activity to word stimuli,  
> even in regions of the brain that were never included in the  
> training data.
>
> http://www.cs.cmu.edu/~ajit/pubs/ajit-thesis-submitted.pdf
>
> Thesis Committee:
> Geoff Gordon (chair)
> Tom Mitchell
> Christos Faloutsos
> Pedro Domingos (University of Washington)
>
> -- 
> *******************************************************************
> Diane Stidle
> Business & Graduate Programs Manager
> Machine Learning Department
> School of Computer Science
> 8203 Gates Hillman Complex			
> Carnegie Mellon University		
> 5000 Forbes Avenue		
> Pittsburgh, PA  15213-3891
> Phone: 412-268-1299
> Fax:   412-268-3431
> Email: diane at cs.cmu.edu	
> URL:http://www.ml.cmu.edu			
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mailman.srv.cs.cmu.edu/mailman/private/autonlab-research/attachments/20090927/1a9d846c/attachment.html>


More information about the Autonlab-research mailing list