[Research] Auton Lab meeting: Wednesday April 21 at 12noon
Artur Dubrawski
awd at cs.cmu.edu
Fri Apr 16 17:17:50 EDT 2010
Dear Autonians,
The Lab meeting next week will involve practice talks by Liang and Yi
before their appearances at the upcoming SDM conference.
Abstracts are provided below.
Please come along and provide them with your feedback!
Time: Wednesday April 21, 12noon
Karen will confirm the location.
See you,
Artur
---
Liang:
Temporal Collaborative Filtering with Bayesian Probabilistic Tensor
Factorization
Real-world relational data are seldom stationary, yet traditional
collaborative filtering algorithms generally rely on this assumption.
Motivated by our sales prediction problem, we propose a factor-based
algorithm
that is able to take time into account. By introducing additional
factors for
time, we formalize this problem as a tensor factorization with a special
constraint on the time dimension. Further, we provide a fully Bayesian
treatment to avoid tuning parameters and achieve automatic model
complexity
control. To learn the model we develop an efficient sampling procedure
that is
capable of analysing large-scale data sets. This new algorithm, called
Bayesian Probabilistic Tensor Factorization (BPTF), is evaluated on
several real-world problems including sales prediction and movie
recommendation. Empirical results demonstrate the superiority of our
temporal
model.
---
Yi:
Learning Compressible Models.
In this paper, we study the combination of compression and L1-norm
regularization in a machine learning context: learning compressible
models. By including a compression operation into the L-1
regularization, the assumption on model sparsity is relaxed to
compressibility: model coefficients are compressed before being
penalized, and sparsity is achieved in a compressed domain rather than
the original space. We focus on the design of different compression
operations, by which we can encode various compressibility assumptions
and inductive biases, e.g., piecewise local smoothness, compacted energy
in the frequency domain, and semantic correlation. We show that use of a
compression operation provides an opportunity to leverage auxiliary
information from various sources, e.g., domain knowledge, coding
theories, unlabeled data. We conduct extensive experiments on
brain-computer interfacing, handwritten character recognition and text
classification. Empirical results show clear improvements in prediction
performance by including compression in L-1 regularization. We also
analyze the learned model coefficients under appropriate compressibility
assumptions, which further demonstrate the advantages of learning
compressible models instead of sparse models.
More information about the Autonlab-research
mailing list