- Series
- High-Dimensional Phenomena in Statistics and Machine Learning Seminar
- Time
- Tuesday, February 14, 2012 - 4:00pm for 1.5 hours (actually 80 minutes)
- Location
- Skyles 006
- Speaker
- Vladimir Koltchinskii – Georgia Institute of Technology, School of Mathematics
- Organizer
- Karim Lounici
Let (V, E) be a graph with vertex set V and edge set E. Let
(X, X', Y) V \in V × {-1,1} be a random triple, where X, X' are
independent uniformly distributed vertices and Y is a label
indicating whether X, X' are "similar", or not. Our goal is to
estimate the regression function
S_*(u, v) = \mathbb{E}(Y|X = u, X' = v), u, v \in V
based on n i.i.d. copies (X_1, X'_1, Y_1), ... , (X_n, X'_n, Y_n)
of (X, X', Y). We are interested in this problem in the case
when S_*: V × V \mapsto [-1,1] is a symmetric low rank kernel (or it can
be well approximated by low rank kernels). In addition to this,
assume that S_* is "smooth" on the graph. We study estimators based
on a modified least squares method with complexity penalization
involving both the nuclear norm and Sobolev type norms of symmetric
kernels on the graph (defined in terms of its Laplacian). We prove
error bounds for such estimators that improve the existing bounds
in low rank matrix recovery in the cases when the target kernel is
not only low rank, but also sufficiently smooth.
The talk is based in part on a joint project with Pedro Rangel.