Seminars and Colloquia by Series

Optimal Estimation of Low Rank Density Matrices

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, November 10, 2015 - 15:00 for 1.5 hours (actually 80 minutes)
Location
Skiles 005
Speaker
Dong XiaGeorgia Inst. of Technology, School of Mathematics

Please Note: Joint work with Vladimir Koltchinskii.

The density matrices are positively semi-definite Hermitian matrices of unit trace that describe the state of a quantum system. We develop minimax lower bounds on error rates of estimation of low rank density matrices in trace regression models used in quantum state tomography (in particular, in the case of Pauli measurements) with explicit dependence of the bounds on the rank and other complexity parameters.Such bounds are established for several statistically relevant distances, including quantum versions of Kullback-Leibler divergence (relative entropy distance) and of Hellinger distance (so called Bures distance), and Schatten p-norm distances. Sharp upper bounds and oracle inequalities for least squares estimator with von Neumann entropy penalization are obtained showing that minimax lower bounds are attained (up to logarithmic factors) for these distances.

Generalized Dantzig Selector: Application to the k-support norm

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, October 27, 2015 - 15:00 for 1.5 hours (actually 80 minutes)
Location
Skiles 249
Speaker
Changong LiGeorgia Inst. of Technology, School of Mathematics

Please Note: Review of a recent paper by Chatterjee et al. (Arxiv 1406.5291)

We propose a Generalized Dantzig Selector (GDS) for linear models, in which any norm encoding the parameter structure can be leveraged for estimation. We investigate both computational and statistical aspects of the GDS. Based on conjugate proximal operator, a flexible inexact ADMM framework is designed for solving GDS, and non-asymptotic high-probability bounds are established on the estimation error, which rely on Gaussian width of unit norm ball and suitable set encompassing estimation error. Further, we consider a non-trivial example of the GDS using k-support norm. We derive an efficient method to compute the proximal operator for k-support norm since existing methods are inapplicable in this setting. For statistical analysis, we provide upper bounds for the Gaussian widths needed in the GDS analysis, yielding the first statistical recovery guarantee for estimation with the k-support norm. The experimental results confirm our theoretical analysis.

Perturbation of linear forms of singular vectors under Gaussian noise

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, October 20, 2015 - 15:00 for 1.5 hours (actually 80 minutes)
Location
Skiles 005
Speaker
Dong XiaGeorgia Inst. of Technology
Let A be a mxn matrix with singular value decomposition A=UDV', where the columns of U are left singular vectors and columns of V are right singular vectors of A. Suppose X is a mxn noise matrix whose entries are i.i.d. Gaussian random variables and consider A'=A+X. Let u_k be the k-th left singular vector of A and u'_k be its counterpart of A'. We develop sharp upper bounds for concentration of linear forms for the right singular vectors of A'.The talk is based on a joint work with Vladimir Koltchinskii.

The Gaussian Radon Transform for Infinite-Dimensional Banach Spaces

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, September 23, 2014 - 15:00 for 1 hour (actually 50 minutes)
Location
Skiles 170
Speaker
Irina HolmesSchool of Mathematics, Georgia Tech
In this talk we construct an infinite-dimensional, stochastic version of the Radon transform. We work within the framework of abstract Wiener spaces, introduced by Leonard Gross. We present some basic properties of this transform, as well as compute some specific examples on the classical Wiener space.

Variable Selection Consistency of Linear Programming Discriminant Estimator

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, September 9, 2014 - 15:00 for 1 hour (actually 50 minutes)
Location
Skiles 005
Speaker
Dong XiaSchool of Mathematics, Georgia Tech
The linear programming discriminant(LPD) estimator is used in sparse linear discriminant analysis for high dimensional classification problems. In this talk we will give a sufficient condition for the variable selection property of the LPD estimator and our result provides optimal bound on the requirement of sample size $n$ and magnitude of components of Bayes direction.

Rademacher Averages and Phase Transitions in Glivenko–Cantelli Classes

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, October 23, 2012 - 15:05 for 1.5 hours (actually 80 minutes)
Location
Skyles 005
Speaker
Krishnakumar BalasubramanianGeorgia Institute of Technology
I will be presenting the paper by S. Mendelson titled 'Rademacher Averages and Phase Transitions in Glivenko–Cantelli Classes'. Fat-shattering dimension and its use in characterizing GC classes will be introduced. Then a new quantity based on the growth rate of the Rademacher averages will be introduced. This parameter enables one to provide improved complexity estimates for the agnostic learning problem with respect to any norm. A phase transition phenomenon that appears in the sample complexity estimates, covering numbers estimates, and in the growth rate of the Rademacher averages will be discussed. Further (i) estimates on the covering numbers of a class when considered as a subset of spaces and (ii) estimate the fat-shattering dimension of the convex hull of a given class will be discussed.

Statistical Algorithms and a Lower Bound for Detecting a Planted Clique

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, October 2, 2012 - 15:05 for 1.5 hours (actually 80 minutes)
Location
Skyles 005
Speaker
Santosh VempalaGeorgia Institute of Technology
We present a framework for proving lower bounds on computational problems over distributions, including optimization and unsupervised learning. The framework is based on defining a restricted class of algorithms, called statistical algorithms, that instead of directly accessing samples from the input distribution can only obtain an estimate of the expectation of any given function on the input distribution. Our definition captures many natural algorithms used in theory and practice, e.g., moment-based methods, local search, linear programming, MCMC and simulated annealing. Our techniques are inspired by the statistical query model of Kearns from learning theory, which addresses the complexity of PAC-learning. For specific well-known problems over distributions, we obtain lower bounds on the complexity of any statistical algorithm. These include an exponential lower bounds for moment maximization and a nearly optimal lower bound for detecting planted clique distributions when the planted clique has size n^{1/2-\delta} for any constant \delta > 0. Variants of the latter problem have been assumed to be hard to prove hardness for other problems and for cryptographic applications. Our lower bounds provide concrete evidence of hardness. This is joint work with V. Feldman, E. Grigorescu, L. Reyzin and Y. Xiao.

Generic Chaining

Series
High-Dimensional Phenomena in Statistics and Machine Learning Seminar
Time
Tuesday, September 4, 2012 - 15:05 for 1.5 hours (actually 80 minutes)
Location
Skyles 005
Speaker
William MantzelSchool of Electrical and Computer Engineering, Georgia Tech
Recap of generic chaining from last time and more discussion about it. Then, the lower Dudley bound (Theorem 2.1.1) and the Bernoulli upper bound (4.1.2) and statement of the Bernoulli conjecture (lower bound) will be covered from The Generic Chaining book.

Pages