Seminars and Colloquia by Series

Limit results for distributed estimation of invariant subspaces in multiple networks inference and PCA

Series
Stochastics Seminar
Time
Thursday, September 28, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Minh TangNC State

We study the problem of estimating the left and right singular subspaces for a collection of heterogeneous random graphs with a shared common structure. We analyze an algorithm that first estimates the orthogonal projection matrices corresponding to these subspaces for each individual graph, then computes the average of the projection matrices, and finally finds the matrices whose columns are the eigenvectors corresponding to the d largest eigenvalues of the sample averages. We show that the algorithm yields an estimate of the left and right singular vectors whose row-wise fluctuations are normally distributed around the rows of the true singular vectors. We then consider a two-sample hypothesis test for the null hypothesis that two graphs have the same edge probabilities matrices against the alternative hypothesis that their edge probabilities matrices are different. Using the limiting distributions for the singular subspaces, we present a test statistic whose limiting distribution converges to a central chi-square (resp. non-central chi-square) under the null (resp. alternative) hypothesis. Finally, we adapt the theoretical analysis for multiple networks to the setting of distributed PCA; in particular, we derive normal approximations for the rows of the estimated eigenvectors using distributed PCA when the data exhibit a spiked covariance matrix structure.

Eigenvalues of fractional Brownian matrix process

Series
Stochastics Seminar
Time
Tuesday, September 26, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Victor Pérez-AbreuCIMAT

This talk will present an overview of the behavior of the eigenvalues of the fractional Brownian matrix motion and other related matrix processes. We will do so by emphasizing the dynamics of the eigenvalues processes, the non-colliding property, the limit of the associated empirical process, as well as the free Brownian motion and the non commutative fractional Brownian motion.

Curie-Weiss Model under $\ell^{p}$ constraint

Series
Stochastics Seminar
Time
Thursday, September 21, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Daesung KimGeorgia Tech

We consider the Ising Curie-Weiss model on the complete graph constrained under a given $\ell_{p}$ norm. For $p=\infty$, it reduces to the classical Ising Curie-Weiss model. We prove that for all $p\ge 2$, there exists a critical inverse temperature $\beta_{c}(p)$ such that for $\beta<\beta_{c}(p)$, the magnetization is concentrated at zero and satisfies an appropriate Gaussian CLT. On the other hand, for $\beta>\beta_{c}(p)$, the magnetization is not concentrated at zero similar to the classical case. We further generalize the model for general symmetric spin distributions and prove similar phase transition. In this talk, we discuss a brief overview of classical Curie-Weiss model, a generalized Hubbard-Stratonovich transforms, and how we apply the transform to Curie-Weiss model under $\ell^p$ constraint. This is based on joint work with Partha Dey.

Spectral clustering in the geometric block model

Series
Stochastics Seminar
Time
Thursday, September 7, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Shuangping LiStanford

Gaussian mixture block models are distributions over graphs that strive to model modern networks: to generate a graph from such a model, we associate each vertex with a latent feature vector sampled from a mixture of Gaussians, and we add edge if and only if the feature vectors are sufficiently similar. The different components of the Gaussian mixture represent the fact that there may be different types of nodes with different distributions over features---for example, in a social network each component represents the different attributes of a distinct community. Natural algorithmic tasks associated with these networks are embedding (recovering the latent feature vectors) and clustering (grouping nodes by their mixture component).

In this talk, we focus on clustering and embedding graphs sampled from high-dimensional Gaussian mixture block models, where the dimension of the latent feature vectors goes to infinity as the size of the network goes to infinity. This high-dimensional setting is most appropriate in the context of modern networks, in which we think of the latent feature space as being high-dimensional. We analyze the performance of canonical spectral clustering and embedding algorithms for such graphs in the case of 2-component spherical Gaussian mixtures and begin to sketch out the information-computation landscape for clustering and embedding in these models.

This is based on joint work with Tselil Schramm.

On the geometry of polytopes generated by heavy-tailed random vectors

Series
Stochastics Seminar
Time
Friday, September 1, 2023 - 13:00 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Felix KrahmerTechnical University of Munich

In this talk, we present recent results on the geometry of centrally-symmetric random polytopes generated by N independent copies of a random vector X. We show that under minimal assumptions on X, for N>Cn, and with high probability, the polytope contains a deterministic set that is naturally associated with the random vector - namely, the polar of a certain floating body. This solves the long-standing question on whether such a random polytope contains a canonical body. Moreover, by identifying the floating bodies associated with various random vectors we recover the estimates that have been obtained previously, and thanks to the minimal assumptions on X we derive estimates in cases that had been out of reach, involving random polytopes generated by heavy-tailed random vectors (e.g., when X is q-stable or when X has an unconditional structure). Finally, the structural results are used for the study of a fundamental question in compressive sensing - noise blind sparse recovery. This is joint work with Olivier Guédon (University of Paris-Est Marne La Vallée), Christian Kümmerle (UNC Charlotte), Shahar Mendelson (Sorbonne University Paris), and Holger Rauhut (LMU Munich).

Bio: Felix Krahmer received his PhD in Mathematics in 2009 from New York University under the supervision of Percy Deift and Sinan Güntürk. He was a Hausdorff postdoctoral fellow in the group of Holger Rauhut at the University of Bonn, Germany from 2009-2012. In 2012 he joined the University of Göttingen as an assistant professor for mathematical data analysis, where he has been awarded an Emmy Noether Junior Research Group. From 2015-2021 he was assistant professor for optimization and data analysis in the department of mathematics at the Technical University of Munich, before he was tenured and promoted to associate professor in 2021. His research interests span various areas at the interface of probability, analysis, machine learning, and signal processing including randomized sensing and reconstruction, fast random embeddings, quantization, and the computational sensing paradigm.

(Skew) Gaussian surrogates for high-dimensional posteriors: tighter bounds and tighter approximations

Series
Stochastics Seminar
Time
Thursday, August 31, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Anya KatsevichMIT

Computing integrals against a high-dimensional posterior is the major computational bottleneck in Bayesian inference. A popular technique to reduce this computational burden is to use the Laplace approximation, a Gaussian distribution, in place of the true posterior. Despite its widespread use, the Laplace approximation's accuracy in high dimensions is not well understood.  The body of existing results does not form a cohesive theory, leaving open important questions e.g. on the dimension dependence of the approximation rate. We address many of these questions through the unified framework of a new, leading order asymptotic decomposition of high-dimensional Laplace integrals. In particular, we (1) determine the tight dimension dependence of the approximation error, leading to the tightest known Bernstein von Mises result on the asymptotic normality of the posterior, and (2) derive a simple correction to this Gaussian distribution to obtain a higher-order accurate approximation to the posterior.

Quantitative Generalized CLT with Self-Decomposable Limiting Laws by Spectral Methods

Series
Stochastics Seminar
Time
Thursday, May 18, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Benjamin ArrasUniversité de Lille

In this talk, I will present new stability results for non-degenerate centered self-decomposable laws with finite second moment and for non-degenerate symmetric alpha-stable laws with alpha in (1,2). These stability results are based on Stein's method and closed forms techniques. As an application, explicit rates of convergence are obtained for several instances of the generalized CLTs. Finally, I will discuss the standard Cauchy case.

Coalescence, geodesic density, and bigeodesics in first-passage percolation

Series
Stochastics Seminar
Time
Thursday, April 20, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Jack HansonCity College, CUNY

Several well-known problems in first-passage percolation relate to the behavior of infinite geodesics: whether they coalesce and how rapidly, and whether doubly infinite "bigeodesics'' exist. In the plane, a version of coalescence of "parallel'' geodesics has previously been shown; we will discuss new results that show infinite geodesics from the origin have zero density in the plane. We will describe related forthcoming work showing that geodesics coalesce in dimensions three and higher, under unproven assumptions believed to hold below the model's upper critical dimension. If time permits, we will also discuss results on the bigeodesic question in dimension three and higher.

Random Laplacian Matrices

Series
Stochastics Seminar
Time
Thursday, April 13, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
Skiles 006
Speaker
Andrew CampbellUniversity of Colorado

The Laplacian of a graph is a real symmetric matrix given by $L=D-A$, where $D$ is the degree matrix of the graph and $A$ is the adjacency matrix. The Laplacian is a central object in spectral graph theory, and the spectrum of $L$ contains information on the graph. In the case of a random graph the Laplacian will be a random real symmetric matrix with dependent entries. These random Laplacian matrices can be generalized by taking $A$ to be a random real symmetric matrix and $D$ to be a diagonal matrix with entries equal to the row sums of $A$. We will consider the eigenvalues of general random Laplacian matrices, and the challenges raised by the dependence between $D$ and $A$. After discussing the bulk global eigenvalue behavior of general random Laplacian matrices, we will focus in detail on fluctuations of the largest eigenvalue of $L$ when $A$ is a matrix of independent Gaussian random variables. The asymptotic behavior of these Gaussian Laplacian matrices has a particularly nice free probabilistic interpretation, which can be exploited in the study of their eigenvalues. We will see how this interpretation can locate the largest eigenvalue of $L$ with respect to the largest entry of $D$. This talk is based on joint work with Kyle Luh and Sean O'Rourke.

Stein kernels, functional inequalities and applications in statistics

Series
Stochastics Seminar
Time
Thursday, April 6, 2023 - 15:30 for 1 hour (actually 50 minutes)
Location
ONLINE via Zoom https://gatech.zoom.us/j/94387417679
Speaker
Adrien SaumardENSAI and CREST

Zoom link to the talk: https://gatech.zoom.us/j/94387417679

We will present the notion of Stein kernel, which provides generalizations of the integration by parts, a.k.a. Stein's formula, for the normal distribution (which has a constant Stein kernel, equal to its covariance). We will first focus on dimension one, where under good conditions the Stein kernel has an explicit formula. We will see that the Stein kernel appears naturally as a weighting of a Poincaré type inequality and that it enables precise concentration inequalities, of the Mills' ratio type. In a second part, we will work in higher dimensions, using in particular Max Fathi's construction of a Stein kernel through the so-called "moment maps" transportation. This will allow us to describe the performance of some shrinkage and thresholding estimators, beyond the classical assumption of Gaussian (or spherical) data. This presentation is mostly based on joint works with Max Fathi, Larry Goldstein, Gesine Reinert and Jon Wellner.

Pages