Seminars and Colloquia by Series

Inference, Computation, and Games

Series
Applied and Computational Mathematics Seminar
Time
Monday, September 20, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
Skiles 005 and https://bluejeans.com/457724603/4379
Speaker
Florian SchaeferGT CSE

Please Note: Note the hybrid mode. The speaker will be in person in Skiles 005.

In this talk, we develop algorithms for numerical computation, based on ideas from competitive games and statistical inference. 

 

In the first part, we propose competitive gradient descent (CGD) as a natural generalization of gradient descent to saddle point problems and general sum games. Whereas gradient descent minimizes a local linear approximation at each step, CGD uses the Nash equilibrium of a local bilinear approximation. Explicitly accounting for agent-interaction significantly improves the convergence properties, as demonstrated in applications to GANs, reinforcement learning, and computer graphics.

 

In the second part, we show that the conditional near-independence properties of smooth Gaussian processes imply the near-sparsity of Cholesky factors of their dense covariance matrices. We use this insight to derive simple, fast solvers with state-of-the-art complexity vs. accuracy guarantees for general elliptic differential- and integral equations. Our methods come with rigorous error estimates, are easy to parallelize, and show good performance in practice.

Incorporating Symmetry for Improved Deep Dynamics Learning

Series
Applied and Computational Mathematics Seminar
Time
Monday, September 13, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Prof. Rose YuUCSD

While deep learning has been used for dynamics learning, limited physical accuracy and an inability to generalize under distributional shift limit its applicability to real world. In this talk, I will demonstrate how to incorporate symmetries into deep neural networks and significantly improve the physical consistency, sample efficiency, and generalization in learning dynamics. I will showcase the applications of these models to challenging problems such as turbulence forecasting and trajectory prediction for autonomous vehicles.

Some problems in point-set registration

Series
Applied and Computational Mathematics Seminar
Time
Monday, April 26, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/884917410
Speaker
Yuehaw KhooUniversity of Chicago

In this talk, we discuss variants of the rigid registration problem, i.e aligning objects via rigid transformation. In the simplest scenario of point-set registration where the correspondence between points are known, we investigate the robustness of registration to outliers. We also study a convex programming formulation of point-set registration with exact recovery, in the situation where both the correspondence and alignment are unknown. This talk is based on joint works with Ankur Kapoor, Cindy Orozco, and Lexing Ying.
 

Symmetrically processed splitting integrators for enhanced Hamiltonian Monte Carlo sampling

Series
Applied and Computational Mathematics Seminar
Time
Monday, April 19, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
ONLINE https://bluejeans.com/884917410
Speaker
Prof. Sergio BlanesUniversidad Politécnica de Valencia

We construct integrators to be used in Hamiltonian (or Hybrid) Monte Carlo sampling. The new integrators are easily implementable and, for a given computational budget, may deliver five times as many accepted proposals as standard leapfrog/Verlet without impairing in any way the quality of the samples. They are based on a suitable modification of the   processing technique first introduced by J.C. Butcher. The idea of modified processing may also be useful for other purposes, like the construction of high-order splitting integrators with positive coefficients.

Joint work with Mari Paz Calvo, Fernando Casas, and Jesús M. Sanz-Serna

Optimization in the space of probabilities with MCMC: Uncertainty quantification and sequential decision making

Series
Applied and Computational Mathematics Seminar
Time
Monday, April 5, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
ONLINE https://bluejeans.com/884917410
Speaker
Prof. Yian MaUCSD

I will present MCMC algorithms as optimization over the KL-divergence in the space of probabilities. By incorporating a momentum variable, I will discuss an algorithm which performs accelerated gradient descent over the KL-divergence. Using optimization-like ideas, a suitable Lyapunov function is constructed to prove that an accelerated convergence rate is obtained. I will then discuss how MCMC algorithms compare against variational inference methods in parameterizing the gradient flows in the space of probabilities and how it applies to sequential decision making problems.

Incorporating Invariance to Reduce the Complexity of Parametric Models

Series
Applied and Computational Mathematics Seminar
Time
Monday, March 29, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/884917410
Speaker
Alex CloningerUniversity of California, San Diego

Many scientific problems involve invariant structures, and learning functions that rely on a much lower dimensional set of features than the data itself.   Incorporating these invariances into a parametric model can significantly reduce the model complexity, and lead to a vast reduction in the number of labeled examples required to estimate the parameters.  We display this benefit in two settings.  The first setting concerns ReLU networks, and the size of networks and number of points required to learn certain functions and classification regions.  Here, we assume that the target function has built in invariances, namely that it only depends on the projection onto a very low dimensional, function defined manifold (with dimension possibly significantly smaller than even the intrinsic dimension of the data).  We use this manifold variant of a single or multi index model to establish network complexity and ERM rates that beat even the intrinsic dimension of the data.  We should note that a corollary of this result is developing intrinsic rates for a manifold plus noise data model without needing to assume the distribution of the noise decays exponentially, and we also discuss implications in two-sample testing and statistical distances.  The second setting for building invariances concerns linearized optimal transport (LOT), and using it to build supervised classifiers on distributions.  Here, we construct invariances to families of group actions (e.g., shifts and scalings of a fixed distribution), and show that LOT can learn a classifier on group orbits using a simple linear separator.   We demonstrate the benefit of this on MNIST by constructing robust classifiers with only a small number of labeled examples.  This talk covers joint work with Timo Klock, Xiuyuan Cheng, and Caroline Moosmueller.
 

Group Synchronization via Cycle-Edge Message Passing

Series
Applied and Computational Mathematics Seminar
Time
Monday, March 8, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/884917410
Speaker
Gilad LermanUniversity of Minnesota

The problem of group synchronization asks to recover states of objects associated with group elements given possibly corrupted relative state measurements (or group ratios) between pairs of objects. This problem arises in important data-related tasks, such as structure from motion, simultaneous localization and mapping, Cryo-EM, community detection and sensor network localization. Two common groups in these problems are the rotation and symmetric groups. We propose a general framework for group synchronization with compact groups. The main part of the talk discusses a novel message passing procedure that uses cycle consistency information in order to estimate the corruption levels of group ratios. Under our mathematical model of adversarial corruption, it can be used to infer the corrupted group ratios and thus to solve the synchronization problem. We first explain why the group cycle consistency information is essential for effectively solving group synchronization problems. We then establish exact recovery and linear convergence guarantees for the proposed message passing procedure under a deterministic setting with adversarial corruption. We also establish the stability of the proposed procedure to sub-Gaussian noise. We further establish competitive theoretical results under a uniform corruption model. Finally, we discuss the MPLS (Message Passing Least Squares) or Minneapolis framework for solving real scenarios with high levels of corruption and noise and with nontrivial scenarios of corruption. We demonstrate state-of-the-art results for two different problems that occur in structure from motion and involve the rotation and permutation groups.

Monte Carlo methods for the Hermitian eigenvaue problem

Series
Applied and Computational Mathematics Seminar
Time
Monday, January 25, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
ONLINE https://bluejeans.com/884917410
Speaker
Robert WebberCourant Institute

In quantum mechanics and the analysis of Markov processes, Monte Carlo methods are needed to identify low-lying eigenfunctions of dynamical generators. The standard Monte Carlo approaches for identifying eigenfunctions, however, can be inaccurate or slow to converge. What limits the efficiency of the currently available spectral estimation methods and what is needed to build more efficient methods for the future? Through numerical analysis and computational examples, we begin to answer these questions. We present the first-ever convergence proof and error bounds for the variational approach to conformational dynamics (VAC), the dominant method for estimating eigenfunctions used in biochemistry. Additionally, we analyze and optimize variational Monte Carlo (VMC), which combines Monte Carlo with neural networks to accurately identify low-lying eigenstates of quantum systems.

Time-parallel wave propagation in heterogeneous media aided by deep learning

Series
Applied and Computational Mathematics Seminar
Time
Monday, November 23, 2020 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/884917410
Speaker
Richard TsaiUT Austin

 

We present a deep learning framework for learning multiscale wave propagation in heterogeneous media. The framework involves the construction of linear feed-forward networks (experts) that specialize in different media groups and a nonlinear "committee" network that gives an improved approximation of wave propagation in more complicated media.  The framework is then applied to stabilize the "parareal" schemes of Lions, Maday, and Turinici, which are time-parallelization schemes for evolutionary problems. 

Theoretical guarantees of machine learning methods for statistical sampling and PDEs in high dimensions

Series
Applied and Computational Mathematics Seminar
Time
Monday, November 2, 2020 - 16:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/884917410
Speaker
Yulong LuUniversity of Massachusetts Amherst

Neural network-based machine learning methods, inlcuding the most notably deep learning have achieved extraordinary successes in numerious  fields. In spite of the rapid development of learning algorithms based on neural networks, their mathematical analysis are far from understood. In particular, it has been a big mystery that neural network-based machine learning methods work extremely well for solving high dimensional problems.

In this talk, I will demonstrate the power of  neural network methods for solving two classes of high dimensional problems: statistical sampling and PDEs. In the first part of the talk, I will present a universal approximation theorem of deep neural networks for representing high dimensional probability distributions. In the second part of the talk, I will discuss a generalization error bound of the Deep Ritz Method for solving high dimensional elliptic problems. For both problems,  our theoretical results show that neural networks-based methods  can overcome the curse of dimensionality.

Pages