Seminars and Colloquia by Series

Neural Networks with Inputs Based on Domain of Dependence and A Converging Sequence for Solving Conservation Laws

Series
Applied and Computational Mathematics Seminar
Time
Monday, February 28, 2022 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Haoxiang HuangGT

Recent research on solving partial differential equations with deep neural networks (DNNs) has demonstrated that spatiotemporal-function approximators defined by auto-differentiation are effective    for approximating nonlinear problems. However, it remains a challenge to resolve discontinuities in nonlinear conservation laws using forward methods with DNNs without beginning with part of the solution. In this study, we incorporate first-order numerical schemes into DNNs to set up the loss function approximator instead of auto-differentiation from traditional deep learning framework such as the TensorFlow package, thereby improving the effectiveness of capturing discontinuities in Riemann problems. We introduce a novel neural network method.  A local low-cost solution is first used as the input of a neural network to predict the high-fidelity solution at a space-time location. The challenge lies in the fact that there is no way to distinguish a smeared discontinuity from a steep smooth solution in the input, thus resulting in “multiple predictions” of the neural network. To overcome the difficulty, two solutions of the conservation laws from a converging sequence, computed from low-cost numerical schemes, and in a local domain of dependence of the space-time location, serve as the input. Despite smeared input solutions, the output provides sharp approximations to solutions containing shocks and contact surfaces, and the method is efficient to use, once trained. It works not only for discontinuities, but also for smooth areas of the solution, implying broader applications for other differential equations.

Coarse – Graining of stochastic system

Series
Applied and Computational Mathematics Seminar
Time
Monday, February 7, 2022 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Prof. Xingjie "Helen" LiUNC Charlotte


Efficient simulation of SDEs is essential in many applications, particularly for ergodic
systems that demand efficient simulation of both short-time dynamics and large-time
statistics. To achieve the efficiency, dimension reduction is often required in both space
and time. In this talk, I will talk about our recent work on both spatial and temporal
reductions.
For spatial dimension reduction, the Mori-Zwanzig formalism is applied to derive
equations for the evolution of linear observables of the Langevin dynamics for both
overdamped and general cases.
For temporal dimension reduction, we introduce a framework to construct inference-
based schemes adaptive to large time-steps (ISALT) from data, achieving a reduction in
time by several orders of magnitudes.
This is a joint work with Dr. Thomas Hudson from the University of Warwick, UK; Dr. Fei
Lu from the Johns Hopkins University and Dr Xiaofeng Felix Ye from SUNY at Albany.

How to Break the Curse of Dimensionality

Series
Applied and Computational Mathematics Seminar
Time
Monday, January 31, 2022 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Ming-Jun LaiUniversity of Georgia

We first review the problem of the curse of dimensionality when approximating multi-dimensional functions. Several approximation results from Barron, Petrushev,  Bach, and etc . will be explained. 

Then we present two approaches to break the curse of the dimensionality: one is based on probability approach explained in Barron, 1993 and the other one is based on a deterministic approach using the Kolmogorov superposition theorem.   As the Kolmogorov superposition theorem has been used to explain the approximation of neural network computation, I will use it to explain why the deep learning algorithm works for image classification.
In addition, I will introduce the neural network approximation based on higher order ReLU functions to explain the powerful approximation of multivariate functions using  deep learning algorithms with  multiple layers.

Non-Parametric Estimation of Manifolds from Noisy Data

Series
Applied and Computational Mathematics Seminar
Time
Monday, December 6, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Yariv AizenbudYale University
A common task in many data-driven applications is to find a low dimensional manifold that describes the data accurately. Estimating a manifold from noisy samples has proven to be a challenging task. Indeed, even after decades of research, there is no (computationally tractable) algorithm that accurately estimates a manifold from noisy samples with a constant level of noise.
 
In this talk, we will present a method that estimates a manifold and its tangent in the ambient space. Moreover, we establish rigorous convergence rates, which are essentially as good as existing convergence rates for function estimation.

Model-free Feature Screening and FDR Control with Knockoff Features

Series
Applied and Computational Mathematics Seminar
Time
Monday, November 29, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Yuan KeUniversity of Georgia

This paper proposes a model-free and data-adaptive feature screening method for ultra-high dimensional data. The proposed method is based on the projection correlation which measures the dependence between two random vectors. This projection correlation based method does not require specifying a regression model, and applies to data in the presence of heavy tails and multivariate responses. It enjoys both sure screening and rank consistency properties under weak assumptions.  A two-step approach, with the help of knockoff features, is advocated to specify the threshold for feature screening  such that the false discovery rate (FDR) is controlled under a pre-specified level. The proposed two-step approach enjoys both sure screening and FDR control simultaneously if the pre-specified FDR level is greater or equal to 1/s, where s is the number of active features.  The superior empirical performance of the proposed method is illustrated by simulation examples and real data applications. This is a joint work with Wanjun Liu, Jingyuan Liu and Runze Li.

Local and Optimal Transport Perspectives on Uncertainty Quantification

Series
Applied and Computational Mathematics Seminar
Time
Monday, November 22, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Dr. Amir SagivColumbia

Please Note: remote

In many scientific areas, deterministic models (e.g., differential equations) use numerical parameters. In real-world settings, however, such parameters might be uncertain or noisy. A more comprehensive model should therefore provide a statistical description of the quantity of interest. Underlying this computational problem is a fundamental question - if two "similar" functions push-forward the same measure, would the new resulting measures be close, and if so, in what sense? We will first show how the probability density function (PDF) of the quantity of interest can be approximated, using spectral and local methods. We will then discuss the limitations of PDF approximation, and present an alternative viewpoint: through optimal transport theory, a Wasserstein-distance formulation of our problem yields a much simpler and widely applicable theory.
 

Data Compression in Distributed Learning

Series
Applied and Computational Mathematics Seminar
Time
Monday, November 15, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Ming YanMichigan State University

Large-scale machine learning models are trained by parallel (stochastic) gradient descent algorithms on distributed systems. The communications for gradient aggregation and model synchronization become the major obstacles for efficient learning as the number of nodes and the model's dimension scale up. In this talk, I will introduce several ways to compress the transferred data and reduce the overall communication such that the obstacles can be immensely mitigated. More specifically, I will introduce methods to reduce or eliminate the compression error without additional communication.

Generalization Bounds for Sparse Random Feature Expansions

Series
Applied and Computational Mathematics Seminar
Time
Monday, November 8, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
https://bluejeans.com/457724603/4379
Speaker
Giang TranUniversity of Waterloo

Random feature methods have been successful in various machine learning tasks, are easy to compute, and come with theoretical accuracy bounds. They serve as an alternative approach to standard neural networks since they can represent similar function spaces without a costly training phase. However, for accuracy, random feature methods require more measurements than trainable parameters, limiting their use for data-scarce applications or problems in scientific machine learning. This paper introduces the sparse random feature expansion to obtain parsimonious random feature models. Specifically, we leverage ideas from compressive sensing to generate random feature expansions with theoretical guarantees even in the data-scarce setting. We provide generalization bounds for functions in a certain class (that is dense in a reproducing kernel Hilbert space) depending on the number of samples and the distribution of features. The generalization bounds improve with additional structural conditions, such as coordinate sparsity, compact clusters of the spectrum, or rapid spectral decay. We show that the sparse random feature expansions outperform shallow networks in several scientific machine learning tasks. Applications to signal decompositions for music data, astronomical data, and various complicated signals are also provided.

The Heavy-Tail Phenomenon in SGD

Series
Applied and Computational Mathematics Seminar
Time
Monday, October 18, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
Skiles 005 and https://bluejeans.com/457724603/4379
Speaker
Lingjiong ZhuFSU

Please Note: The speaker will be in person, but there will also be a remote option https://bluejeans.com/457724603/4379

In recent years, various notions of capacity and complexity have been proposed for characterizing the generalization properties of stochastic gradient descent (SGD) in deep learning. Some of the popular notions that correlate well with the performance on unseen data are (i) the flatness of the local minimum found by SGD, which is related to the eigenvalues of the Hessian, (ii) the ratio of the stepsize to the batch-size, which essentially controls the magnitude of the stochastic gradient noise, and (iii) the tail-index, which measures the heaviness of the tails of the network weights at convergence. In this paper, we argue that these three seemingly unrelated perspectives for generalization are deeply linked to each other. We claim that depending on the structure of the Hessian of the loss at the minimum, and the choices of the algorithm parameters, the distribution of the SGD iterates will converge to a heavy-tailed stationary distribution. We rigorously prove this claim in the setting of quadratic optimization: we show that even in a simple linear regression problem with independent and identically distributed data whose distribution has finite moments of all order, the iterates can be heavy-tailed with infinite variance. We further characterize the behavior of the tails with respect to algorithm parameters, the dimension, and the curvature. We then translate our results into insights about the behavior of SGD in deep learning. We support our theory with experiments conducted on synthetic data, fully connected, and convolutional neural networks. This is based on the joint work with Mert Gurbuzbalaban and Umut Simsekli.

High-Order Multirate Explicit Time-Stepping Schemes for the Baroclinic-Barotropic Split Dynamics in Primitive Equations

Series
Applied and Computational Mathematics Seminar
Time
Monday, October 4, 2021 - 14:00 for 1 hour (actually 50 minutes)
Location
online
Speaker
Lili JuUniversity of South Carolina

To treat the multiple time scales of ocean dynamics in an efficient manner, the baroclinic-barotropic splitting technique has been widely used for solving the primitive equations for ocean modeling. In this paper, we propose second and third-order multirate explicit time-stepping schemes for such split systems based on the strong stability-preserving Runge-Kutta (SSPRK) framework. Our method allows for a large time step to be used for advancing the three-dimensional (slow) baroclinic mode and a small time step for the two-dimensional (fast) barotropic mode, so that each of the two mode solves only need satisfy their respective CFL condition to maintain numerical stability. It is well known that the SSPRK method achieves high-order temporal accuracy by utilizing a convex combination of forward-Euler steps. At each time step of our method, the baroclinic velocity is first computed by using the SSPRK scheme to advance the baroclinic-barotropic system with the large time step, then the barotropic velocity is specially corrected by using the same SSPRK scheme with the small time step to advance the barotropic subsystem with a barotropic forcing interpolated based on values from the preceding baroclinic solves. Finally, the fluid thickness and the sea surface height perturbation is updated by coupling the predicted baroclinic and barotropic velocities. Two benchmark tests drawn from the ``MPAS-Ocean" platform are used to numerically demonstrate the accuracy and parallel performance of the proposed schemes.

 

The bluejeans link for the seminar is https://bluejeans.com/457724603/4379

Pages