Mathematical theory of structured deep neural networks

Series
Applied and Computational Mathematics Seminar
Time
Monday, April 28, 2025 - 2:00pm for 1 hour (actually 50 minutes)
Location
Skiles 005
Speaker
Ding-Xuan Zhou – School of Mathematics and Statistics, University of Sydney, Australia – dingxuan.zhou@sydney.edu.au
Organizer
Haomin Zhou

Deep learning has been widely applied and brought breakthroughs in speech recognition, computer vision, natural language processing, and many other domains. The involved deep neural network architectures and computational issues have been well studied in machine learning. But there is much less theoretical understanding about the modelling, approximation or generalization abilities of deep learning models with network architectures. An important family of structured deep neural networks is deep convolutional neural networks (CNNs) induced by convolutions. The convolutional architecture gives essential differences between deep CNNs and fully-connected neural networks, and the classical approximation theory for fully-connected networks developed around 30 years ago does not apply.  This talk describes approximation and generalization analysis of deep CNNs and related structured deep neural networks.