Robert Nowak - What Kinds of Functions Do Neural Networks Learn?
Presentation given by Robert Nowak on 13th October in the one world seminar on the mathematics of machine learning on the topic "What Kinds of Functions Do Neural Networks Learn?".
Abstract: Neural nets have made an amazing comeback during the past decade. Their empirical success has been truly phenomenal, but neural nets are poorly understood in a mathematical sense compared to classical methods like splines, kernels, and wavelets. This talk describes recent steps towards a mathematical theory of neural networks comparable to the foundations we have for classical nonparametric methods. Surprisingly, neural nets are minimax optimal in a wide variety of classical univariate function spaces, including those handled by splines and wavelets. In multivariate settings, neural nets are solutions to data-fitting problems cast in entirely new types of multivariate function spaces characterized through total variation (TV) measured in the Radon transform domain. Deep multilayer neural nets naturally represent compositions of functions in these Radon-TV spaces. This theory provides novel explanations for many notable empirical discoveries in deep learning and suggests new approaches to training neural networks.
This is joint work with Rahul Parhi.
Видео Robert Nowak - What Kinds of Functions Do Neural Networks Learn? канала One world theoretical machine learning
Abstract: Neural nets have made an amazing comeback during the past decade. Their empirical success has been truly phenomenal, but neural nets are poorly understood in a mathematical sense compared to classical methods like splines, kernels, and wavelets. This talk describes recent steps towards a mathematical theory of neural networks comparable to the foundations we have for classical nonparametric methods. Surprisingly, neural nets are minimax optimal in a wide variety of classical univariate function spaces, including those handled by splines and wavelets. In multivariate settings, neural nets are solutions to data-fitting problems cast in entirely new types of multivariate function spaces characterized through total variation (TV) measured in the Radon transform domain. Deep multilayer neural nets naturally represent compositions of functions in these Radon-TV spaces. This theory provides novel explanations for many notable empirical discoveries in deep learning and suggests new approaches to training neural networks.
This is joint work with Rahul Parhi.
Видео Robert Nowak - What Kinds of Functions Do Neural Networks Learn? канала One world theoretical machine learning
Показать
Комментарии отсутствуют
Информация о видео
14 октября 2021 г. 15:11:51
00:55:44
Другие видео канала
![Lukasz Szpruch - Mean-Field Neural ODEs, Relaxed Control and Generalization Errors](https://i.ytimg.com/vi/f9SUMJNeGEw/default.jpg)
![Matthew Colbrook - Smale’s 18th Problem and the Barriers of Deep Learning](https://i.ytimg.com/vi/OYMMgnaP7Kc/default.jpg)
![Micah Goldblum - Bridging the Gap between Deep Learning Theory and Practice](https://i.ytimg.com/vi/SCvpURzW-44/default.jpg)
![Yu Bai - How Important is the Train-Validation Split in Meta-Learning?](https://i.ytimg.com/vi/uGDZsDjXfMQ/default.jpg)
![Anna Korba - Kernel Stein Discrepancy Descent](https://i.ytimg.com/vi/TLDSVfm1Pyw/default.jpg)
![Anirbit Mukherjee - Provable Training of Neural Nets With One Layer of Activation](https://i.ytimg.com/vi/RxhipqdlFrM/default.jpg)
![Kevin Miller - Ensuring Exploration and Exploitation in Graph-Based Active Learning](https://i.ytimg.com/vi/ayU-FO-XcJY/default.jpg)
![Theo Bourdais - Computational Hypergraph Discovery, a Gaussian Process framework](https://i.ytimg.com/vi/XIz2Va_wXrc/default.jpg)
![Yaoqing Yang - Predicting & improving generalization by measuring loss landscapes & weight matrices](https://i.ytimg.com/vi/dDsW8H9i6GQ/default.jpg)
![Konstantinos Spiliopoulos - Mean field limits of neural networks: typical behavior and fluctuations](https://i.ytimg.com/vi/euNJpB4QAP8/default.jpg)
![Nadia Drenska - A PDE Interpretation of Prediction with Expert Advice](https://i.ytimg.com/vi/ISC_7YZXOq4/default.jpg)
![Matthias Ehrhardt - Bilevel Learning for Inverse Problems](https://i.ytimg.com/vi/_U5NsS6hWLw/default.jpg)
![Peter Richtarik - The Resolution of a Question Related to Local Training in Federated Learning](https://i.ytimg.com/vi/IhoonUttMg0/default.jpg)
![Marcus Hutter - Testing Independence of Exchangeable Random Variables](https://i.ytimg.com/vi/xb1ZYG4rP_Y/default.jpg)
![Yury Korolev - Approximation properties of two-layer neural networks with values in a Banach space](https://i.ytimg.com/vi/jhGtE20aiqg/default.jpg)
![Sophie Langer - Circumventing the curse of dimensionality with deep neural networks](https://i.ytimg.com/vi/FopfGs4LkcI/default.jpg)
![Stephan Mandt - Compressing Variational Bayes: From neural data compression to video prediction](https://i.ytimg.com/vi/ztx_KRnAtvk/default.jpg)
![Derek Driggs - Barriers to Deploying Deep Learning Models During the COVID-19 Pandemic](https://i.ytimg.com/vi/wIOhrCo8uto/default.jpg)
![Gal Vardi - Implications of the implicit bias in neural networks](https://i.ytimg.com/vi/t9ZRPpVKt4c/default.jpg)
![Ziwei Ji - The dual of the margin: improved analyses and rates for gradient descent’s implicit bias](https://i.ytimg.com/vi/VRf0c7NMp7g/default.jpg)
![Qi Lei - Predicting What You Already Know Helps: Provable Self-Supervised Learning](https://i.ytimg.com/vi/5A1iKkPv1Y0/default.jpg)