Tess Smidt - Symmetry’s made to be broken: Learning how to break symmetry with symmetry-preservin...
Prof. Tess Smidt of the MIT speaking in the UW Data-driven methods in science and engineering seminar on January 6, 2023.
Sign up for notifications of future talks: https://mailman11.u.washington.edu/mailman/listinfo/datadriven-seminar
Title: Symmetry’s made to be broken: Learning how to break symmetry with symmetry-preserving neural networks
Abstract: Symmetry-preserving (equivariant) neural networks are extremely data-efficient and generalize well when applied to diverse domains (e.g. computer vision and atomic systems). But, are there circumstances where equivariance is too strict or otherwise undesirable? In this talk, I’ll discuss how symmetry-preserving neural networks can learn symmetry-breaking information in order to fit a dataset that may have missing information unbeknownst to the researcher. Furthermore, due to the mathematical guarantees of equivariant neural networks, these learned parameters are guaranteed to be minimally symmetry breaking. I’ll describe network architectures that can learn these symmetry-breaking parameters in two distinct settings: 1) symmetry-breaking parameters are learned for an entire dataset to capture global asymmetries in the data and 2) symmetry-breaking parameters are predicted in an equivariant manner for individual examples. Finally, I’ll demonstrate these networks on several prototypical examples and apply them to predicting structural distortions of crystalline materials.
Видео Tess Smidt - Symmetry’s made to be broken: Learning how to break symmetry with symmetry-preservin... канала Physics Informed Machine Learning
Sign up for notifications of future talks: https://mailman11.u.washington.edu/mailman/listinfo/datadriven-seminar
Title: Symmetry’s made to be broken: Learning how to break symmetry with symmetry-preserving neural networks
Abstract: Symmetry-preserving (equivariant) neural networks are extremely data-efficient and generalize well when applied to diverse domains (e.g. computer vision and atomic systems). But, are there circumstances where equivariance is too strict or otherwise undesirable? In this talk, I’ll discuss how symmetry-preserving neural networks can learn symmetry-breaking information in order to fit a dataset that may have missing information unbeknownst to the researcher. Furthermore, due to the mathematical guarantees of equivariant neural networks, these learned parameters are guaranteed to be minimally symmetry breaking. I’ll describe network architectures that can learn these symmetry-breaking parameters in two distinct settings: 1) symmetry-breaking parameters are learned for an entire dataset to capture global asymmetries in the data and 2) symmetry-breaking parameters are predicted in an equivariant manner for individual examples. Finally, I’ll demonstrate these networks on several prototypical examples and apply them to predicting structural distortions of crystalline materials.
Видео Tess Smidt - Symmetry’s made to be broken: Learning how to break symmetry with symmetry-preservin... канала Physics Informed Machine Learning
Показать
Комментарии отсутствуют
Информация о видео
3 февраля 2023 г. 4:04:23
01:04:15
Другие видео канала
![Benjamin Peherstorfer - Physics-based machine learning for quickly simulating transport-dominated...](https://i.ytimg.com/vi/9Cxgu9TuY2U/default.jpg)
![Aditya Nair/Floris van Breugel - Leveraging Machine Learning for interaction-based models and contro](https://i.ytimg.com/vi/TFGMAtlqaJQ/default.jpg)
![Melanie Weber - Exploiting Geometric Structure in Machine Learning and Optimization](https://i.ytimg.com/vi/EOdnedBuJN0/default.jpg)
![Na Li - Scalable and Safe Learning-based Decision-Making in Dynamical Systems](https://i.ytimg.com/vi/5pHjryDp6lw/default.jpg)
![Alex Gorodetsky - Scalable Learning of Dynamical Systems](https://i.ytimg.com/vi/nrd4lw2qYMk/default.jpg)
![Rose Yu - Incorporating Symmetry for Learning Spatiotemporal Dynamics](https://i.ytimg.com/vi/h7TfFssBFEs/default.jpg)
![Hod Lipson - The AI Scientist Automating Discovery, From Cognitive Robotics to Computational Biology](https://i.ytimg.com/vi/0mKguCNHlEo/default.jpg)
![Steve Brunton - Discovering interpretable and generalizable dynamical systems from data](https://i.ytimg.com/vi/UeyRETIwcYw/default.jpg)
![Jane Bae - Wall-models of turbulent flows via scientific multi-agent reinforcement learning](https://i.ytimg.com/vi/dDS1TUx70J0/default.jpg)
![Kathleen Champion - Data-driven discovery of coordinates and governing equations](https://i.ytimg.com/vi/Z-l7G8zq8I0/default.jpg)
![Kevin Carlberg - Breaking Komolgorov-Width Barriers using Deep Learning](https://i.ytimg.com/vi/kxEiJLNqW_4/default.jpg)
![David Bortz - The Surprising Robustness and Computational Efficiency of Weak Form System Identif...](https://i.ytimg.com/vi/LLDFkOoxDBo/default.jpg)
![John Wright - Deep Networks and the Multiple Manifold Problem](https://i.ytimg.com/vi/xCVV4eDDysY/default.jpg)
![Dominique Zosso - Graph-Based Geometric Data Analysis](https://i.ytimg.com/vi/kjZG2ryLmkU/default.jpg)
![Laure Zanna - Data-driven turbulence closures for ocean and climate models: advances and challenges](https://i.ytimg.com/vi/J48yprEYNHk/default.jpg)
![Benjamin Peherstorfer - Data generation for learning reduced models with operator inference](https://i.ytimg.com/vi/odZmGIJYZ14/default.jpg)
![Andrea Bertozzi - Total Variation Minimization on Graphs for Semisupervised and Unsupervised ML](https://i.ytimg.com/vi/EOlG5SGqfL4/default.jpg)
![Andrew Stuart - Supervised Learning For Operators](https://i.ytimg.com/vi/tASot9j7-Cc/default.jpg)
![Krithika Manohar - Optimal Sensors for Empowering AI](https://i.ytimg.com/vi/pDxdjhRU7ak/default.jpg)
![Jörn Dunkel - Symmetry-informed model inference for active matter](https://i.ytimg.com/vi/A53PciowtBE/default.jpg)
![Aleksandr Aravkin - Algorithms for Nonsmooth, Nonconvex Problems in Data-Driven Discovery](https://i.ytimg.com/vi/UnIcBCPO1zg/default.jpg)