Linear algebra with Transformers – Paper Explained
Why would one build a transformer to solve linear algebra problems when there is numpy.linalg? Check out the video to find out why this is a cool idea and understand how the transformer works that can solve 9 linear algebra problems (e.g. matrix multiplication, inversion).
► SPONSOR: Weights & Biases 👉 https://wandb.me/ai-coffee-break
❓ Quiz Questions: https://www.youtube.com/c/AICoffeeBreak/community
➡️ AI Coffee Break Merch! 🛍️ https://aicoffeebreak.creator-spring.com/
📺 Symbolic Mathematics with transformers: https://youtu.be/l7ofrfmVsd0
📺 Transformer explained: https://youtube.com/playlist?list=PLpZBeKTZRGPNdymdEsSSSod5YQ3Vu0sKY
📺 GPT-3: https://youtu.be/5fqxPOaaqi0
📺 Foundation models: https://youtu.be/4Cxz4rnnZ7Q
📺 Interpolation vs. Generalization in Deep Learning: https://youtu.be/-GH9vW5FWUs
Thanks to our Patrons who support us in Tier 2, 3, 4: 🙏
donor, Dres. Trost GbR, Yannik Schneider
Paper 📜: Charton, François. "Linear algebra with transformers." arXiv preprint arXiv:2112.01898 (2021). https://arxiv.org/abs/2112.01898
🔗 Openreview discussion between author and reviewers: https://openreview.net/forum?id=L2a_bcarHcF
🔗 A cat can be an author, too! https://en.wikipedia.org/wiki/F._D._C._Willard
🔗 Gary Marcus: extrapolationhttps://twitter.com/GaryMarcus/status/1411401507610796032
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔥 Optionally, pay us a coffee to help with our Coffee Bean production! ☕
Patreon: https://www.patreon.com/AICoffeeBreak
Ko-fi: https://ko-fi.com/aicoffeebreak
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
Outline:
00:00 Linear algebra with transformers
00:41 Weights & Biases (Sponsor)
02:21 Why throwing transformers at linear algebra is cool.
08:08 How do transformers solve linear algebra?
09:50 Encoding matrices for transformers
11:28 Training data and results
12:43 Generalization!?
16:05 Few-shot learning!?
17:36 AI Coffee Break Quiz call to action
Music 🎵 : Secret Job – Godmode
Cat meowing sound by ignotus : https://freesound.org/people/ignotus/sounds/26104/
----------------------------------
🔗 Links:
AICoffeeBreakQuiz: https://www.youtube.com/c/AICoffeeBreak/community
Twitter: https://twitter.com/AICoffeeBreak
Reddit: https://www.reddit.com/r/AICoffeeBreak/
YouTube: https://www.youtube.com/AICoffeeBreak
#AICoffeeBreak #MsCoffeeBean #MachineLearning #AI #research
Видео Linear algebra with Transformers – Paper Explained канала AI Coffee Break with Letitia
► SPONSOR: Weights & Biases 👉 https://wandb.me/ai-coffee-break
❓ Quiz Questions: https://www.youtube.com/c/AICoffeeBreak/community
➡️ AI Coffee Break Merch! 🛍️ https://aicoffeebreak.creator-spring.com/
📺 Symbolic Mathematics with transformers: https://youtu.be/l7ofrfmVsd0
📺 Transformer explained: https://youtube.com/playlist?list=PLpZBeKTZRGPNdymdEsSSSod5YQ3Vu0sKY
📺 GPT-3: https://youtu.be/5fqxPOaaqi0
📺 Foundation models: https://youtu.be/4Cxz4rnnZ7Q
📺 Interpolation vs. Generalization in Deep Learning: https://youtu.be/-GH9vW5FWUs
Thanks to our Patrons who support us in Tier 2, 3, 4: 🙏
donor, Dres. Trost GbR, Yannik Schneider
Paper 📜: Charton, François. "Linear algebra with transformers." arXiv preprint arXiv:2112.01898 (2021). https://arxiv.org/abs/2112.01898
🔗 Openreview discussion between author and reviewers: https://openreview.net/forum?id=L2a_bcarHcF
🔗 A cat can be an author, too! https://en.wikipedia.org/wiki/F._D._C._Willard
🔗 Gary Marcus: extrapolationhttps://twitter.com/GaryMarcus/status/1411401507610796032
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔥 Optionally, pay us a coffee to help with our Coffee Bean production! ☕
Patreon: https://www.patreon.com/AICoffeeBreak
Ko-fi: https://ko-fi.com/aicoffeebreak
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
Outline:
00:00 Linear algebra with transformers
00:41 Weights & Biases (Sponsor)
02:21 Why throwing transformers at linear algebra is cool.
08:08 How do transformers solve linear algebra?
09:50 Encoding matrices for transformers
11:28 Training data and results
12:43 Generalization!?
16:05 Few-shot learning!?
17:36 AI Coffee Break Quiz call to action
Music 🎵 : Secret Job – Godmode
Cat meowing sound by ignotus : https://freesound.org/people/ignotus/sounds/26104/
----------------------------------
🔗 Links:
AICoffeeBreakQuiz: https://www.youtube.com/c/AICoffeeBreak/community
Twitter: https://twitter.com/AICoffeeBreak
Reddit: https://www.reddit.com/r/AICoffeeBreak/
YouTube: https://www.youtube.com/AICoffeeBreak
#AICoffeeBreak #MsCoffeeBean #MachineLearning #AI #research
Видео Linear algebra with Transformers – Paper Explained канала AI Coffee Break with Letitia
Показать
Комментарии отсутствуют
Информация о видео
22 декабря 2021 г. 17:30:23
00:18:18
Другие видео канала
![Deep Learning for Symbolic Mathematics!? | Paper EXPLAINED](https://i.ytimg.com/vi/l7ofrfmVsd0/default.jpg)
![[Quiz] Eigenfaces, Domain adaptation, Causality, Manifold Hypothesis, Denoising Autoencoder](https://i.ytimg.com/vi/yPXNQ6Ig7hQ/default.jpg)
![[RANT] Adversarial attack on OpenAI’s CLIP? Are we the fools or the foolers?](https://i.ytimg.com/vi/xqdHfLrevuo/default.jpg)
![Our paper at CVPR 2020 - MUL Workshop and ACL 2020 - ALVR Workshop](https://i.ytimg.com/vi/Ev17hz52FGo/default.jpg)
![Data BAD | What Will it Take to Fix Benchmarking for NLU?](https://i.ytimg.com/vi/W57u1j16iC8/default.jpg)
![Preparing for Virtual Conferences – 7 Tips for recording a good conference talk](https://i.ytimg.com/vi/b6Gad5edd18/default.jpg)
![Can a neural network tell if an image is mirrored? – Visual Chirality](https://i.ytimg.com/vi/rbg1Mdo2LZM/default.jpg)
![AI Coffee Break - Channel Trailer](https://i.ytimg.com/vi/h9xPrgTYP_0/default.jpg)
![[Quiz] Interpretable ML, VQ-VAE w/o Quantization / infinite codebook, Pearson’s, PointClouds](https://i.ytimg.com/vi/KS7UlN9SCg4/default.jpg)
![What is the model identifiability problem? | Explained in 60 seconds! | ❓ #AICoffeeBreakQuiz #Shorts](https://i.ytimg.com/vi/mI4sXRSkzE8/default.jpg)
![Adding vs. concatenating positional embeddings & Learned positional encodings](https://i.ytimg.com/vi/M2ToEXF6Olw/default.jpg)
![Transformer in Transformer: Paper explained and visualized | TNT](https://i.ytimg.com/vi/HWna2c5VXDg/default.jpg)
![GaLore EXPLAINED: Memory-Efficient LLM Training by Gradient Low-Rank Projection](https://i.ytimg.com/vi/VC9NbOir7q0/default.jpg)
![Training learned optimizers: VeLO paper EXPLAINED](https://i.ytimg.com/vi/9a6PQJxzUpM/default.jpg)
![Pre-training of BERT-based Transformer architectures explained – language and vision!](https://i.ytimg.com/vi/dabFOBE4eZI/default.jpg)
![What is tokenization and how does it work? Tokenizers explained.](https://i.ytimg.com/vi/D8j1c4NJRfo/default.jpg)
![[Quiz] Regularization in Deep Learning, Lipschitz continuity, Gradient regularization](https://i.ytimg.com/vi/zAyDhZFup9k/default.jpg)
![Adversarial Machine Learning explained! | With examples.](https://i.ytimg.com/vi/YyTyWGUUhmo/default.jpg)
![Are Pre-trained Convolutions Better than Pre-trained Transformers? – Paper Explained](https://i.ytimg.com/vi/xchDU2VMR4M/default.jpg)
![Do Transformers process sequences of FIXED or of VARIABLE length? | #AICoffeeBreakQuiz](https://i.ytimg.com/vi/Xxts1ithupI/default.jpg)
![FNet: Mixing Tokens with Fourier Transforms – Paper Explained](https://i.ytimg.com/vi/j7pWPdGEfMA/default.jpg)