Week 12 – Practicum: Attention and the Transformer
Course website: http://bit.ly/pDL-home
Playlist: http://bit.ly/pDL-YouTube
Speaker: Alfredo Canziani
Week 12: http://bit.ly/pDL-en-12
0:00:00 – Week 12 – Practicum
PRACTICUM: http://bit.ly/pDL-en-12-3
We introduce attention, focusing on self-attention and its hidden layer representations of the inputs. Then, we introduce the key-value store paradigm and discuss how to represent queries, keys, and values as rotations of an input. Finally, we use attention to interpret the transformer architecture, taking a forward pass through a basic transformer, and comparing the encoder-decoder paradigm to sequential architectures.
0:01:09 – Attention
0:17:36 – Key-value store
0:35:14 – Transformer and PyTorch implementation
0:54:00 – Q&A
Видео Week 12 – Practicum: Attention and the Transformer канала Alfredo Canziani
Playlist: http://bit.ly/pDL-YouTube
Speaker: Alfredo Canziani
Week 12: http://bit.ly/pDL-en-12
0:00:00 – Week 12 – Practicum
PRACTICUM: http://bit.ly/pDL-en-12-3
We introduce attention, focusing on self-attention and its hidden layer representations of the inputs. Then, we introduce the key-value store paradigm and discuss how to represent queries, keys, and values as rotations of an input. Finally, we use attention to interpret the transformer architecture, taking a forward pass through a basic transformer, and comparing the encoder-decoder paradigm to sequential architectures.
0:01:09 – Attention
0:17:36 – Key-value store
0:35:14 – Transformer and PyTorch implementation
0:54:00 – Q&A
Видео Week 12 – Practicum: Attention and the Transformer канала Alfredo Canziani
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
![10 – Self / cross, hard / soft attention and the Transformer](https://i.ytimg.com/vi/fEVyfT-gLqQ/default.jpg)
![An image is worth 16x16 words: ViT | Is this the extinction of CNNs? Long live the Transformer?](https://i.ytimg.com/vi/DVoHvmww2lQ/default.jpg)
![](https://i.ytimg.com/vi/ULRaPiMM47Y/default.jpg)
![NLP for Developers: Transformers | Rasa](https://i.ytimg.com/vi/KN3ZL65Dze0/default.jpg)
![Transformer Networks - Attention is all you need easy introduction!](https://i.ytimg.com/vi/HcYKTsq4v0w/default.jpg)
![Transformer Neural Networks - EXPLAINED! (Attention is all you need)](https://i.ytimg.com/vi/TQQlZhbC5ps/default.jpg)
![Code Review: Transformer - Attention Is All You Need | AISC](https://i.ytimg.com/vi/KMY2Knr4iAs/default.jpg)
![07 – Unsupervised learning: autoencoding the targets](https://i.ytimg.com/vi/IuXsG3sN3zY/default.jpg)
![Pytorch for Beginners #24 | Transformer Model: Self Attention - Simplest Explanation](https://i.ytimg.com/vi/QUk6jvB9RGk/default.jpg)
![Open3D: A Modern Open-Source Library for 3D Data Processing](https://i.ytimg.com/vi/Rsh4poEpahI/default.jpg)
![Rasa Algorithm Whiteboard - Transformers & Attention 3: Multi Head Attention](https://i.ytimg.com/vi/23XUv0T9L5c/default.jpg)
![Gradient descent, how neural networks learn | Chapter 2, Deep learning](https://i.ytimg.com/vi/IHZwWFHWa-w/default.jpg)
![Week 9 – Practicum: (Energy-based) Generative adversarial networks](https://i.ytimg.com/vi/xYc11zyZ26M/default.jpg)
![Week 14 – Lecture: Structured prediction with energy based models](https://i.ytimg.com/vi/gYayCG6YyO8/default.jpg)
![Week 10 – Practicum: The Truck Backer-Upper](https://i.ytimg.com/vi/A3klBqEWR-I/default.jpg)
![Developing a deep learning project (case study on transformer)](https://i.ytimg.com/vi/px4rtkWHFvM/default.jpg)
![CS480/680 Lecture 19: Attention and Transformer Networks](https://i.ytimg.com/vi/OyFJWRnt_AY/default.jpg)
![[Transformer] Attention Is All You Need | AISC Foundational](https://i.ytimg.com/vi/S0KakHcj_rs/default.jpg)
![NLP with Neural Networks & Transformers](https://i.ytimg.com/vi/BGKumht1qLA/default.jpg)
![Vision Transformer (ViT) - An image is worth 16x16 words | Paper Explained](https://i.ytimg.com/vi/j6kuz_NqkG0/default.jpg)