Knowledge Distillation - Keras Code Examples
This Keras Code Examples show you how to implement Knowledge Distillation! Knowledge Distillation has lead to new advances in compression, training state of the art models, and stabilizing Transformers for Computer Vision. All you need to do to build on this is swap out the Teacher and Student architectures. I think the example of how to overwrite keras.Model and integrate two loss functions controlled with an alpha hyperparameter weighting is very useful as well.
Content Links
Knowledge Distillation (Keras Code Examples): https://keras.io/examples/vision/knowledge_distillation/
DistilBERT: https://arxiv.org/pdf/1910.01108.pdf
Self-Training with Noisy Student: https://arxiv.org/pdf/1911.04252.pdf
Data-efficient Image Transformers: https://ai.facebook.com/blog/data-efficient-image-transformers-a-promising-new-technique-for-image-classification/
KL Divergence: https://en.wikipedia.org/wiki/Kullback%E2%80%93Leibler_divergence
0:00 Beginning
0:44 Motivation, Success Stories
2:47 Custom keras.Model
11:18 Teacher and Student models
12:17 Data Loading, Train the Teacher
14:05 Distill Teacher to Student
Видео Knowledge Distillation - Keras Code Examples канала Henry AI Labs
Content Links
Knowledge Distillation (Keras Code Examples): https://keras.io/examples/vision/knowledge_distillation/
DistilBERT: https://arxiv.org/pdf/1910.01108.pdf
Self-Training with Noisy Student: https://arxiv.org/pdf/1911.04252.pdf
Data-efficient Image Transformers: https://ai.facebook.com/blog/data-efficient-image-transformers-a-promising-new-technique-for-image-classification/
KL Divergence: https://en.wikipedia.org/wiki/Kullback%E2%80%93Leibler_divergence
0:00 Beginning
0:44 Motivation, Success Stories
2:47 Custom keras.Model
11:18 Teacher and Student models
12:17 Data Loading, Train the Teacher
14:05 Distill Teacher to Student
Видео Knowledge Distillation - Keras Code Examples канала Henry AI Labs
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
![Knowledge Distillation](https://i.ytimg.com/vi/b3zf-JylUus/default.jpg)
![](https://i.ytimg.com/vi/hwnXbnm9Ucw/default.jpg)
![Convolutional Autoencoder for Image Denoising - Keras Code Examples](https://i.ytimg.com/vi/Qz09h3lLSEk/default.jpg)
![Next-frame prediction with Conv-LSTM - Keras Code Examples](https://i.ytimg.com/vi/P5yv8HDFc_M/default.jpg)
![Zettelkasten Note-Taking: What I wish I knew about Smart Notes](https://i.ytimg.com/vi/yqKspwjXu18/default.jpg)
![Knowledge Distillation with TAs](https://i.ytimg.com/vi/ueUAtFLtukM/default.jpg)
![Point Cloud Classification - Keras Code Examples](https://i.ytimg.com/vi/ctdi4Fjp_50/default.jpg)
![Getting Started with Keras](https://i.ytimg.com/vi/J6Ok8p463C4/default.jpg)
![Distilling Neural Networks | Two Minute Papers #218](https://i.ytimg.com/vi/zjaz2mC1KhM/default.jpg)
![Don't Stop Pretraining!](https://i.ytimg.com/vi/zNjiTcF3FZE/default.jpg)
![Training BERT #5 - Training With BertForPretraining](https://i.ytimg.com/vi/IC9FaVPKlYc/default.jpg)
![Convolutional Neural Networks - Deep Learning basics with Python, TensorFlow and Keras p.3](https://i.ytimg.com/vi/WvoLTXIjBYU/default.jpg)
![Knowledge Distillation in Deep Learning - DistilBERT Explained](https://i.ytimg.com/vi/rNOuDKWtrAE/default.jpg)
![Knowledge Distillation: A Good Teacher is Patient and Consistent](https://i.ytimg.com/vi/gZPUGje1PCI/default.jpg)
![Paper towns and why learning is awesome | John Green](https://i.ytimg.com/vi/NgDGlcxYrhQ/default.jpg)
![Knowledge Distillation in Deep Learning - Basics](https://i.ytimg.com/vi/gADXP5daZeM/default.jpg)
![Distilling the Knowledge in a Neural Network](https://i.ytimg.com/vi/k63qGsH1jLo/default.jpg)
![214 - Improving semantic segmentation (U-Net) performance via ensemble of multiple trained networks](https://i.ytimg.com/vi/NFIYdYjJams/default.jpg)
![Self-Training improves Pre-Training for Natural Language Understanding](https://i.ytimg.com/vi/8aV48bwbLTY/default.jpg)
![Knowledge Distillation | Machine Learning](https://i.ytimg.com/vi/BWuAeCBvavw/default.jpg)