Data BAD | What Will it Take to Fix Benchmarking for NLU?
The Coffee Bean explains and comments the sobering take of the paper "What Will it Take to Fix Benchmarking in Natural Language Understanding?"
See more videos from Ms. Coffee Bean about natural language understanding:
📺 The road to NLU: https://youtube.com/playlist?list=PLpZBeKTZRGPMjF-Ob-NYjaTtewbMNXKcU
► Thanks to our Patrons who support us in Tier 2, 3, 4: 🙏
donor, Dres. Trost GbR, Yannik Schneider
➡️ AI Coffee Break Merch! 🛍️ https://aicoffeebreak.creator-spring.com/
Paper:
📜 Bowman, Samuel R., and George E. Dahl. "What Will it Take to Fix Benchmarking in Natural Language Understanding?." arXiv preprint arXiv:2104.02145 (2021). https://arxiv.org/abs/2104.02145
🔗 SuperGLUE: https://super.gluebenchmark.com/tasks
🔗 WiC: The Word-in-Context Dataset (English): https://pilehvar.github.io/wic/
Outline:
00:00 NLU Benchmarking – Motivation
01:04 How to measure NLU advances?
02:31 Why is NLU benchmarking broken?
04:43 What are the fixes?
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔥 Optionally, pay us a coffee to help with our Coffee Bean production! ☕
Patreon: https://www.patreon.com/AICoffeeBreak
Ko-fi: https://ko-fi.com/aicoffeebreak
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔗 Links:
AICoffeeBreakQuiz: https://www.youtube.com/c/AICoffeeBreak/community
Twitter: https://twitter.com/AICoffeeBreak
Reddit: https://www.reddit.com/r/AICoffeeBreak/
YouTube: https://www.youtube.com/AICoffeeBreak
#AICoffeeBreak #MsCoffeeBean #MachineLearning #AI #research
Видео Data BAD | What Will it Take to Fix Benchmarking for NLU? канала AI Coffee Break with Letitia
See more videos from Ms. Coffee Bean about natural language understanding:
📺 The road to NLU: https://youtube.com/playlist?list=PLpZBeKTZRGPMjF-Ob-NYjaTtewbMNXKcU
► Thanks to our Patrons who support us in Tier 2, 3, 4: 🙏
donor, Dres. Trost GbR, Yannik Schneider
➡️ AI Coffee Break Merch! 🛍️ https://aicoffeebreak.creator-spring.com/
Paper:
📜 Bowman, Samuel R., and George E. Dahl. "What Will it Take to Fix Benchmarking in Natural Language Understanding?." arXiv preprint arXiv:2104.02145 (2021). https://arxiv.org/abs/2104.02145
🔗 SuperGLUE: https://super.gluebenchmark.com/tasks
🔗 WiC: The Word-in-Context Dataset (English): https://pilehvar.github.io/wic/
Outline:
00:00 NLU Benchmarking – Motivation
01:04 How to measure NLU advances?
02:31 Why is NLU benchmarking broken?
04:43 What are the fixes?
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔥 Optionally, pay us a coffee to help with our Coffee Bean production! ☕
Patreon: https://www.patreon.com/AICoffeeBreak
Ko-fi: https://ko-fi.com/aicoffeebreak
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔗 Links:
AICoffeeBreakQuiz: https://www.youtube.com/c/AICoffeeBreak/community
Twitter: https://twitter.com/AICoffeeBreak
Reddit: https://www.reddit.com/r/AICoffeeBreak/
YouTube: https://www.youtube.com/AICoffeeBreak
#AICoffeeBreak #MsCoffeeBean #MachineLearning #AI #research
Видео Data BAD | What Will it Take to Fix Benchmarking for NLU? канала AI Coffee Break with Letitia
Показать
Комментарии отсутствуют
Информация о видео
10 октября 2021 г. 17:00:12
00:12:56
Другие видео канала
![Deep Learning for Symbolic Mathematics!? | Paper EXPLAINED](https://i.ytimg.com/vi/l7ofrfmVsd0/default.jpg)
![[Quiz] Eigenfaces, Domain adaptation, Causality, Manifold Hypothesis, Denoising Autoencoder](https://i.ytimg.com/vi/yPXNQ6Ig7hQ/default.jpg)
![[RANT] Adversarial attack on OpenAI’s CLIP? Are we the fools or the foolers?](https://i.ytimg.com/vi/xqdHfLrevuo/default.jpg)
![Our paper at CVPR 2020 - MUL Workshop and ACL 2020 - ALVR Workshop](https://i.ytimg.com/vi/Ev17hz52FGo/default.jpg)
![Preparing for Virtual Conferences – 7 Tips for recording a good conference talk](https://i.ytimg.com/vi/b6Gad5edd18/default.jpg)
![Can a neural network tell if an image is mirrored? – Visual Chirality](https://i.ytimg.com/vi/rbg1Mdo2LZM/default.jpg)
![AI Coffee Break - Channel Trailer](https://i.ytimg.com/vi/h9xPrgTYP_0/default.jpg)
![[Quiz] Interpretable ML, VQ-VAE w/o Quantization / infinite codebook, Pearson’s, PointClouds](https://i.ytimg.com/vi/KS7UlN9SCg4/default.jpg)
![What is the model identifiability problem? | Explained in 60 seconds! | ❓ #AICoffeeBreakQuiz #Shorts](https://i.ytimg.com/vi/mI4sXRSkzE8/default.jpg)
![Adding vs. concatenating positional embeddings & Learned positional encodings](https://i.ytimg.com/vi/M2ToEXF6Olw/default.jpg)
![GaLore EXPLAINED: Memory-Efficient LLM Training by Gradient Low-Rank Projection](https://i.ytimg.com/vi/VC9NbOir7q0/default.jpg)
![Transformer in Transformer: Paper explained and visualized | TNT](https://i.ytimg.com/vi/HWna2c5VXDg/default.jpg)
![Training learned optimizers: VeLO paper EXPLAINED](https://i.ytimg.com/vi/9a6PQJxzUpM/default.jpg)
![Pre-training of BERT-based Transformer architectures explained – language and vision!](https://i.ytimg.com/vi/dabFOBE4eZI/default.jpg)
![What is tokenization and how does it work? Tokenizers explained.](https://i.ytimg.com/vi/D8j1c4NJRfo/default.jpg)
![[Quiz] Regularization in Deep Learning, Lipschitz continuity, Gradient regularization](https://i.ytimg.com/vi/zAyDhZFup9k/default.jpg)
![Adversarial Machine Learning explained! | With examples.](https://i.ytimg.com/vi/YyTyWGUUhmo/default.jpg)
![Are Pre-trained Convolutions Better than Pre-trained Transformers? – Paper Explained](https://i.ytimg.com/vi/xchDU2VMR4M/default.jpg)
![Do Transformers process sequences of FIXED or of VARIABLE length? | #AICoffeeBreakQuiz](https://i.ytimg.com/vi/Xxts1ithupI/default.jpg)
![FNet: Mixing Tokens with Fourier Transforms – Paper Explained](https://i.ytimg.com/vi/j7pWPdGEfMA/default.jpg)