Learned data augmentation in natural language processing - Kyunghyun Cho
Data augmentation has been found as a key aspect in modern machine learning. Especially in domains and problems in which there is no knowledge of important invariances and equivariances, a data augmentation procedure can be designed to encourage a machine learning model to encode those invariances and equivairances. In the case of natural language processing, it is unfortunately difficult to come up with such a data augmentation procedure, as the knowledge of invariances and equivariances is limited. Instead, one needs to rely on a vast amount of unlabelled data to “learn to augment” data. In his presentation, Kyunghyun will talk about two different approaches. In the first approach, a standard masked language model is used to produce a set of samples given a training sequence to augment the data along the data (text) manifold learned by the masked language model. In the second approach, an algorithm is designed that learns to interpolate two text snippets, allowing to use a successful data augmentation method, called mixup, which requires a mechanism to mix in contents from two different examples. At the end, Professor Kyunghyun Cho will talk briefly about how this learned data augmentation can be used to predict generalization as well.
______
Kyunghyun Cho is an associate professor of computer science and data science at New York University and CIFAR Fellow of Learning in Machines & Brains. He is also a senior director of frontier research at the Prescient Design team within Genentech Research & Early Development (gRED). He was a research scientist at Facebook AI Research from June 2017 to May 2020 and a postdoctoral fellow at University of Montreal until summer 2015 under the supervision of Prof. Yoshua Bengio, after receiving PhD and MSc degrees from Aalto University April 2011 and April 2014, respectively, under the supervision of Prof. Juha Karhunen, Dr. Tapani Raiko and Dr. Alexander Ilin. He tries his best to find a balance among machine learning, natural language processing, and life, but almost always fails to do so.
Subscribe to our newsletter and stay in the know:
https://www.iarai.ac.at/event-type/seminars/
___________________________________________________________________
IARAI | Institute of Advanced Research in Artificial Intelligence
www.iarai.ac.at
Видео Learned data augmentation in natural language processing - Kyunghyun Cho канала IARAI Research
______
Kyunghyun Cho is an associate professor of computer science and data science at New York University and CIFAR Fellow of Learning in Machines & Brains. He is also a senior director of frontier research at the Prescient Design team within Genentech Research & Early Development (gRED). He was a research scientist at Facebook AI Research from June 2017 to May 2020 and a postdoctoral fellow at University of Montreal until summer 2015 under the supervision of Prof. Yoshua Bengio, after receiving PhD and MSc degrees from Aalto University April 2011 and April 2014, respectively, under the supervision of Prof. Juha Karhunen, Dr. Tapani Raiko and Dr. Alexander Ilin. He tries his best to find a balance among machine learning, natural language processing, and life, but almost always fails to do so.
Subscribe to our newsletter and stay in the know:
https://www.iarai.ac.at/event-type/seminars/
___________________________________________________________________
IARAI | Institute of Advanced Research in Artificial Intelligence
www.iarai.ac.at
Видео Learned data augmentation in natural language processing - Kyunghyun Cho канала IARAI Research
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
Evaluating Machine (Human) Accuracy and Robustness on ImageNet - Ludwig SchmidtA Number Sense as an Emergent Property of the Manipulating Brain - Pietro PeronaEmbedding and Language Modeling for Effective Text Mining - Jiawei HanModern Hopfield Networks - Dr Sepp HochreiterCDCEO 22: Session I - Invited talk by Vipin KumarCDCEO 22: Session III - Invited talk by Nebojsa JojicTowards General and Robust AI at Scale - Irina RishPerformers & Memory - fireside chat: Sepp Hochreiter, Krzysztof Choromanski & Johannes BrandstetterScience4cast Special Session - Special Prize: Francisco AndradesScience4cast Special Session - 2nd Place: Ngoc TranProtein structure prediction with AlphaFold - Andrew SeniorTraffic4cast Special Session: Part II - NeurIPS 2020Neural diffusion PDEs, differential geometry, and graph neural networks - Michael BronsteinWeather4cast 2021 Special Session - Part 2Hopfield Networks in 2021 - Fireside chat between Sepp Hochreiter and Dmitry Krotov | NeurIPS 2020Machine Learning for Location Based Services - Prof. Dr. Ioannis GiannopoulosCDCEO22: Session II - Invited talk by Nantheera AnantrasirichaiScience4cast Special Session - 3rd Place: Milad AghajohariPatenting AI: Why, What and How? - Dr. Alexander KorenbergThe Importance of Motion Perception in Visual Recognition - Roman Pflugfelder