2022 Adversarial Machine Learning Rising Star Award Presentation by Fatemehsadat Mireshghallah
Presentation of 2022 Adversarial Machine Learning Rising Star Award by Fatemehsadat Mireshghallah (https://cseweb.ucsd.edu/~fmireshg/)
More details about the workshop and AdvML Rising Star Award: https://sites.google.com/view/advml
___
Talk Title: How much can we trust large language models?
Abstract: Large language Models (LLMs, e.g., GPT-3, TNLG, T-5) are shown to have a remarkably high performance on standard benchmarks, due to their high parameter count, extremely large training datasets, and significant compute. Although the high parameter count in these models leads to more expressiveness, it can also lead to higher memorization, which, coupled with large unvetted, web-scraped datasets can cause multiple different negative societal and ethical impacts: leakage of private, sensitive information— i.e. LLMs are ‘leaky’, generation of biased text—i.e. LLMs are ‘sneaky, and generation hateful or stereotypical text— i.e. LLMs are ‘creepy’. In this talk, I will go over how the issues mentioned above affect the trustworthiness of LLMs, and zoom in on how we can measure the leakage and memorization of these models. Finally I will discuss what it would actually mean for large LLMs to be privacy preserving, and what are the future research directions on making large models trustworthy.
Видео 2022 Adversarial Machine Learning Rising Star Award Presentation by Fatemehsadat Mireshghallah канала TrustworthyAI
More details about the workshop and AdvML Rising Star Award: https://sites.google.com/view/advml
___
Talk Title: How much can we trust large language models?
Abstract: Large language Models (LLMs, e.g., GPT-3, TNLG, T-5) are shown to have a remarkably high performance on standard benchmarks, due to their high parameter count, extremely large training datasets, and significant compute. Although the high parameter count in these models leads to more expressiveness, it can also lead to higher memorization, which, coupled with large unvetted, web-scraped datasets can cause multiple different negative societal and ethical impacts: leakage of private, sensitive information— i.e. LLMs are ‘leaky’, generation of biased text—i.e. LLMs are ‘sneaky, and generation hateful or stereotypical text— i.e. LLMs are ‘creepy’. In this talk, I will go over how the issues mentioned above affect the trustworthiness of LLMs, and zoom in on how we can measure the leakage and memorization of these models. Finally I will discuss what it would actually mean for large LLMs to be privacy preserving, and what are the future research directions on making large models trustworthy.
Видео 2022 Adversarial Machine Learning Rising Star Award Presentation by Fatemehsadat Mireshghallah канала TrustworthyAI
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
2022 AdvML workshop Best Paper Award Presentation by Lei XuAAAI'22 Tutorial on "Adversarial Machine Learning for Good", presented by Pin-Yu Chen @ IBM ResearchPresentations at 3rd Workshop on Adversarial Learning Methods for Machine Learning and Data Mining2021 Adversarial Machine Learning Rising Star Award Presentation by Huan Zhang at AdvML workshop[AAAI 2024] Model Reprogramming: Resource-efficient Cross-Domain Machine Learning by Pin-Yu ChenCrossFL-2022: Workshop on Cross-Community Federated Learning: Algorithms, Systems and Co-designsAAAI 2023: Holistic Adversarial Robustness of Deep Learning Models by Pin-Yu Chen2022 Adversarial Machine Learning Rising Star Award Presentation by Linyi Li at AdvML workshopCVPR 2021 Tutorial on "Practical Adversarial Robustness in Deep Learning: Problems and Solutions"CVPR 2020 Tutorial on Zeroth Order Optimization: Theory and Applications to Deep LearningECCV 2020 Tutorial on Adversarial Robustness of Deep Learning Models by Pin-Yu Chen (IBM Research)2021 Adversarial Machine Learning Rising Star Award Presentation by Florian Tramer at AdvML workshop