Загрузка страницы

Tin Nguyen: "Sensitivity of MCMC-based analyses to small-data removal"

Talk Title: Sensitivity of MCMC-based analyses to small-data removal
Thesis Committee: Tamara Broderick, Ashia Wilson, and Stefanie Jegelka

Talk Abstract: If the conclusion of a data analysis is sensitive to dropping very few data points, that conclusion might hinge on the particular data at hand rather than representing a more broadly applicable truth. How could we check whether this sensitivity holds? One idea is to consider every small subset of data, drop it from the dataset, and re-run our analysis. But running MCMC to approximate a Bayesian posterior is already very expensive; running multiple times is prohibitive, and the number of re-runs needed here is combinatorially large. Recent work proposes a fast and accurate approximation to find the worst-case dropped data subset, but that work was developed for problems based on estimating equations --- and does not directly handle Bayesian posterior approximations using MCMC. We make two principal contributions in the present work. We adapt the existing data-dropping approximation to estimators computed via MCMC. Observing that Monte Carlo errors induce variability in the approximation, we use a variant of the bootstrap to quantify this uncertainty. We demonstrate how to use our approximation in practice to determine whether there is non-robustness in a problem. Empirically, our method is accurate in simple models, such as linear regression. In models with complicated structure, such as hierarchical models, the performance of our method is mixed.

Видео Tin Nguyen: "Sensitivity of MCMC-based analyses to small-data removal" канала Tamara Broderick
Показать
Комментарии отсутствуют
Введите заголовок:

Введите адрес ссылки:

Введите адрес видео с YouTube:

Зарегистрируйтесь или войдите с
Информация о видео
10 мая 2024 г. 3:10:52
01:01:37
Другие видео канала
Nicholas Bonaker: "Nomon: A Flexible, Bayesian Interface for Motor-Impaired Users"Nicholas Bonaker: "Nomon: A Flexible, Bayesian Interface for Motor-Impaired Users"MIT: Machine Learning 6.036, Lecture 9: State machines and Markov decision processes (Fall 2020)MIT: Machine Learning 6.036, Lecture 9: State machines and Markov decision processes (Fall 2020)MIT: Machine Learning 6.036, Lecture 7: Brief intermission (Fall 2020)MIT: Machine Learning 6.036, Lecture 7: Brief intermission (Fall 2020)MIT: Machine Learning 6.036, Lecture 8: Convolutional neural networks (Fall 2020)MIT: Machine Learning 6.036, Lecture 8: Convolutional neural networks (Fall 2020)Tamara Broderick: "Approximate Cross-Validation for Complex Models"Tamara Broderick: "Approximate Cross-Validation for Complex Models"Tamara Broderick: "Toward a taxonomy of trust for probabilistic data analysis"Tamara Broderick: "Toward a taxonomy of trust for probabilistic data analysis"Lorenzo Masoero: "Predicting and maximizing genomic variety discovery via Bayesian nonparametrics"Lorenzo Masoero: "Predicting and maximizing genomic variety discovery via Bayesian nonparametrics"Lorenzo Masoero: "Bayesian nonparametrics for maximizing power in rare variants association studies"Lorenzo Masoero: "Bayesian nonparametrics for maximizing power in rare variants association studies"Soumya Ghosh: "Approximate Cross-Validation for Structured Models"Soumya Ghosh: "Approximate Cross-Validation for Structured Models"Brian Trippe: "Advances in Bayesian Linear Modeling in High Dimensions"Brian Trippe: "Advances in Bayesian Linear Modeling in High Dimensions"MIT: Machine Learning 6.036, Lecture 11: Recurrent neural networks (Fall 2020)MIT: Machine Learning 6.036, Lecture 11: Recurrent neural networks (Fall 2020)William Stephenson: "Can we globally optimize cross-validation loss?"William Stephenson: "Can we globally optimize cross-validation loss?"MIT: Machine Learning 6.036, Lecture 6: Neural networks (Fall 2020)MIT: Machine Learning 6.036, Lecture 6: Neural networks (Fall 2020)Brian Trippe: "Bayes Estimates for Multiple Related Regressions" (JSM 2020)Brian Trippe: "Bayes Estimates for Multiple Related Regressions" (JSM 2020)Raj Agrawal: "High-Dimensional Variable Selection & Nonlinear Interaction Discovery in Linear Time"Raj Agrawal: "High-Dimensional Variable Selection & Nonlinear Interaction Discovery in Linear Time"Soumya Ghosh: "Are you using test log-likelihood correctly?"Soumya Ghosh: "Are you using test log-likelihood correctly?"MIT: Machine Learning 6.036, Lecture 4: Logistic regression (Fall 2020)MIT: Machine Learning 6.036, Lecture 4: Logistic regression (Fall 2020)MIT: Machine Learning 6.036, Lecture 2: Perceptrons (Fall 2020)MIT: Machine Learning 6.036, Lecture 2: Perceptrons (Fall 2020)MIT: Machine Learning 6.036, Lecture 12: Decision trees and random forests (Fall 2020)MIT: Machine Learning 6.036, Lecture 12: Decision trees and random forests (Fall 2020)MIT: Machine Learning 6.036, Lecture 1: Basics (Fall 2020)MIT: Machine Learning 6.036, Lecture 1: Basics (Fall 2020)
Яндекс.Метрика