Causal Matrix Completion
Anish Agarwal (UC Berkeley)
https://simons.berkeley.edu/talks/causal-matrix-completion
Algorithmic Aspects of Causal Inference
Matrix completion is the study of recovering an underlying matrix from a sparse subset of noisy observations. Traditionally, it is assumed that the entries of the matrix are “missing completely at random” (MCAR), i.e., each entry is revealed at random, independent of everything else, with uniform probability. This is likely unrealistic due to the presence of “latent confounders”, i.e., unobserved factors that determine both the entries of the underlying matrix and the missingness pattern in the observed matrix. For example, in the context of movie recommender systems—a canonical application for matrix completion—a user who vehemently dislikes horror films is unlikely to ever watch horror films. In general, these confounders yield “missing not at random” (MNAR) data, which can severely impact any inference procedure that does not correct for this bias. We develop a formal causal model for matrix completion through the language of potential outcomes, and provide novel identification arguments for a variety of causal estimands of interest. We design a procedure, which we call “synthetic nearest neighbors” (SNN), to estimate these causal estimands. We prove finite-sample consistency and asymptotic normality of our estimator. Our analysis also leads to new theoretical results for the matrix completion literature. In particular, we establish entry-wise, i.e., max-norm, finite-sample consistency and asymptotic normality results for matrix completion with MNAR data. As a special case, this also provides entry-wise bounds for matrix completion with MCAR data. Across simulated and real data, we demonstrate the efficacy of our proposed estimator
Видео Causal Matrix Completion канала Simons Institute
https://simons.berkeley.edu/talks/causal-matrix-completion
Algorithmic Aspects of Causal Inference
Matrix completion is the study of recovering an underlying matrix from a sparse subset of noisy observations. Traditionally, it is assumed that the entries of the matrix are “missing completely at random” (MCAR), i.e., each entry is revealed at random, independent of everything else, with uniform probability. This is likely unrealistic due to the presence of “latent confounders”, i.e., unobserved factors that determine both the entries of the underlying matrix and the missingness pattern in the observed matrix. For example, in the context of movie recommender systems—a canonical application for matrix completion—a user who vehemently dislikes horror films is unlikely to ever watch horror films. In general, these confounders yield “missing not at random” (MNAR) data, which can severely impact any inference procedure that does not correct for this bias. We develop a formal causal model for matrix completion through the language of potential outcomes, and provide novel identification arguments for a variety of causal estimands of interest. We design a procedure, which we call “synthetic nearest neighbors” (SNN), to estimate these causal estimands. We prove finite-sample consistency and asymptotic normality of our estimator. Our analysis also leads to new theoretical results for the matrix completion literature. In particular, we establish entry-wise, i.e., max-norm, finite-sample consistency and asymptotic normality results for matrix completion with MNAR data. As a special case, this also provides entry-wise bounds for matrix completion with MCAR data. Across simulated and real data, we demonstrate the efficacy of our proposed estimator
Видео Causal Matrix Completion канала Simons Institute
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
Berkeley in the 80s, Episode 5: Richard KarpAdversarial Examples in Deep LearningQuantum Advantage Without StructureRemarks on the Discrete CubeFast Reinforcement Learning With Generalized Policy Updates"The Problem with Qubits"Stability and Learning in Repeated GamesPlanning and Markov Decision Processes Part 1 (reupload)Panel | Quantum ColloquiumTractable Probabilistic CircuitsAlgorithmic Fairness From The Lens Of Causality And Information TheoryProject CETI Next Steps: Industrial-Scale Whale Bioacoustic Data Collection and AnalysisMaximum Satisfiability SolvingLearning and Incentives (Part I)Tutorial: Implicit Bias IUsing Theories of Decision-Making Under Uncertainty to Improve Data VisualizationEquivariant RLBerkeley in the 80s, Episode 1: Shafi GoldwasserThe Role of Conventions in Adaptive Human-AI InteractionAttacking the Off-Policy Problem With Duality