[2508.02812] Evaluating and Learning Robust Bandit Policies Under Uncertain Causal Mechanisms
About this article
Abstract page for arXiv paper 2508.02812: Evaluating and Learning Robust Bandit Policies Under Uncertain Causal Mechanisms
Computer Science > Machine Learning arXiv:2508.02812 (cs) [Submitted on 4 Aug 2025 (v1), last revised 3 Apr 2026 (this version, v2)] Title:Evaluating and Learning Robust Bandit Policies Under Uncertain Causal Mechanisms Authors:Katherine Avery, Chinmay Pendse, David Jensen View a PDF of the paper titled Evaluating and Learning Robust Bandit Policies Under Uncertain Causal Mechanisms, by Katherine Avery and 2 other authors View PDF Abstract:Causal graphical models can encode large amounts structural knowledge, both from the background knowledge of domain experts and the structural knowledge discovered from randomized experiments or observational data. However, though we may know the general structure of causal relationships, we often do not know the exact causal mechanisms. In this work, we propose a causal multi-armed bandit evaluation and learning algorithm that can reason effectively despite uncertainty over conditional probability distributions. Further, we show how conditional independence testing can be used to choose variables for modeling. We find that the structural equation model (SEM) approach gives more accurate evaluations compared to traditional approaches, particularly as the range of possible causal mechanisms grows. Further, the SEM approach learns low-variance policies, and it learns an optimal policy, assuming the model is sufficiently well-specified. Traditional approaches can converge to local extrema or fail to converge at all. Comments: Subjects: Mach...