[2603.04582] Self-Attribution Bias: When AI Monitors Go Easy on Themselves
About this article
Abstract page for arXiv paper 2603.04582: Self-Attribution Bias: When AI Monitors Go Easy on Themselves
Computer Science > Artificial Intelligence arXiv:2603.04582 (cs) [Submitted on 4 Mar 2026] Title:Self-Attribution Bias: When AI Monitors Go Easy on Themselves Authors:Dipika Khullar, Jack Hopkins, Rowan Wang, Fabien Roger View a PDF of the paper titled Self-Attribution Bias: When AI Monitors Go Easy on Themselves, by Dipika Khullar and 3 other authors View PDF HTML (experimental) Abstract:Agentic systems increasingly rely on language models to monitor their own behavior. For example, coding agents may self critique generated code for pull request approval or assess the safety of tool-use actions. We show that this design pattern can fail when the action is presented in a previous or in the same assistant turn instead of being presented by the user in a user turn. We define self-attribution bias as the tendency of a model to evaluate an action as more correct or less risky when the action is implicitly framed as its own, compared to when the same action is evaluated under off-policy attribution. Across four coding and tool-use datasets, we find that monitors fail to report high-risk or low-correctness actions more often when evaluation follows a previous assistant turn in which the action was generated, compared to when the same action is evaluated in a new context presented in a user turn. In contrast, explicitly stating that the action comes from the monitor does not by itself induce self-attribution bias. Because monitors are often evaluated on fixed examples rather than...