[2603.04831] Missingness Bias Calibration in Feature Attribution Explanations
About this article
Abstract page for arXiv paper 2603.04831: Missingness Bias Calibration in Feature Attribution Explanations
Computer Science > Machine Learning arXiv:2603.04831 (cs) [Submitted on 5 Mar 2026] Title:Missingness Bias Calibration in Feature Attribution Explanations Authors:Shailesh Sridhar, Anton Xue, Eric Wong View a PDF of the paper titled Missingness Bias Calibration in Feature Attribution Explanations, by Shailesh Sridhar and 2 other authors View PDF HTML (experimental) Abstract:Popular explanation methods often produce unreliable feature importance scores due to missingness bias, a systematic distortion that arises when models are probed with ablated, out-of-distribution inputs. Existing solutions treat this as a deep representational flaw that requires expensive retraining or architectural modifications. In this work, we challenge this assumption and show that missingness bias can be effectively treated as a superficial artifact of the model's output space. We introduce MCal, a lightweight post-hoc method that corrects this bias by fine-tuning a simple linear head on the outputs of a frozen base model. Surprisingly, we find this simple correction consistently reduces missingness bias and is competitive with, or even outperforms, prior heavyweight approaches across diverse medical benchmarks spanning vision, language, and tabular domains. Subjects: Machine Learning (cs.LG) Cite as: arXiv:2603.04831 [cs.LG] (or arXiv:2603.04831v1 [cs.LG] for this version) https://doi.org/10.48550/arXiv.2603.04831 Focus to learn more arXiv-issued DOI via DataCite (pending registration) Submi...