[2604.27914] Geometry-Calibrated Conformal Abstention for Language Models
About this article
Abstract page for arXiv paper 2604.27914: Geometry-Calibrated Conformal Abstention for Language Models
Computer Science > Computation and Language arXiv:2604.27914 (cs) [Submitted on 30 Apr 2026] Title:Geometry-Calibrated Conformal Abstention for Language Models Authors:Rui Xu, Yi Chen, Sihong Xie, Hui Xiong View a PDF of the paper titled Geometry-Calibrated Conformal Abstention for Language Models, by Rui Xu and 3 other authors View PDF HTML (experimental) Abstract:When language models lack relevant knowledge for a given query, they frequently generate plausible responses that can be hallucinations, rather than admitting being agnostic about the answer. Retraining models to reward admitting ignorance can lead to overly conservative behaviors and poor generalization due to scarce evaluation benchmarks. We propose a post hoc framework, Conformal Abstention (CA), adapted from conformal prediction (CP) to determine whether to abstain from answering a query. CA provides finite-sample guarantees on both the probability of participation (i.e., not abstaining) and the probability that the generated response is correct. Importantly, the abstention decision relies on prediction confidence rather than the non-conformity scores used in CP, which are intractable for open-ended generation. To better align prediction confidence with the model's ignorance, we introduce a calibration strategy using representation geometry within the model to measure knowledge involvement in shaping the response. Experiments demonstrate that we improve selective answering significantly with 75 percent condi...