[2603.21607] INTRYGUE: Induction-Aware Entropy Gating for Reliable RAG Uncertainty Estimation
About this article
Abstract page for arXiv paper 2603.21607: INTRYGUE: Induction-Aware Entropy Gating for Reliable RAG Uncertainty Estimation
Computer Science > Artificial Intelligence arXiv:2603.21607 (cs) [Submitted on 23 Mar 2026] Title:INTRYGUE: Induction-Aware Entropy Gating for Reliable RAG Uncertainty Estimation Authors:Alexandra Bazarova, Andrei Volodichev, Daria Kotova, Alexey Zaytsev View a PDF of the paper titled INTRYGUE: Induction-Aware Entropy Gating for Reliable RAG Uncertainty Estimation, by Alexandra Bazarova and 3 other authors View PDF HTML (experimental) Abstract:While retrieval-augmented generation (RAG) significantly improves the factual reliability of LLMs, it does not eliminate hallucinations, so robust uncertainty quantification (UQ) remains essential. In this paper, we reveal that standard entropy-based UQ methods often fail in RAG settings due to a mechanistic paradox. An internal "tug-of-war" inherent to context utilization appears: while induction heads promote grounded responses by copying the correct answer, they collaterally trigger the previously established "entropy neurons". This interaction inflates predictive entropy, causing the model to signal false uncertainty on accurate outputs. To address this, we propose INTRYGUE (Induction-Aware Entropy Gating for Uncertainty Estimation), a mechanistically grounded method that gates predictive entropy based on the activation patterns of induction heads. Evaluated across four RAG benchmarks and six open-source LLMs (4B to 13B parameters), INTRYGUE consistently matches or outperforms a wide range of UQ baselines. Our findings demonstrat...