[2604.13068] Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive Language Models

[2604.13068] Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive Language Models

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2604.13068: Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive Language Models

Computer Science > Computation and Language arXiv:2604.13068 (cs) [Submitted on 20 Mar 2026] Title:Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive Language Models Authors:Dip Roy, Rajiv Misra, Sanjay Kumar Singh, Anisha Roy View a PDF of the paper titled Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive Language Models, by Dip Roy and 3 other authors View PDF Abstract:When do large language models decide to hallucinate? Despite serious consequences in healthcare, law, and finance, few formal answers exist. Recent work shows autoregressive models maintain internal representations distinguishing factual from fictional outputs, but when these representations peak as a function of model scale remains poorly understood. We study the temporal dynamics of hallucination-indicative internal representations across 7 autoregressive transformers (117M--7B parameters) using three fact-based datasets (TriviaQA, Simple Facts, Biography; 552 labeled examples). We identify a scale-dependent phase transition: models below 400M parameters show chance-level probe accuracy at every generation position (AUC = 0.48--0.67), indicating no reliable factuality signal. Above $\sim$1B parameters, a qualitatively different regime emerges where peak detectability occurs at position zero -- before any tokens are generated -- then declines during generation. This pre-generation signal is statistically significant in bo...

Originally published on April 16, 2026. Curated by AI News.

Related Articles

Treating enterprise AI as an operating layer | MIT Technology Review
Llms

Treating enterprise AI as an operating layer | MIT Technology Review

There’s a fault line running through enterprise AI, and it’s not the one getting the most attention. The public conversation still tracks...

MIT Technology Review - AI · 7 min ·
Llms

emotion in llms

you know most human emotion is constructed, inferred, there is no root object, you can kind of create the emotion you want? well, i was l...

Reddit - Artificial Intelligence · 1 min ·
Making AI operational in constrained public sector environments | MIT Technology Review
Llms

Making AI operational in constrained public sector environments | MIT Technology Review

The AI boom has hit across industries, and public sector organizations are facing pressure to accelerate adoption. At the same time, gove...

MIT Technology Review - AI · 8 min ·
[2510.19268] Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models
Llms

[2510.19268] Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models

Abstract page for arXiv paper 2510.19268: Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models

arXiv - Machine Learning · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime