[2603.03415] Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs
About this article
Abstract page for arXiv paper 2603.03415: Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs
Computer Science > Computation and Language arXiv:2603.03415 (cs) [Submitted on 3 Mar 2026] Title:Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs Authors:Mingyu Jin, Yutong Yin, Jingcheng Niu, Qingcheng Zeng, Wujiang Xu, Mengnan Du, Wei Cheng, Zhaoran Wang, Tianlong Chen, Dimitris N. Metaxas View a PDF of the paper titled Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs, by Mingyu Jin and 9 other authors View PDF Abstract:In this work, we investigate how Large Language Models (LLMs) adapt their internal representations when encountering inputs of increasing difficulty, quantified as the degree of out-of-distribution (OOD) shift. We reveal a consistent and quantifiable phenomenon: as task difficulty increases, whether through harder reasoning questions, longer contexts, or adding answer choices, the last hidden states of LLMs become substantially sparser. In short, \textbf{\textit{the farther the shift, the sparser the representations}}. This sparsity--difficulty relation is observable across diverse models and domains, suggesting that language models respond to unfamiliar or complex inputs by concentrating computation into specialized subspaces in the last hidden state. Through a series of controlled analyses with a learning dynamic explanation, we demonstrate that this sparsity is not incidental but an adaptive mechanism for stabilizing reasoning under OOD. Leveraging this insight, we design \textit{Sparsity-G...