[2604.04182] Comparative reversal learning reveals rigid adaptation in LLMs under non-stationary uncertainty
About this article
Abstract page for arXiv paper 2604.04182: Comparative reversal learning reveals rigid adaptation in LLMs under non-stationary uncertainty
Computer Science > Artificial Intelligence arXiv:2604.04182 (cs) [Submitted on 5 Apr 2026] Title:Comparative reversal learning reveals rigid adaptation in LLMs under non-stationary uncertainty Authors:Haomiaomiao Wang, Tomás E Ward, Lili Zhang View a PDF of the paper titled Comparative reversal learning reveals rigid adaptation in LLMs under non-stationary uncertainty, by Haomiaomiao Wang and 2 other authors View PDF HTML (experimental) Abstract:Non-stationary environments require agents to revise previously learned action values when contingencies change. We treat large language models (LLMs) as sequential decision policies in a two-option probabilistic reversal-learning task with three latent states and switch events triggered by either a performance criterion or timeout. We compare a deterministic fixed transition cycle to a stochastic random schedule that increases volatility, and evaluate DeepSeek-V3.2, Gemini-3, and GPT-5.2, with human data as a behavioural reference. Across models, win-stay was near ceiling while lose-shift was markedly attenuated, revealing asymmetric use of positive versus negative evidence. DeepSeek-V3.2 showed extreme perseveration after reversals and weak acquisition, whereas Gemini-3 and GPT-5.2 adapted more rapidly but still remained less loss-sensitive than humans. Random transitions amplified reversal-specific persistence across LLMs yet did not uniformly reduce total wins, demonstrating that high aggregate payoff can coexist with rigid ada...