[2509.24156] Reasoning or Retrieval? A Study of Answer Attribution on Large Reasoning Models
About this article
Abstract page for arXiv paper 2509.24156: Reasoning or Retrieval? A Study of Answer Attribution on Large Reasoning Models
Computer Science > Artificial Intelligence arXiv:2509.24156 (cs) [Submitted on 29 Sep 2025 (v1), last revised 2 Mar 2026 (this version, v2)] Title:Reasoning or Retrieval? A Study of Answer Attribution on Large Reasoning Models Authors:Yuhui Wang, Changjiang Li, Guangke Chen, Jiacheng Liang, Ting Wang View a PDF of the paper titled Reasoning or Retrieval? A Study of Answer Attribution on Large Reasoning Models, by Yuhui Wang and 4 other authors View PDF HTML (experimental) Abstract:Large reasoning models (LRMs) exhibit unprecedented capabilities in solving complex problems through Chain-of-Thought (CoT) reasoning. However, recent studies reveal that their final answers often contradict their own reasoning traces. We hypothesize that this inconsistency stems from two competing mechanisms for generating answers: CoT reasoning and memory retrieval. To test this hypothesis, we conduct controlled experiments that challenge LRMs with misleading cues during reasoning and/or corrupted answers during retrieval. Our results across models and datasets confirm that both mechanisms operate simultaneously, with their relative dominance influenced by multiple factors: problem domains, model scales, and fine-tuning approaches (e.g., reinforcement learning vs. distillation). The findings reveal a critical limitation in current reasoning fine-tuning paradigms: models can exploit the retrieval mechanism as a shortcut, effectively "hacking" the reward signal and undermining genuine reasoning d...