[2603.28038] Beyond the Answer: Decoding the Behavior of LLMs as Scientific Reasoners
About this article
Abstract page for arXiv paper 2603.28038: Beyond the Answer: Decoding the Behavior of LLMs as Scientific Reasoners
Computer Science > Artificial Intelligence arXiv:2603.28038 (cs) [Submitted on 30 Mar 2026] Title:Beyond the Answer: Decoding the Behavior of LLMs as Scientific Reasoners Authors:Rohan Pandey, Eric Ye, Michael Li View a PDF of the paper titled Beyond the Answer: Decoding the Behavior of LLMs as Scientific Reasoners, by Rohan Pandey and 2 other authors View PDF HTML (experimental) Abstract:As Large Language Models (LLMs) achieve increasingly sophisticated performance on complex reasoning tasks, current architectures serve as critical proxies for the internal heuristics of frontier models. Characterizing emergent reasoning is vital for long-term interpretability and safety. Furthermore, understanding how prompting modulates these processes is essential, as natural language will likely be the primary interface for interacting with AGI systems. In this work, we use a custom variant of Genetic Pareto (GEPA) to systematically optimize prompts for scientific reasoning tasks, and analyze how prompting can affect reasoning behavior. We investigate the structural patterns and logical heuristics inherent in GEPA-optimized prompts, and evaluate their transferability and brittleness. Our findings reveal that gains in scientific reasoning often correspond to model-specific heuristics that fail to generalize across systems, which we call "local" logic. By framing prompt optimization as a tool for model interpretability, we argue that mapping these preferred reasoning structures for LLMs ...