[2506.11128] Theory-Grounded Evaluation of Human-Like Fallacy Patterns in LLM Reasoning
About this article
Abstract page for arXiv paper 2506.11128: Theory-Grounded Evaluation of Human-Like Fallacy Patterns in LLM Reasoning
Computer Science > Computation and Language arXiv:2506.11128 (cs) [Submitted on 10 Jun 2025 (v1), last revised 20 Mar 2026 (this version, v3)] Title:Theory-Grounded Evaluation of Human-Like Fallacy Patterns in LLM Reasoning Authors:Andrew Keenan Richardson, Ryan Othniel Kearns, Sean Moss, Vincent Wang-Mascianica, Philipp Koralus View a PDF of the paper titled Theory-Grounded Evaluation of Human-Like Fallacy Patterns in LLM Reasoning, by Andrew Keenan Richardson and 4 other authors View PDF HTML (experimental) Abstract:We study logical reasoning in language models by asking whether their errors follow established human fallacy patterns. Using the Erotetic Theory of Reasoning (ETR) and its open-source implementation, PyETR, we programmatically generate 383 formally specified reasoning problems and evaluate 38 models. For each response, we judge logical correctness and, when incorrect, whether it matches an ETR-predicted fallacy. Two results stand out: (i) as a capability proxy (Chatbot Arena Elo) increases, a larger share of a model's incorrect answers are ETR-predicted fallacies $(\rho=0.360, p=0.0265)$, while overall correctness on this dataset shows no correlation with capability; (ii) reversing premise order significantly reduces fallacy production for many models, mirroring human order effects. Methodologically, PyETR provides an open-source pipeline for unbounded, synthetic, contamination-resistant reasoning tests linked to a cognitive theory, enabling analyses that fo...