[2603.21006] How AI Systems Think About Education: Analyzing Latent Preference Patterns in Large Language Models
About this article
Abstract page for arXiv paper 2603.21006: How AI Systems Think About Education: Analyzing Latent Preference Patterns in Large Language Models
Computer Science > Computers and Society arXiv:2603.21006 (cs) [Submitted on 28 Feb 2026] Title:How AI Systems Think About Education: Analyzing Latent Preference Patterns in Large Language Models Authors:Daniel Autenrieth View a PDF of the paper titled How AI Systems Think About Education: Analyzing Latent Preference Patterns in Large Language Models, by Daniel Autenrieth View PDF HTML (experimental) Abstract:This paper presents the first systematic measurement of educational alignment in Large Language Models. Using a Delphi-validated instrument comprising 48 items across eight educational-theoretical dimensions, the study reveals that GPT-5.1 exhibits highly coherent preference patterns (99.78% transitivity; 92.79% model accuracy) that largely align with humanistic educational principles where expert consensus exists. Crucially, divergences from expert opinion occur precisely in domains of normative disagreement among human experts themselves, particularly emotional dimensions and epistemic normativity. This raises a fundamental question for alignment research: When human values are contested, what should models be aligned to? The findings demonstrate that GPT-5.1 does not remain neutral in contested domains but adopts coherent positions, prioritizing emotional responsiveness and rejecting false balance. The methodology, combining Delphi consensus-building with Structured Preference Elicitation and Thurstonian Utility modeling, provides a replicable framework for domain-...