[2603.26539] How Open Must Language Models be to Enable Reliable Scientific Inference?
About this article
Abstract page for arXiv paper 2603.26539: How Open Must Language Models be to Enable Reliable Scientific Inference?
Computer Science > Computation and Language arXiv:2603.26539 (cs) [Submitted on 27 Mar 2026] Title:How Open Must Language Models be to Enable Reliable Scientific Inference? Authors:James A. Michaelov, Catherine Arnett, Tyler A. Chang, Pamela D. Rivière, Samuel M. Taylor, Cameron R. Jones, Sean Trott, Roger P. Levy, Benjamin K. Bergen, Micah Altman View a PDF of the paper titled How Open Must Language Models be to Enable Reliable Scientific Inference?, by James A. Michaelov and 9 other authors View PDF HTML (experimental) Abstract:How does the extent to which a model is open or closed impact the scientific inferences that can be drawn from research that involves it? In this paper, we analyze how restrictions on information about model construction and deployment threaten reliable inference. We argue that current closed models are generally ill-suited for scientific purposes, with some notable exceptions, and discuss ways in which the issues they present to reliable inference can be resolved or mitigated. We recommend that when models are used in research, potential threats to inference should be systematically identified along with the steps taken to mitigate them, and that specific justifications for model selection should be provided. Subjects: Computation and Language (cs.CL); Artificial Intelligence (cs.AI) Cite as: arXiv:2603.26539 [cs.CL] (or arXiv:2603.26539v1 [cs.CL] for this version) https://doi.org/10.48550/arXiv.2603.26539 Focus to learn more arXiv-issued DOI...