[2603.26718] Toward Evaluation Frameworks for Multi-Agent Scientific AI Systems
About this article
Abstract page for arXiv paper 2603.26718: Toward Evaluation Frameworks for Multi-Agent Scientific AI Systems
Computer Science > Computers and Society arXiv:2603.26718 (cs) [Submitted on 18 Mar 2026] Title:Toward Evaluation Frameworks for Multi-Agent Scientific AI Systems Authors:Marcin Abram View a PDF of the paper titled Toward Evaluation Frameworks for Multi-Agent Scientific AI Systems, by Marcin Abram View PDF HTML (experimental) Abstract:We analyze the challenges of benchmarking scientific (multi)-agentic systems, including the difficulty of distinguishing reasoning from retrieval, the risks of data/model contamination, the lack of reliable ground truth for novel research problems, the complications introduced by tool use, and the replication challenges due to the continuously changing/updating knowledge base. We discuss strategies for constructing contamination-resistant problems, generating scalable families of tasks, and the need for evaluating systems through multi-turn interactions that better reflect real scientific practice. As an early feasibility test, we demonstrate how to construct a dataset of novel research ideas to test the out-of-sample performance of our system. We also discuss the results of interviews with several researchers and engineers working in quantum science. Through those interviews, we examine how scientists expect to interact with AI systems and how these expectations should shape evaluation methods. Comments: Subjects: Computers and Society (cs.CY); Artificial Intelligence (cs.AI); Multiagent Systems (cs.MA); Quantum Physics (quant-ph) Cite as: a...