[2602.10140] Can Large Language Models Implement Agent-Based Models? An ODD-based Replication Study

[2602.10140] Can Large Language Models Implement Agent-Based Models? An ODD-based Replication Study

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2602.10140: Can Large Language Models Implement Agent-Based Models? An ODD-based Replication Study

Computer Science > Software Engineering arXiv:2602.10140 (cs) [Submitted on 8 Feb 2026 (v1), last revised 30 Apr 2026 (this version, v2)] Title:Can Large Language Models Implement Agent-Based Models? An ODD-based Replication Study Authors:Nuno Fachada, Daniel Fernandes, Carlos M. Fernandes, João P. Matos-Carvalho View a PDF of the paper titled Can Large Language Models Implement Agent-Based Models? An ODD-based Replication Study, by Nuno Fachada and 3 other authors View PDF HTML (experimental) Abstract:Large language models (LLMs) can now synthesize non-trivial executable code from textual descriptions, raising an important question: can LLMs reliably implement agent-based models from standardized specifications in a way that supports replication, verification, and validation? We address this question by evaluating 17 contemporary LLMs on a controlled ODD-to-code translation task, using the PPHPC predator-prey model as a fully specified reference. Generated Python implementations are assessed through staged executability checks, model-independent statistical comparison against a validated NetLogo baseline, and quantitative measures of runtime efficiency and maintainability. Results show that behaviorally faithful implementations are achievable but not guaranteed, and that executability alone is insufficient for scientific use. GPT-4.1 consistently produces statistically valid and efficient implementations, with Claude 3.7 Sonnet performing well but less reliably. Overall, ...

Originally published on May 01, 2026. Curated by AI News.

Related Articles

[2604.17460] Agentic Education: Using Claude Code to Teach Claude Code
Llms

[2604.17460] Agentic Education: Using Claude Code to Teach Claude Code

Abstract page for arXiv paper 2604.17460: Agentic Education: Using Claude Code to Teach Claude Code

arXiv - AI · 4 min ·
[2603.09117] Decoupling Reasoning and Confidence: Resurrecting Calibration in Reinforcement Learning from Verifiable Rewards
Llms

[2603.09117] Decoupling Reasoning and Confidence: Resurrecting Calibration in Reinforcement Learning from Verifiable Rewards

Abstract page for arXiv paper 2603.09117: Decoupling Reasoning and Confidence: Resurrecting Calibration in Reinforcement Learning from Ve...

arXiv - AI · 3 min ·
[2601.22228] Lost in Space? Vision-Language Models Struggle with Relative Camera Pose Estimation
Llms

[2601.22228] Lost in Space? Vision-Language Models Struggle with Relative Camera Pose Estimation

Abstract page for arXiv paper 2601.22228: Lost in Space? Vision-Language Models Struggle with Relative Camera Pose Estimation

arXiv - AI · 4 min ·
[2601.14289] RPC-Bench: A Fine-grained Benchmark for Research Paper Comprehension
Llms

[2601.14289] RPC-Bench: A Fine-grained Benchmark for Research Paper Comprehension

Abstract page for arXiv paper 2601.14289: RPC-Bench: A Fine-grained Benchmark for Research Paper Comprehension

arXiv - AI · 3 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime