[2603.28052] Meta-Harness: End-to-End Optimization of Model Harnesses
About this article
Abstract page for arXiv paper 2603.28052: Meta-Harness: End-to-End Optimization of Model Harnesses
Computer Science > Artificial Intelligence arXiv:2603.28052 (cs) [Submitted on 30 Mar 2026] Title:Meta-Harness: End-to-End Optimization of Model Harnesses Authors:Yoonho Lee, Roshen Nair, Qizheng Zhang, Kangwook Lee, Omar Khattab, Chelsea Finn View a PDF of the paper titled Meta-Harness: End-to-End Optimization of Model Harnesses, by Yoonho Lee and 5 other authors View PDF HTML (experimental) Abstract:The performance of large language model (LLM) systems depends not only on model weights, but also on their harness: the code that determines what information to store, retrieve, and present to the model. Yet harnesses are still designed largely by hand, and existing text optimizers are poorly matched to this setting because they compress feedback too aggressively. We introduce Meta-Harness, an outer-loop system that searches over harness code for LLM applications. It uses an agentic proposer that accesses the source code, scores, and execution traces of all prior candidates through a filesystem. On online text classification, Meta-Harness improves over a state-of-the-art context management system by 7.7 points while using 4x fewer context tokens. On retrieval-augmented math reasoning, a single discovered harness improves accuracy on 200 IMO-level problems by 4.7 points on average across five held-out models. On agentic coding, discovered harnesses surpass the best hand-engineered baselines on TerminalBench-2. Together, these results show that richer access to prior experience...