[2603.27905] ATLAS-RTC: Closing the Loop on LLM Agent Output with Token-Level Runtime Control

[2603.27905] ATLAS-RTC: Closing the Loop on LLM Agent Output with Token-Level Runtime Control

arXiv - Machine Learning 3 min read

About this article

Abstract page for arXiv paper 2603.27905: ATLAS-RTC: Closing the Loop on LLM Agent Output with Token-Level Runtime Control

Computer Science > Machine Learning arXiv:2603.27905 (cs) [Submitted on 29 Mar 2026] Title:ATLAS-RTC: Closing the Loop on LLM Agent Output with Token-Level Runtime Control Authors:Christopher Cruz View a PDF of the paper titled ATLAS-RTC: Closing the Loop on LLM Agent Output with Token-Level Runtime Control, by Christopher Cruz View PDF HTML (experimental) Abstract:We present ATLAS-RTC, a runtime control system for autoregressive language models that enforces structured output during decoding. ATLAS-RTC monitors generation at each step, detects drift from output contracts using lightweight signals, and applies targeted interventions such as biasing, masking, and rollback. Unlike post-hoc validation or static constrained decoding, it operates in a closed loop, enabling correction before errors materialize. Across structured generation and tool-calling tasks, ATLAS-RTC improves first-attempt success rates by 20 to 37.8 percentage points, with up to 88% latency reduction in failure-dominated settings. Results show that many failures arise from decoding artifacts rather than task misunderstanding, motivating runtime control as a distinct layer in LLM systems. Subjects: Machine Learning (cs.LG) ACM classes: I.2.8 Cite as: arXiv:2603.27905 [cs.LG]   (or arXiv:2603.27905v1 [cs.LG] for this version)   https://doi.org/10.48550/arXiv.2603.27905 Focus to learn more arXiv-issued DOI via DataCite (pending registration) Submission history From: Christopher Cruz [view email] [v1] Sun, 29...

Originally published on March 31, 2026. Curated by AI News.

Related Articles

Llms

Agents Can Now Propose and Deploy Their Own Code Changes

150 clones yesterday. 43 stars in 3 days. Every agent framework you've used (LangChain, LangGraph, Claude Code) assumes agents are tools ...

Reddit - Artificial Intelligence · 1 min ·
[2603.17839] How do LLMs Compute Verbal Confidence
Llms

[2603.17839] How do LLMs Compute Verbal Confidence

Abstract page for arXiv paper 2603.17839: How do LLMs Compute Verbal Confidence

arXiv - AI · 4 min ·
[2603.15970] 100x Cost & Latency Reduction: Performance Analysis of AI Query Approximation using Lightweight Proxy Models
Llms

[2603.15970] 100x Cost & Latency Reduction: Performance Analysis of AI Query Approximation using Lightweight Proxy Models

Abstract page for arXiv paper 2603.15970: 100x Cost & Latency Reduction: Performance Analysis of AI Query Approximation using Lightweight...

arXiv - AI · 4 min ·
[2603.10062] Multi-Agent Memory from a Computer Architecture Perspective: Visions and Challenges Ahead
Llms

[2603.10062] Multi-Agent Memory from a Computer Architecture Perspective: Visions and Challenges Ahead

Abstract page for arXiv paper 2603.10062: Multi-Agent Memory from a Computer Architecture Perspective: Visions and Challenges Ahead

arXiv - AI · 3 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime