[2603.00991] Tracking Capabilities for Safer Agents

[2603.00991] Tracking Capabilities for Safer Agents

arXiv - AI 3 min read

About this article

Abstract page for arXiv paper 2603.00991: Tracking Capabilities for Safer Agents

Computer Science > Artificial Intelligence arXiv:2603.00991 (cs) [Submitted on 1 Mar 2026] Title:Tracking Capabilities for Safer Agents Authors:Martin Odersky, Yaoyu Zhao, Yichen Xu, Oliver Bračevac, Cao Nguyen Pham View a PDF of the paper titled Tracking Capabilities for Safer Agents, by Martin Odersky and 4 other authors View PDF Abstract:AI agents that interact with the real world through tool calls pose fundamental safety challenges: agents might leak private information, cause unintended side effects, or be manipulated through prompt injection. To address these challenges, we propose to put the agent in a programming-language-based "safety harness": instead of calling tools directly, agents express their intentions as code in a capability-safe language: Scala 3 with capture checking. Capabilities are program variables that regulate access to effects and resources of interest. Scala's type system tracks capabilities statically, providing fine-grained control over what an agent can do. In particular, it enables local purity, the ability to enforce that sub-computations are side-effect-free, preventing information leakage when agents process classified data. We demonstrate that extensible agent safety harnesses can be built by leveraging a strong type system with tracked capabilities. Our experiments show that agents can generate capability-safe code with no significant loss in task performance, while the type system reliably prevents unsafe behaviors such as information...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

Okta CEO: The next frontier of security is AI agent identity | The Verge
Ai Agents

Okta CEO: The next frontier of security is AI agent identity | The Verge

Todd McKinnon on why AI agents need an identity, security in an OpenClaw era, and being “paranoid” in preparing for the SaaSpocalypse.

The Verge - AI · 61 min ·
[2506.20964] Evidence-based diagnostic reasoning with multi-agent copilot for human pathology
Llms

[2506.20964] Evidence-based diagnostic reasoning with multi-agent copilot for human pathology

Abstract page for arXiv paper 2506.20964: Evidence-based diagnostic reasoning with multi-agent copilot for human pathology

arXiv - AI · 4 min ·
[2601.08323] AtomMem : Learnable Dynamic Agentic Memory with Atomic Memory Operation
Ai Agents

[2601.08323] AtomMem : Learnable Dynamic Agentic Memory with Atomic Memory Operation

Abstract page for arXiv paper 2601.08323: AtomMem : Learnable Dynamic Agentic Memory with Atomic Memory Operation

arXiv - AI · 3 min ·
[2603.18349] Large-Scale Analysis of Persuasive Content on Moltbook
Llms

[2603.18349] Large-Scale Analysis of Persuasive Content on Moltbook

Abstract page for arXiv paper 2603.18349: Large-Scale Analysis of Persuasive Content on Moltbook

arXiv - AI · 3 min ·
More in Ai Agents: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime