[2602.17913] From Lossy to Verified: A Provenance-Aware Tiered Memory for Agents

[2602.17913] From Lossy to Verified: A Provenance-Aware Tiered Memory for Agents

arXiv - AI 4 min read Article

Summary

The paper presents TierMem, a novel memory framework for agents that balances the need for accurate evidence with efficiency, reducing latency and token consumption while maintaining high accuracy.

Why It Matters

This research addresses the critical challenge of managing memory in AI agents, particularly in balancing the trade-off between using compressed summaries and retaining detailed logs. By improving evidence retrieval, it enhances the reliability of AI responses, which is vital for applications in various fields, including healthcare and customer service.

Key Takeaways

  • TierMem utilizes a two-tier memory system to optimize evidence retrieval.
  • The framework reduces input tokens by 54.1% and latency by 60.7% while maintaining high accuracy.
  • It addresses the issue of unverifiable omissions in AI responses by linking summaries to raw logs.

Computer Science > Databases arXiv:2602.17913 (cs) [Submitted on 20 Feb 2026] Title:From Lossy to Verified: A Provenance-Aware Tiered Memory for Agents Authors:Qiming Zhu, Shunian Chen, Rui Yu, Zhehao Wu, Benyou Wang View a PDF of the paper titled From Lossy to Verified: A Provenance-Aware Tiered Memory for Agents, by Qiming Zhu and 4 other authors View PDF HTML (experimental) Abstract:Long-horizon agents often compress interaction histories into write-time summaries. This creates a fundamental write-before-query barrier: compression decisions are made before the system knows what a future query will hinge on. As a result, summaries can cause unverifiable omissions -- decisive constraints (e.g., allergies) may be dropped, leaving the agent unable to justify an answer with traceable evidence. Retaining raw logs restores an authoritative source of truth, but grounding on raw logs by default is expensive: many queries are answerable from summaries, yet raw grounding still requires processing far longer contexts, inflating token consumption and latency. We propose TierMem, a provenance-linked framework that casts retrieval as an inference-time evidence allocation problem. TierMem uses a two-tier memory hierarchy to answer with the cheapest sufficient evidence: it queries a fast summary index by default, and a runtime sufficiency router Escalates to an immutable raw-log store only when summary evidence is insufficient. TierMem then writes back verified findings as new summary u...

Related Articles

Ai Agents

[P] Easily provide Wandb logs as context to agents for analysis and planning.

It is frustrating to use the Wandb CLI and MCP tools with my agents. For one, the MCP tool basically floods the context window and freque...

Reddit - Machine Learning · 1 min ·
Deepmind's 'AI Agent Traps' Paper Maps How Hackers Could Weaponize AI Agents Against Users
Ai Agents

Deepmind's 'AI Agent Traps' Paper Maps How Hackers Could Weaponize AI Agents Against Users

AI Tools & Products · 7 min ·
Llms

Claude, OpenClaw and the new reality: AI agents are here — and so is the chaos

AI Tools & Products ·
Agentic AI in Beauty: How ChatGPT Is Reshaping Discovery, Trust, and Conversion
Llms

Agentic AI in Beauty: How ChatGPT Is Reshaping Discovery, Trust, and Conversion

AI Tools & Products · 7 min ·
More in Ai Agents: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime