[2602.16953] LLM4Cov: Execution-Aware Agentic Learning for High-coverage Testbench Generation

[2602.16953] LLM4Cov: Execution-Aware Agentic Learning for High-coverage Testbench Generation

arXiv - Machine Learning 3 min read Article

Summary

The paper presents LLM4Cov, an offline learning framework for high-coverage testbench generation, addressing challenges in hardware verification with execution-aware learning techniques.

Why It Matters

This research is significant as it proposes a novel approach to improve the efficiency of hardware verification processes, which are critical in ensuring the reliability of complex systems. By leveraging execution-aware learning, it aims to reduce the costs and time associated with traditional reinforcement learning methods.

Key Takeaways

  • LLM4Cov utilizes offline agent-learning to enhance hardware verification.
  • The framework introduces execution-validated data curation and policy-aware data synthesis.
  • A compact model achieves a 69.2% coverage pass rate, outperforming larger models.
  • The study addresses the limitations of online reinforcement learning in high-coverage scenarios.
  • A reality-aligned benchmark is developed to assess the proposed framework's effectiveness.

Computer Science > Artificial Intelligence arXiv:2602.16953 (cs) [Submitted on 18 Feb 2026] Title:LLM4Cov: Execution-Aware Agentic Learning for High-coverage Testbench Generation Authors:Hejia Zhang, Zhongming Yu, Chia-Tung Ho, Haoxing Ren, Brucek Khailany, Jishen Zhao View a PDF of the paper titled LLM4Cov: Execution-Aware Agentic Learning for High-coverage Testbench Generation, by Hejia Zhang and 5 other authors View PDF HTML (experimental) Abstract:Execution-aware LLM agents offer a promising paradigm for learning from tool feedback, but such feedback is often expensive and slow to obtain, making online reinforcement learning (RL) impractical. High-coverage hardware verification exemplifies this challenge due to its reliance on industrial simulators and non-differentiable execution signals. We propose LLM4Cov, an offline agent-learning framework that models verification as memoryless state transitions guided by deterministic evaluators. Building on this formulation, we introduce execution-validated data curation, policy-aware agentic data synthesis, and worst-state-prioritized sampling to enable scalable learning under execution constraints. We further curate a reality-aligned benchmark adapted from an existing verification suite through a revised evaluation protocol. Using the proposed pipeline, a compact 4B-parameter model achieves 69.2% coverage pass rate under agentic evaluation, outperforming its teacher by 5.3% and demonstrating competitive performance against mod...

Related Articles

Llms

Nvidia goes all-in on AI agents while Anthropic pulls the plug

TLDR: Nvidia is partnering with 17 major companies to build a platform specifically for enterprise AI agents, basically trying to become ...

Reddit - Artificial Intelligence · 1 min ·
Anthropic says Claude Code subscribers will need to pay extra for OpenClaw usage | TechCrunch
Llms

Anthropic says Claude Code subscribers will need to pay extra for OpenClaw usage | TechCrunch

It’s about to become more expensive for Claude Code subscribers to use Anthropic’s coding assistant with OpenClaw and other third-party t...

TechCrunch - AI · 4 min ·
Llms

I am seeing Claude everywhere

Every single Instagram reel or TikTok I scroll i see people mentioning Claude and glazing it like it’s some kind of master tool that’s be...

Reddit - Artificial Intelligence · 1 min ·
Llms

Claude Opus 4.6 API at 40% below Anthropic pricing – try free before you pay anything

Hey everyone I've set up a self-hosted API gateway using [New-API](QuantumNous/new-ap) to manage and distribute Claude Opus 4.6 access ac...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime