[2604.03962] Predict, Don't React: Value-Based Safety Forecasting for LLM Streaming

[2604.03962] Predict, Don't React: Value-Based Safety Forecasting for LLM Streaming

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2604.03962: Predict, Don't React: Value-Based Safety Forecasting for LLM Streaming

Computer Science > Computation and Language arXiv:2604.03962 (cs) [Submitted on 5 Apr 2026] Title:Predict, Don't React: Value-Based Safety Forecasting for LLM Streaming Authors:Pride Kavumba, Koki Wataoka, Huy H. Nguyen, Jiaxuan Li, Masaya Ohagi View a PDF of the paper titled Predict, Don't React: Value-Based Safety Forecasting for LLM Streaming, by Pride Kavumba and 4 other authors View PDF HTML (experimental) Abstract:In many practical LLM deployments, a single guardrail is used for both prompt and response moderation. Prompt moderation operates on fully observed text, whereas streaming response moderation requires safety decisions to be made over partial generations. Existing text-based streaming guardrails commonly frame this output-side problem as boundary detection, training models to identify the earliest prefix at which a response has already become unsafe. In this work, we introduce StreamGuard, a unified model-agnostic streaming guardrail that instead formulates moderation as a forecasting problem: given a partial prefix, the model predicts the expected harmfulness of likely future continuations. We supervise this prediction using Monte Carlo rollouts, which enables early intervention without requiring exact token-level boundary annotations. Across standard safety benchmarks, StreamGuard performs strongly both for input moderation and for streaming output moderation. At the 8B scale, StreamGuard improves aggregated input-moderation F1 from 86.7 to 88.2 and aggreg...

Originally published on April 07, 2026. Curated by AI News.

Related Articles

[2603.07475] A Comparative analysis of Layer-wise Representational Capacity in AR and Diffusion LLMs
Llms

[2603.07475] A Comparative analysis of Layer-wise Representational Capacity in AR and Diffusion LLMs

Abstract page for arXiv paper 2603.07475: A Comparative analysis of Layer-wise Representational Capacity in AR and Diffusion LLMs

arXiv - Machine Learning · 3 min ·
[2601.22925] BEAR: Towards Beam-Search-Aware Optimization for Recommendation with Large Language Models
Llms

[2601.22925] BEAR: Towards Beam-Search-Aware Optimization for Recommendation with Large Language Models

Abstract page for arXiv paper 2601.22925: BEAR: Towards Beam-Search-Aware Optimization for Recommendation with Large Language Models

arXiv - Machine Learning · 4 min ·
[2512.10551] LLM-Auction: Generative Auction towards LLM-Native Advertising
Llms

[2512.10551] LLM-Auction: Generative Auction towards LLM-Native Advertising

Abstract page for arXiv paper 2512.10551: LLM-Auction: Generative Auction towards LLM-Native Advertising

arXiv - Machine Learning · 3 min ·
[2511.17411] SPEAR-1: Scaling Beyond Robot Demonstrations via 3D Understanding
Llms

[2511.17411] SPEAR-1: Scaling Beyond Robot Demonstrations via 3D Understanding

Abstract page for arXiv paper 2511.17411: SPEAR-1: Scaling Beyond Robot Demonstrations via 3D Understanding

arXiv - Machine Learning · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime