[2604.04855] The Role of Generator Access in Autoregressive Post-Training

[2604.04855] The Role of Generator Access in Autoregressive Post-Training

arXiv - Machine Learning 3 min read

About this article

Abstract page for arXiv paper 2604.04855: The Role of Generator Access in Autoregressive Post-Training

Computer Science > Machine Learning arXiv:2604.04855 (cs) [Submitted on 6 Apr 2026] Title:The Role of Generator Access in Autoregressive Post-Training Authors:Amit Kiran Rege View a PDF of the paper titled The Role of Generator Access in Autoregressive Post-Training, by Amit Kiran Rege View PDF HTML (experimental) Abstract:We study how generator access constrains autoregressive post-training. The central question is whether the learner is confined to fresh root-start rollouts or can return to previously built prefixes and query the next-token rule there. In the root-start regime, output sampling, generated-token log probabilities, top-$k$ reports, and full next-token distributions along sampled trajectories all reduce to one canonical experiment, limited by the on-policy probability of reaching informative prefixes. Weak prefix control breaks this barrier, and once control is available, richer observations such as conditional sampling or logits can outperform top-$1$ access. Changing only the generator interface creates an exponential gap for KL-regularized outcome-reward post-training. Comments: Subjects: Machine Learning (cs.LG) Cite as: arXiv:2604.04855 [cs.LG]   (or arXiv:2604.04855v1 [cs.LG] for this version)   https://doi.org/10.48550/arXiv.2604.04855 Focus to learn more arXiv-issued DOI via DataCite (pending registration) Submission history From: Amit Kiran Rege [view email] [v1] Mon, 6 Apr 2026 16:58:20 UTC (39 KB) Full-text links: Access Paper: View a PDF of the p...

Originally published on April 07, 2026. Curated by AI News.

Related Articles

Improving AI models’ ability to explain their predictions
Machine Learning

Improving AI models’ ability to explain their predictions

AI News - General · 9 min ·
New technique makes AI models leaner and faster while they’re still learning
Machine Learning

New technique makes AI models leaner and faster while they’re still learning

AI News - General · 9 min ·
Machine Learning

Question regarding Transformer's pipeline module [D]

from transformers import pipeline , DistilBertTokenizer , DistilBertModel model = DistilBertModel . from_pretrained ('distilbert-base-cas...

Reddit - Machine Learning · 1 min ·
Llms

Could the best LLM be able to generate a symbolic AI that is superior to itself, or is there something superior about matrices vs graphs?

Deep neural network AIs have beaten symbolic AIs across the board on many tasks, but is there a chance that symbolic AIs written by DNNs(...

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime