[2602.21765] Generalisation of RLHF under Reward Shift and Clipped KL Regularisation
Summary
This paper explores the generalization of Reinforcement Learning from Human Feedback (RLHF) under conditions of reward shift and clipped KL regularization, providing theoretical insights and practical implications.
Why It Matters
Understanding the generalization of RLHF is crucial for improving the alignment and adaptability of large language models. This research addresses gaps in the theoretical framework, particularly regarding how reward shifts and regularization techniques affect model performance, which is vital for developing more robust AI systems.
Key Takeaways
- The paper develops a generalization theory for RLHF that accounts for reward shifts and clipped KL regularization.
- Generalization error in RLHF arises from sampling errors, reward shifts, and KL clipping errors.
- Practical implications include determining optimal KL clipping thresholds and budget allocation for prompts and rollouts.
Computer Science > Machine Learning arXiv:2602.21765 (cs) [Submitted on 25 Feb 2026] Title:Generalisation of RLHF under Reward Shift and Clipped KL Regularisation Authors:Kenton Tang, Yuzhu Chen, Fengxiang He View a PDF of the paper titled Generalisation of RLHF under Reward Shift and Clipped KL Regularisation, by Kenton Tang and 2 other authors View PDF HTML (experimental) Abstract:Alignment and adaptation in large language models heavily rely on reinforcement learning from human feedback (RLHF); yet, theoretical understanding of its generalisability remains premature, especially when the learned reward could shift, and the KL control is estimated and clipped. To address this issue, we develop generalisation theory for RLHF that explicitly accounts for (1) \emph{reward shift}: reward models are trained on preference data from earlier or mixed behaviour policies while RLHF optimises the current policy on its own rollouts; and (2) \emph{clipped KL regularisation}: the KL regulariser is estimated from sampled log-probability ratios and then clipped for stabilisation, resulting in an error to RLHF. We present generalisation bounds for RLHF, suggesting that the generalisation error stems from a sampling error from prompts and rollouts, a reward shift error, and a KL clipping error. We also discuss special cases of (1) initialising RLHF parameters with a uniform prior over a finite space, and (2) training RLHF by stochastic gradient descent, as an Ornstein-Uhlenbeck process. Th...