[2604.08905] StaRPO: Stability-Augmented Reinforcement Policy Optimization
About this article
Abstract page for arXiv paper 2604.08905: StaRPO: Stability-Augmented Reinforcement Policy Optimization
Computer Science > Artificial Intelligence arXiv:2604.08905 (cs) [Submitted on 10 Apr 2026] Title:StaRPO: Stability-Augmented Reinforcement Policy Optimization Authors:Jinghan Zhang, Fengran Mo, Tharindu Cyril Weerasooriya, Ruimin Dai, Xiaoyan Han, Yanjie Fu, Dakuo Wang, Kunpeng Liu View a PDF of the paper titled StaRPO: Stability-Augmented Reinforcement Policy Optimization, by Jinghan Zhang and 7 other authors View PDF HTML (experimental) Abstract:Reinforcement learning (RL) is effective in enhancing the accuracy of large language models in complex reasoning tasks. Existing RL policy optimization frameworks rely on final-answer correctness as feedback signals and rarely capture the internal logical structure of the reasoning process. Consequently, the models would generate fluent and semantically relevant responses but logically inconsistent, structurally erratic, or redundant. To this end, we propose StaRPO, a stability-augmented reinforcement learning framework that explicitly incorporates reasoning stability into the optimization objective. Our StaRPO decomposes stability into two computable lightweight metrics: the Autocorrelation Function (ACF) to evaluate local step-to-step coherence, and Path Efficiency (PE) to evaluate global goal-directedness of the reasoning trajectory. These stability rewards are combined with task rewards to provide complementary and process-aware feedback. We validate the effectiveness of using ACF and PE rewards by showing their correlation ...