[2603.03843] Invariance-Based Dynamic Regret Minimization
About this article
Abstract page for arXiv paper 2603.03843: Invariance-Based Dynamic Regret Minimization
Statistics > Machine Learning arXiv:2603.03843 (stat) [Submitted on 4 Mar 2026] Title:Invariance-Based Dynamic Regret Minimization Authors:Margherita Lazzaretto, Jonas Peters, Niklas Pfister View a PDF of the paper titled Invariance-Based Dynamic Regret Minimization, by Margherita Lazzaretto and 2 other authors View PDF HTML (experimental) Abstract:We consider stochastic non-stationary linear bandits where the linear parameter connecting contexts to the reward changes over time. Existing algorithms in this setting localize the policy by gradually discarding or down-weighting past data, effectively shrinking the time horizon over which learning can occur. However, in many settings historical data may still carry partial information about the reward model. We propose to leverage such data while adapting to changes, by assuming the reward model decomposes into stationary and non-stationary components. Based on this assumption, we introduce ISD-linUCB, an algorithm that uses past data to learn invariances in the reward model and subsequently exploits them to improve online performance. We show both theoretically and empirically that leveraging invariance reduces the problem dimensionality, yielding significant regret improvements in fast-changing environments when sufficient historical data is available. Comments: Subjects: Machine Learning (stat.ML); Machine Learning (cs.LG) Cite as: arXiv:2603.03843 [stat.ML] (or arXiv:2603.03843v1 [stat.ML] for this version) https://doi...