[2604.02505] Optimal Projection-Free Adaptive SGD for Matrix Optimization
About this article
Abstract page for arXiv paper 2604.02505: Optimal Projection-Free Adaptive SGD for Matrix Optimization
Mathematics > Optimization and Control arXiv:2604.02505 (math) [Submitted on 2 Apr 2026] Title:Optimal Projection-Free Adaptive SGD for Matrix Optimization Authors:Dmitry Kovalev View a PDF of the paper titled Optimal Projection-Free Adaptive SGD for Matrix Optimization, by Dmitry Kovalev View PDF HTML (experimental) Abstract:Recently, Jiang et al. [2026] developed Leon, a practical variant of One-sided Shampoo [Xie et al., 2025a, An et al., 2025] algorithm for online convex optimization, which does not require computing a costly quadratic projection at each iteration. Unfortunately, according to the existing analysis, Leon requires tuning an additional hyperparameter in its preconditioner and cannot achieve dimension-independent convergence guarantees for convex optimization problems beyond the bounded gradients assumption. In this paper, we resolve this issue by proving certain stability properties of Leon's preconditioner. Using our improved analysis, we show that tuning the extra hyperparameter can be avoided and, more importantly, develop the first practical variant of One-sided Shampoo with Nesterov acceleration, which does not require computing projections at each iteration. As a side contribution, we obtain improved dimension-independent rates in the non-smooth non-convex setting and develop a unified analysis of the proposed algorithm, which yields accelerated projection-free adaptive SGD with (block-)diagonal preconditioners. Subjects: Optimization and Control (m...