[2603.26954] High dimensional theory of two-phase optimizers
About this article
Abstract page for arXiv paper 2603.26954: High dimensional theory of two-phase optimizers
Computer Science > Machine Learning arXiv:2603.26954 (cs) [Submitted on 27 Mar 2026] Title:High dimensional theory of two-phase optimizers Authors:Atish Agarwala View a PDF of the paper titled High dimensional theory of two-phase optimizers, by Atish Agarwala View PDF HTML (experimental) Abstract:The trend towards larger training setups has brought a renewed interest in partially asynchronous two-phase optimizers which optimize locally and then synchronize across workers. Additionally, recent work suggests that the one-worker version of one of these algorithms, DiLoCo, shows promising results as a (synchronous) optimizer. Motivated by these studies we present an analysis of LA-DiLoCo, a simple member of the DiLoCo family, on a high-dimensional linear regression problem. We show that the one-worker variant, LA, provides a different tradeoff between signal and noise than SGD, which is beneficial in many scenarios. We also show that the multi-worker version generates more noise than the single worker version, but that this additional noise generation can be ameliorated by appropriate choice of hyperparameters. We conclude with an analysis of SLA -- LA with momentum -- and show that stacking two momentum operators gives an opportunity for acceleration via a non-linear transformation of the "effective'' Hessian spectrum, which is maximized for Nesterov momentum. Altogether our results show that two-phase optimizers represent a fruitful new paradigm for understanding and improvi...