[2602.06320] High-Dimensional Limit of Stochastic Gradient Flow via Dynamical Mean-Field Theory

[2602.06320] High-Dimensional Limit of Stochastic Gradient Flow via Dynamical Mean-Field Theory

arXiv - Machine Learning 4 min read Article

Summary

This paper explores the high-dimensional dynamics of stochastic gradient flow (SGF) in machine learning, providing a closed system of equations that characterizes the asymptotic behavior of multi-pass SGD with small batch sizes.

Why It Matters

Understanding the dynamics of SGD in high dimensions is crucial for improving machine learning models. This research fills a gap in analytical frameworks, offering insights that can enhance model training and performance across various applications, including neural networks.

Key Takeaways

  • The paper derives a closed system of low-dimensional equations for SGF in high dimensions.
  • It utilizes dynamical mean-field theory to analyze the asymptotic behavior of SGD.
  • The findings unify existing frameworks and provide a broader understanding of SGD dynamics.
  • The approach is applicable to various models, including generalized linear models and neural networks.
  • The research extends existing techniques to handle stochasticity in gradient flows.

Statistics > Machine Learning arXiv:2602.06320 (stat) [Submitted on 6 Feb 2026 (v1), last revised 16 Feb 2026 (this version, v2)] Title:High-Dimensional Limit of Stochastic Gradient Flow via Dynamical Mean-Field Theory Authors:Sota Nishiyama, Masaaki Imaizumi View a PDF of the paper titled High-Dimensional Limit of Stochastic Gradient Flow via Dynamical Mean-Field Theory, by Sota Nishiyama and Masaaki Imaizumi View PDF Abstract:Modern machine learning models are typically trained via multi-pass stochastic gradient descent (SGD) with small batch sizes, and understanding their dynamics in high dimensions is of great interest. However, an analytical framework for describing the high-dimensional asymptotic behavior of multi-pass SGD with small batch sizes for nonlinear models is currently missing. In this study, we address this gap by analyzing the high-dimensional dynamics of a stochastic differential equation called a \emph{stochastic gradient flow} (SGF), which approximates multi-pass SGD in this regime. In the limit where the number of data samples $n$ and the dimension $d$ grow proportionally, we derive a closed system of low-dimensional and continuous-time equations and prove that it characterizes the asymptotic distribution of the SGF parameters. Our theory is based on the dynamical mean-field theory (DMFT) and is applicable to a wide range of models encompassing generalized linear models and two-layer neural networks. We further show that the resulting DMFT equations r...

Related Articles

Machine Learning

[For Hire] Ex-Microsoft Senior Data Engineer | Databricks, Palantir Foundry, MLOps | $55/hr

submitted by /u/mcheetirala2510 [link] [comments]

Reddit - ML Jobs · 1 min ·
Meta AI app climbs to No. 5 on the App Store after Muse Spark launch | TechCrunch
Machine Learning

Meta AI app climbs to No. 5 on the App Store after Muse Spark launch | TechCrunch

The app was ranking No. 57 on the App Store just before Meta AI's new model launched. Now it's No. 5 — and rising.

TechCrunch - AI · 4 min ·
Machine Learning

Detecting mirrored selfie images: OCR the best way? [D]

I'm trying to catch backwards "selfie" images before passing them to our VLM text reader and/or face embedding extraction. Since models l...

Reddit - Machine Learning · 1 min ·
Llms

Google’s Gemini AI can answer your questions with 3D models and simulations

submitted by /u/tekz [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime