[2602.18946] Exponential Convergence of (Stochastic) Gradient Descent for Separable Logistic Regression

[2602.18946] Exponential Convergence of (Stochastic) Gradient Descent for Separable Logistic Regression

arXiv - Machine Learning 4 min read Article

Summary

This paper presents a novel approach to gradient descent and stochastic gradient descent, demonstrating exponential convergence for separable logistic regression using a stable step-size schedule.

Why It Matters

Understanding the convergence behavior of gradient descent methods is crucial for optimizing machine learning algorithms. This research provides insights that can enhance the efficiency of training models, particularly in scenarios where stability is a concern, thus broadening the applicability of these methods in real-world applications.

Key Takeaways

  • Exponential convergence can be achieved with a stable step-size schedule.
  • The proposed method does not require prior knowledge of optimization horizons or target accuracy.
  • Stochastic gradient descent can also achieve exponential convergence using a lightweight adaptive step-size rule.

Computer Science > Machine Learning arXiv:2602.18946 (cs) [Submitted on 21 Feb 2026] Title:Exponential Convergence of (Stochastic) Gradient Descent for Separable Logistic Regression Authors:Sacchit Kale, Piyushi Manupriya, Pierre Marion, Francis bach, Anant Raj View a PDF of the paper titled Exponential Convergence of (Stochastic) Gradient Descent for Separable Logistic Regression, by Sacchit Kale and 3 other authors View PDF HTML (experimental) Abstract:Gradient descent and stochastic gradient descent are central to modern machine learning, yet their behavior under large step sizes remains theoretically unclear. Recent work suggests that acceleration often arises near the edge of stability, where optimization trajectories become unstable and difficult to analyze. Existing results for separable logistic regression achieve faster convergence by explicitly leveraging such unstable regimes through constant or adaptive large step sizes. In this paper, we show that instability is not inherent to acceleration. We prove that gradient descent with a simple, non-adaptive increasing step-size schedule achieves exponential convergence for separable logistic regression under a margin condition, while remaining entirely within a stable optimization regime. The resulting method is anytime and does not require prior knowledge of the optimization horizon or target accuracy. We also establish exponential convergence of stochastic gradient descent using a lightweight adaptive step-size rule...

Related Articles

Llms

Study: LLMs Able to De-Anonymize User Accounts on Reddit, Hacker News & Other "Pseudonymous" Platforms; Report Co-Author Expands, Advises

Advice from the study's co-author: "Be aware that it’s not any single post that identifies you, but the combination of small details acro...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[D] Best websites for pytorch/numpy interviews

Hello, I’m at the last year of my PHD and I’m starting to prepare interviews. I’m mainly aiming at applied scientist/research engineer or...

Reddit - Machine Learning · 1 min ·
Llms

[P] Remote sensing foundation models made easy to use.

This project enables the idea of tasking remote sensing models to acquire embeddings like we task satellites to acquire data! https://git...

Reddit - Machine Learning · 1 min ·
Machine Learning

Can AI truly be creative?

AI has no imagination. “Creativity is the ability to generate novel and valuable ideas or works through the exercise of imagination” http...

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime