[2602.15925] Robust Stochastic Gradient Posterior Sampling with Lattice Based Discretisation

[2602.15925] Robust Stochastic Gradient Posterior Sampling with Lattice Based Discretisation

arXiv - Machine Learning 3 min read Article

Summary

The paper presents a novel method, Stochastic Gradient Lattice Random Walk (SGLRW), to enhance Bayesian posterior sampling by improving robustness against minibatch size and gradient noise.

Why It Matters

This research addresses critical challenges in scalable Bayesian methods, particularly the sensitivity of stochastic-gradient MCMC techniques to noise and minibatch size. By proposing SGLRW, the authors contribute to more stable and efficient Bayesian inference, which is essential for advancing machine learning applications.

Key Takeaways

  • SGLRW improves robustness to minibatch size and gradient noise.
  • The method retains asymptotic correctness while enhancing stability.
  • Experimental results show SGLRW outperforms SGLD in challenging scenarios.

Statistics > Machine Learning arXiv:2602.15925 (stat) [Submitted on 17 Feb 2026] Title:Robust Stochastic Gradient Posterior Sampling with Lattice Based Discretisation Authors:Zier Mensch, Lars Holdijk, Samuel Duffield, Maxwell Aifer, Patrick J. Coles, Max Welling, Miranda C. N. Cheng View a PDF of the paper titled Robust Stochastic Gradient Posterior Sampling with Lattice Based Discretisation, by Zier Mensch and 6 other authors View PDF HTML (experimental) Abstract:Stochastic-gradient MCMC methods enable scalable Bayesian posterior sampling but often suffer from sensitivity to minibatch size and gradient noise. To address this, we propose Stochastic Gradient Lattice Random Walk (SGLRW), an extension of the Lattice Random Walk discretization. Unlike conventional Stochastic Gradient Langevin Dynamics (SGLD), SGLRW introduces stochastic noise only through the off-diagonal elements of the update covariance; this yields greater robustness to minibatch size while retaining asymptotic correctness. Furthermore, as comparison we analyze a natural analogue of SGLD utilizing gradient clipping. Experimental validation on Bayesian regression and classification demonstrates that SGLRW remains stable in regimes where SGLD fails, including in the presence of heavy-tailed gradient noise, and matches or improves predictive performance. Subjects: Machine Learning (stat.ML); Machine Learning (cs.LG) Cite as: arXiv:2602.15925 [stat.ML]   (or arXiv:2602.15925v1 [stat.ML] for this version)   htt...

Related Articles

Llms

[R] The Lyra Technique — A framework for interpreting internal cognitive states in LLMs (Zenodo, open access)

We're releasing a paper on a new framework for reading and interpreting the internal cognitive states of large language models: "The Lyra...

Reddit - Machine Learning · 1 min ·
Machine Learning

[P] citracer: a small CLI tool to trace where a concept comes from in a citation graph

Hi all, I made a small tool that I've been using for my own literature reviews and figured I'd share in case it's useful to anyone else. ...

Reddit - Machine Learning · 1 min ·
Llms

Looking to build a production-level AI/ML project (agentic systems), need guidance on what to build

Hi everyone, I’m a final-year undergraduate AI/ML student currently focusing on applied AI / agentic systems. So far, I’ve spent time und...

Reddit - ML Jobs · 1 min ·
Meta is reentering the AI race with a new model called Muse Spark | The Verge
Machine Learning

Meta is reentering the AI race with a new model called Muse Spark | The Verge

Meta Superintelligence Labs has unveiled a new AI model called Muse Spark that will soon roll out across apps like Instagram and Facebook.

The Verge - AI · 5 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime