[2602.19893] Generalized Random Direction Newton Algorithms for Stochastic Optimization

[2602.19893] Generalized Random Direction Newton Algorithms for Stochastic Optimization

arXiv - Machine Learning 3 min read Article

Summary

This paper introduces generalized Hessian estimators for stochastic optimization using random direction stochastic approximation, demonstrating lower-order bias with increased function measurements and providing convergence analyses.

Why It Matters

The research addresses challenges in stochastic optimization, which is crucial for various machine learning applications. By improving Hessian estimators, the findings could enhance algorithm performance in noisy environments, making them more applicable in real-world scenarios.

Key Takeaways

  • Introduces generalized Hessian estimators for stochastic optimization.
  • Demonstrates that more function measurements reduce estimation bias.
  • Provides asymptotic and non-asymptotic convergence analyses.
  • Validates theoretical findings through numerical experiments.
  • Enhances understanding of stochastic Newton methods.

Computer Science > Machine Learning arXiv:2602.19893 (cs) [Submitted on 23 Feb 2026] Title:Generalized Random Direction Newton Algorithms for Stochastic Optimization Authors:Soumen Pachal, Prashanth L.A., Shalabh Bhatnagar, Avinash Achar View a PDF of the paper titled Generalized Random Direction Newton Algorithms for Stochastic Optimization, by Soumen Pachal and 3 other authors View PDF HTML (experimental) Abstract:We present a family of generalized Hessian estimators of the objective using random direction stochastic approximation (RDSA) by utilizing only noisy function measurements. The form of each estimator and the order of the bias depend on the number of function measurements. In particular, we demonstrate that estimators with more function measurements exhibit lower-order estimation bias. We show the asymptotic unbiasedness of the estimators. We also perform asymptotic and non-asymptotic convergence analyses for stochastic Newton methods that incorporate our generalized Hessian estimators. Finally, we perform numerical experiments to validate our theoretical findings. Subjects: Machine Learning (cs.LG); Machine Learning (stat.ML) Cite as: arXiv:2602.19893 [cs.LG]   (or arXiv:2602.19893v1 [cs.LG] for this version)   https://doi.org/10.48550/arXiv.2602.19893 Focus to learn more arXiv-issued DOI via DataCite (pending registration) Submission history From: Soumen Pachal [view email] [v1] Mon, 23 Feb 2026 14:33:39 UTC (35 KB) Full-text links: Access Paper: View a PDF of...

Related Articles

Ai Safety

NHS staff resist using Palantir software. Staff reportedly cite ethics concerns, privacy worries, and doubt the platform adds much

submitted by /u/esporx [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

AI assistants are optimized to seem helpful. That is not the same thing as being helpful.

RLHF trains models on human feedback. Humans rate responses they like. And it turns out humans consistently rate confident, fluent, agree...

Reddit - Artificial Intelligence · 1 min ·
Computer Vision

House Democrat Questions Anthropic on AI Safety After Source Code Leak

Rep. Josh Gottheimer, who is generally tough on China, just sent a letter to Anthropic questioning their decision to reduce certain safet...

Reddit - Artificial Intelligence · 1 min ·
[2512.21106] Semantic Refinement with LLMs for Graph Representations
Llms

[2512.21106] Semantic Refinement with LLMs for Graph Representations

Abstract page for arXiv paper 2512.21106: Semantic Refinement with LLMs for Graph Representations

arXiv - Machine Learning · 4 min ·
More in Ai Safety: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime