[2510.18259] Learning under Quantization for High-Dimensional Linear Regression

[2510.18259] Learning under Quantization for High-Dimensional Linear Regression

arXiv - Machine Learning 4 min read Article

Summary

This paper explores the impact of low-bit quantization on high-dimensional linear regression, providing a theoretical framework for understanding its effects on learning performance.

Why It Matters

As machine learning models grow in complexity, efficient training techniques like quantization become crucial. This study offers insights into how different quantization methods affect learning dynamics, which can inform better model training under hardware constraints.

Key Takeaways

  • Quantization can amplify noise during training, affecting learning outcomes.
  • Different quantization schemes (additive vs. multiplicative) have distinct impacts on data distortion and noise amplification.
  • The study provides theoretical bounds that characterize the risks associated with various quantization methods.
  • Understanding quantization's effects can lead to improved optimization algorithms.
  • This research lays the groundwork for further exploration of learning theory in practical hardware contexts.

Statistics > Machine Learning arXiv:2510.18259 (stat) [Submitted on 21 Oct 2025 (v1), last revised 16 Feb 2026 (this version, v3)] Title:Learning under Quantization for High-Dimensional Linear Regression Authors:Dechen Zhang, Junwei Su, Difan Zou View a PDF of the paper titled Learning under Quantization for High-Dimensional Linear Regression, by Dechen Zhang and 2 other authors View PDF Abstract:The use of low-bit quantization has emerged as an indispensable technique for enabling the efficient training of large-scale models. Despite its widespread empirical success, a rigorous theoretical understanding of its impact on learning performance remains notably absent, even in the simplest linear regression setting. We present the first systematic theoretical study of this fundamental question, analyzing finite-step stochastic gradient descent (SGD) for high-dimensional linear regression under a comprehensive range of quantization targets: data, label, parameter, activation, and gradient. Our novel analytical framework establishes precise algorithm-dependent and data-dependent excess risk bounds that characterize how different quantization affects learning: parameter, activation, and gradient quantization amplify noise during training; data quantization distorts the data spectrum and introduces additional approximation error. Crucially, we distinguish the effects of two quantization schemes: we prove that for additive quantization (with constant quantization steps), the noise ...

Related Articles

Machine Learning

[For Hire] Ex-Microsoft Senior Data Engineer | Databricks, Palantir Foundry, MLOps | $55/hr

submitted by /u/mcheetirala2510 [link] [comments]

Reddit - ML Jobs · 1 min ·
Meta AI app climbs to No. 5 on the App Store after Muse Spark launch | TechCrunch
Machine Learning

Meta AI app climbs to No. 5 on the App Store after Muse Spark launch | TechCrunch

The app was ranking No. 57 on the App Store just before Meta AI's new model launched. Now it's No. 5 — and rising.

TechCrunch - AI · 4 min ·
Machine Learning

Detecting mirrored selfie images: OCR the best way? [D]

I'm trying to catch backwards "selfie" images before passing them to our VLM text reader and/or face embedding extraction. Since models l...

Reddit - Machine Learning · 1 min ·
Llms

Google’s Gemini AI can answer your questions with 3D models and simulations

submitted by /u/tekz [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime