[2603.29078] PolarQuant: Optimal Gaussian Weight Quantization via Hadamard Rotation for LLM Compression

[2603.29078] PolarQuant: Optimal Gaussian Weight Quantization via Hadamard Rotation for LLM Compression

arXiv - Machine Learning 3 min read

About this article

Abstract page for arXiv paper 2603.29078: PolarQuant: Optimal Gaussian Weight Quantization via Hadamard Rotation for LLM Compression

Computer Science > Computation and Language arXiv:2603.29078 (cs) This paper has been withdrawn by Caio Vicentino [Submitted on 30 Mar 2026 (v1), last revised 20 Apr 2026 (this version, v2)] Title:PolarQuant: Optimal Gaussian Weight Quantization via Hadamard Rotation for LLM Compression Authors:Caio Vicentino View a PDF of the paper titled PolarQuant: Optimal Gaussian Weight Quantization via Hadamard Rotation for LLM Compression, by Caio Vicentino No PDF available, click to view other formats Abstract:We present PolarQuant, a post-training weight quantization method for large language models (LLMs) that exploits the distributional structure of neural network weights to achieve near-lossless compression. PolarQuant operates in three stages: (1) block-wise normalization to the unit hypersphere, (2) Walsh-Hadamard rotation to transform coordinates into approximately Gaussian random variables, and (3) quantization with centroids matched to the Gaussian distribution. Our ablation reveals that Hadamard rotation alone accounts for 98% of the quality improvement, reducing Qwen3.5-9B perplexity from 6.90 (absmax Q5) to 6.40 (Delta = +0.03 from FP16), making it practically lossless without any calibration data. Furthermore, PolarQuant functions as an effective preprocessing step for downstream INT4 quantizers: PolarQuant Q5 dequantized and re-quantized by torchao INT4 achieves perplexity 6.56 versus 6.68 for direct absmax INT4, while maintaining 43.1 tok/s throughput at 6.5 GB VRAM....

Originally published on April 22, 2026. Curated by AI News.

Related Articles

I tried Gemini, ChatGPT, and Claude for a month on Android, and I have a clear winner for you
Llms

I tried Gemini, ChatGPT, and Claude for a month on Android, and I have a clear winner for you

The ultimate Android AI showdown

AI Tools & Products · 5 min ·
[2602.20409] CLIPoint3D: Language-Grounded Few-Shot Unsupervised 3D Point Cloud Domain Adaptation
Llms

[2602.20409] CLIPoint3D: Language-Grounded Few-Shot Unsupervised 3D Point Cloud Domain Adaptation

Abstract page for arXiv paper 2602.20409: CLIPoint3D: Language-Grounded Few-Shot Unsupervised 3D Point Cloud Domain Adaptation

arXiv - Machine Learning · 4 min ·
[2602.11199] When and What to Ask: AskBench and Rubric-Guided RLVR for LLM Clarification
Llms

[2602.11199] When and What to Ask: AskBench and Rubric-Guided RLVR for LLM Clarification

Abstract page for arXiv paper 2602.11199: When and What to Ask: AskBench and Rubric-Guided RLVR for LLM Clarification

arXiv - Machine Learning · 3 min ·
[2512.23805] Fitted Q Evaluation Without Bellman Completeness via Stationary Weighting
Llms

[2512.23805] Fitted Q Evaluation Without Bellman Completeness via Stationary Weighting

Abstract page for arXiv paper 2512.23805: Fitted Q Evaluation Without Bellman Completeness via Stationary Weighting

arXiv - Machine Learning · 3 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime