[2505.13963] Through a Compressed Lens: Investigating The Impact of Quantization on Factual Knowledge Recall

[2505.13963] Through a Compressed Lens: Investigating The Impact of Quantization on Factual Knowledge Recall

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2505.13963: Through a Compressed Lens: Investigating The Impact of Quantization on Factual Knowledge Recall

Computer Science > Computation and Language arXiv:2505.13963 (cs) [Submitted on 20 May 2025 (v1), last revised 29 Apr 2026 (this version, v3)] Title:Through a Compressed Lens: Investigating The Impact of Quantization on Factual Knowledge Recall Authors:Qianli Wang, Mingyang Wang, Nils Feldhus, Simon Ostermann, Yuan Cao, Hinrich Schütze, Sebastian Möller, Vera Schmitt View a PDF of the paper titled Through a Compressed Lens: Investigating The Impact of Quantization on Factual Knowledge Recall, by Qianli Wang and 7 other authors View PDF HTML (experimental) Abstract:Quantization methods are widely used to accelerate inference and streamline the deployment of large language models (LLMs). Although quantization's effects on various LLM capabilities have been extensively studied, one critical area remains underexplored: factual knowledge recall (FKR), the process by which LLMs access stored knowledge. To this end, we conduct comprehensive experiments using three common quantization techniques at distinct bit widths, in conjunction with interpretability-driven analyses on two tasks, knowledge memorization and latent multi-hop reasoning. We show that quantization typically results in information loss within LLMs, consequently diminishing their capacity for FKR. This effect is particularly amplified in smaller models within the same architectural families. However, models quantized at reduced bit precision do not consistently exhibit inferior performance and occasionally quantizat...

Originally published on April 30, 2026. Curated by AI News.

Related Articles

[2604.16552] Co-generation of Layout and Shape from Text via Autoregressive 3D Diffusion
Llms

[2604.16552] Co-generation of Layout and Shape from Text via Autoregressive 3D Diffusion

Abstract page for arXiv paper 2604.16552: Co-generation of Layout and Shape from Text via Autoregressive 3D Diffusion

arXiv - AI · 4 min ·
[2604.17612] Provable Coordination for LLM Agents via Message Sequence Charts
Llms

[2604.17612] Provable Coordination for LLM Agents via Message Sequence Charts

Abstract page for arXiv paper 2604.17612: Provable Coordination for LLM Agents via Message Sequence Charts

arXiv - AI · 3 min ·
[2603.12249] SciMDR: Advancing Scientific Multimodal Document Reasoning
Llms

[2603.12249] SciMDR: Advancing Scientific Multimodal Document Reasoning

Abstract page for arXiv paper 2603.12249: SciMDR: Advancing Scientific Multimodal Document Reasoning

arXiv - AI · 3 min ·
[2512.03992] Value-Guided Iterative Refinement and the DIQ-H Benchmark for Evaluating VLM Robustness
Llms

[2512.03992] Value-Guided Iterative Refinement and the DIQ-H Benchmark for Evaluating VLM Robustness

Abstract page for arXiv paper 2512.03992: Value-Guided Iterative Refinement and the DIQ-H Benchmark for Evaluating VLM Robustness

arXiv - AI · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime