[2603.25035] Mechanistically Interpreting Compression in Vision-Language Models
About this article
Abstract page for arXiv paper 2603.25035: Mechanistically Interpreting Compression in Vision-Language Models
Computer Science > Artificial Intelligence arXiv:2603.25035 (cs) [Submitted on 26 Mar 2026] Title:Mechanistically Interpreting Compression in Vision-Language Models Authors:Veeraraju Elluru, Arth Singh, Roberto Aguero, Ajay Agarwal, Debojyoti Das, Hreetam Paul View a PDF of the paper titled Mechanistically Interpreting Compression in Vision-Language Models, by Veeraraju Elluru and 5 other authors View PDF HTML (experimental) Abstract:Compressed vision-language models (VLMs) are widely used to reduce memory and compute costs, making them a suitable choice for real-world deployment. However, compressing these models raises concerns about whether internal computations and safety behaviors are preserved. In this work, we use causal circuit analysis and crosscoder-based feature comparisons to examine how pruning and quantization fundamentally change the internals across representative VLMs. We observe that pruning generally keeps circuit structure intact but rotates and attenuates internal features, while quantization modifies the circuits at a higher level yet leaves the surviving features better aligned. Leveraging this insight, we also introduce VLMSafe-420, a novel benchmark that pairs harmful inputs with matched benign counterfactuals across various safety categories. Our findings show that pruning causes a sharp drop in genuine refusal behavior, suggesting that the choice of compression has safety implications. Comments: Subjects: Artificial Intelligence (cs.AI) Cite as: ...