[2604.08591] From Dispersion to Attraction: Spectral Dynamics of Hallucination Across Whisper Model Scales

[2604.08591] From Dispersion to Attraction: Spectral Dynamics of Hallucination Across Whisper Model Scales

arXiv - AI 3 min read

About this article

Abstract page for arXiv paper 2604.08591: From Dispersion to Attraction: Spectral Dynamics of Hallucination Across Whisper Model Scales

Computer Science > Machine Learning arXiv:2604.08591 (cs) [Submitted on 31 Mar 2026] Title:From Dispersion to Attraction: Spectral Dynamics of Hallucination Across Whisper Model Scales Authors:Ivan Viakhirev, Kirill Borodin, Grach Mkrtchian View a PDF of the paper titled From Dispersion to Attraction: Spectral Dynamics of Hallucination Across Whisper Model Scales, by Ivan Viakhirev and 2 other authors View PDF HTML (experimental) Abstract:Hallucinations in large ASR models present a critical safety risk. In this work, we propose the \textit{Spectral Sensitivity Theorem}, which predicts a phase transition in deep networks from a dispersive regime (signal decay) to an attractor regime (rank-1 collapse) governed by layer-wise gain and alignment. We validate this theory by analyzing the eigenspectra of activation graphs in Whisper models (Tiny to Large-v3-Turbo) under adversarial stress. Our results confirm the theoretical prediction: intermediate models exhibit \textit{Structural Disintegration} (Regime I), characterized by a $13.4\%$ collapse in Cross-Attention rank. Conversely, large models enter a \textit{Compression-Seeking Attractor} state (Regime II), where Self-Attention actively compresses rank ($-2.34\%$) and hardens the spectral slope, decoupling the model from acoustic evidence. Comments: Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI) Cite as: arXiv:2604.08591 [cs.LG]   (or arXiv:2604.08591v1 [cs.LG] for this version)   https://doi.org/10.48550...

Originally published on April 13, 2026. Curated by AI News.

Related Articles

Machine Learning

How much can a video generated by the same diffusion model differ across GPU architectures if the initial noise latent is fixed? [D]

Hi! I am trying to sanity-check an assumption for diffusion video generation reproducibility. Suppose I run the same video diffusion mode...

Reddit - Machine Learning · 1 min ·
Llms

I am not an "anti" like this guy, but still an interesting video of person interacting with chat 4o

(Posting Here because removed by Chatgpt Complaints moderators because the model here is 4o, and refuse to believe there were any safety ...

Reddit - Artificial Intelligence · 1 min ·
Llms

Unsolved AI Mystery Is Solved Along With Lessons Learned On Why ChatGPT Became Oddly Obsessed With Gremlins And Goblins

This article discusses the resolution of an AI mystery regarding ChatGPT's unusual focus on gremlins and goblins, along with insights gai...

AI Tools & Products · 1 min ·
[2602.06869] Uncovering Cross-Objective Interference in Multi-Objective Alignment
Llms

[2602.06869] Uncovering Cross-Objective Interference in Multi-Objective Alignment

Abstract page for arXiv paper 2602.06869: Uncovering Cross-Objective Interference in Multi-Objective Alignment

arXiv - Machine Learning · 3 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime