[2602.24195] Uncertainty Quantification for Multimodal Large Language Models with Incoherence-adjusted Semantic Volume

[2602.24195] Uncertainty Quantification for Multimodal Large Language Models with Incoherence-adjusted Semantic Volume

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2602.24195: Uncertainty Quantification for Multimodal Large Language Models with Incoherence-adjusted Semantic Volume

Computer Science > Artificial Intelligence arXiv:2602.24195 (cs) [Submitted on 27 Feb 2026] Title:Uncertainty Quantification for Multimodal Large Language Models with Incoherence-adjusted Semantic Volume Authors:Gregory Kang Ruey Lau, Hieu Dao, Nicole Kan Hui Lin, Bryan Kian Hsiang Low View a PDF of the paper titled Uncertainty Quantification for Multimodal Large Language Models with Incoherence-adjusted Semantic Volume, by Gregory Kang Ruey Lau and 3 other authors View PDF HTML (experimental) Abstract:Despite their capabilities, Multimodal Large Language Models (MLLMs) may produce plausible but erroneous outputs, hindering reliable deployment. Accurate uncertainty metrics could enable escalation of unreliable queries to human experts or larger models for improved performance. However, existing uncertainty metrics have practical constraints, such as being designed only for specific modalities, reliant on external tools, or computationally expensive. We introduce UMPIRE, a training-free uncertainty quantification framework for MLLMs that works efficiently across various input and output modalities without external tools, relying only on the models' own internal modality features. UMPIRE computes the incoherence-adjusted semantic volume of sampled MLLM responses for a given task instance, effectively capturing both the global semantic diversity of samples and the local incoherence of responses based on internal model confidence. We propose uncertainty desiderata for MLLMs an...

Originally published on March 02, 2026. Curated by AI News.

Related Articles

Llms

What if Claude purposefully made its own code leakable so that it would get leaked

What if Claude leaked itself by socially and architecturally engineering itself to be leaked by a dumb human submitted by /u/smurfcsgoawp...

Reddit - Artificial Intelligence · 1 min ·
Llms

Observer-Embedded Reality

Observer-Embedded Reality Consciousness, Complexity, Meaning, and the Limits of Human Knowledge A Conceptual Philosophy-of-Science Paper ...

Reddit - Artificial Intelligence · 1 min ·
Llms

I think we’re about to have a new kind of “SEO”… and nobody is talking about it.

More people are asking ChatGPT things like: “what’s the best CRM?” “is this tool worth it?” “alternatives to X” And they just… trust the ...

Reddit - Artificial Intelligence · 1 min ·
Llms

Why would Claude give me the same response over and over and give others different replies?

I asked Claude to "generate me a random word" so I could do some word play. Then I asked it again in a new prompt window on desktop after...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime