[2508.09639] UbiQTree: Uncertainty Quantification in XAI with Tree Ensembles

[2508.09639] UbiQTree: Uncertainty Quantification in XAI with Tree Ensembles

arXiv - AI 4 min read Article

Summary

The paper presents UbiQTree, a method for decomposing uncertainty in SHAP values used in explainable AI, focusing on aleatoric and epistemic uncertainties in tree ensembles.

Why It Matters

Understanding uncertainty in AI models is crucial, especially in high-stakes fields like healthcare. This research enhances the interpretability and reliability of SHAP values, guiding better decision-making and model refinement.

Key Takeaways

  • UbiQTree decomposes SHAP value uncertainty into aleatoric and epistemic components.
  • The method integrates Dempster-Shafer theory and Dirichlet processes for improved uncertainty quantification.
  • Findings indicate that high SHAP values may not correlate with stable features.
  • Better data representation can reduce epistemic uncertainty in model predictions.
  • Tree-based models, particularly bagging, are effective for quantifying uncertainty.

Computer Science > Artificial Intelligence arXiv:2508.09639 (cs) [Submitted on 13 Aug 2025 (v1), last revised 24 Feb 2026 (this version, v2)] Title:UbiQTree: Uncertainty Quantification in XAI with Tree Ensembles Authors:Akshat Dubey, Aleksandar Anžel, Bahar İlgen, Georges Hattab View a PDF of the paper titled UbiQTree: Uncertainty Quantification in XAI with Tree Ensembles, by Akshat Dubey and 3 other authors View PDF HTML (experimental) Abstract:Explainable Artificial Intelligence (XAI) techniques, such as SHapley Additive exPlanations (SHAP), have become essential tools for interpreting complex ensemble tree-based models, especially in high-stakes domains such as healthcare analytics. However, SHAP values are usually treated as point estimates, which disregards the inherent and ubiquitous uncertainty in predictive models and data. This uncertainty has two primary sources: aleatoric and epistemic. The aleatoric uncertainty, which reflects the irreducible noise in the data. The epistemic uncertainty, which arises from a lack of data. In this work, we propose an approach for decomposing uncertainty in SHAP values into aleatoric, epistemic, and entanglement components. This approach integrates Dempster-Shafer evidence theory and hypothesis sampling via Dirichlet processes over tree ensembles. We validate the method across three real-world use cases with descriptive statistical analyses that provide insight into the nature of epistemic uncertainty embedded in SHAP explanations...

Related Articles

Llms

[P] Remote sensing foundation models made easy to use.

This project enables the idea of tasking remote sensing models to acquire embeddings like we task satellites to acquire data! https://git...

Reddit - Machine Learning · 1 min ·
Machine Learning

Can AI truly be creative?

AI has no imagination. “Creativity is the ability to generate novel and valuable ideas or works through the exercise of imagination” http...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

AI video generation seems fundamentally more expensive than text, not just less optimized

There’s been a lot of discussion recently about how expensive AI video generation is compared to text, and it feels like this is more tha...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[D] When to transition from simple heuristics to ML models (e.g., DensityFunction)?

Two questions: What are the recommendations around when to transition from a simple heuristic baseline to machine learning ML models for ...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime