[2602.21036] Empirically Calibrated Conditional Independence Tests

[2602.21036] Empirically Calibrated Conditional Independence Tests

arXiv - Machine Learning 3 min read Article

Summary

The paper presents Empirically Calibrated Conditional Independence Tests (ECCIT), a method designed to enhance the reliability of conditional independence tests in statistical analysis, particularly addressing issues of miscalibration in small and large samples.

Why It Matters

Conditional independence tests are crucial for causal discovery and feature selection in statistics and machine learning. The proposed ECCIT method aims to improve the accuracy of these tests, which is essential for researchers relying on statistical validity in their analyses. By correcting miscalibration, ECCIT offers a more robust framework for obtaining valid results, thereby enhancing the integrity of findings in various applications.

Key Takeaways

  • ECCIT addresses common failure modes in conditional independence tests.
  • The method provides valid false discovery rate control with higher power.
  • ECCIT is test agnostic, making it applicable across various models.
  • Empirical benchmarks demonstrate ECCIT's effectiveness over existing strategies.
  • The approach is particularly beneficial for small sample sizes and misspecified models.

Statistics > Methodology arXiv:2602.21036 (stat) [Submitted on 24 Feb 2026] Title:Empirically Calibrated Conditional Independence Tests Authors:Milleno Pan, Antoine de Mathelin, Wesley Tansey View a PDF of the paper titled Empirically Calibrated Conditional Independence Tests, by Milleno Pan and 2 other authors View PDF HTML (experimental) Abstract:Conditional independence tests (CIT) are widely used for causal discovery and feature selection. Even with false discovery rate (FDR) control procedures, they often fail to provide frequentist guarantees in practice. We highlight two common failure modes: (i) in small samples, asymptotic guarantees for many CITs can be inaccurate and even correctly specified models fail to estimate the noise levels and control the error, and (ii) when sample sizes are large but models are misspecified, unaccounted dependencies skew the test's behavior and fail to return uniform p-values under the null. We propose Empirically Calibrated Conditional Independence Tests (ECCIT), a method that measures and corrects for miscalibration. For a chosen base CIT (e.g., GCM, HRT), ECCIT optimizes an adversary that selects features and response functions to maximize a miscalibration metric. ECCIT then fits a monotone calibration map that adjusts the base-test p-values in proportion to the observed miscalibration. Across empirical benchmarks on synthetic and real data, ECCIT achieves valid FDR with higher power than existing calibration strategies while remai...

Related Articles

Machine Learning

[D] ICML 2026 Average Score

Hi all, I’m curious about the current review dynamics for ICML 2026, especially after the rebuttal phase. For those who are reviewers (or...

Reddit - Machine Learning · 1 min ·
Machine Learning

[R] VOID: Video Object and Interaction Deletion (physically-consistent video inpainting)

We present VOID, a model for video object removal that aims to handle *physical interactions*, not just appearance. Most existing video i...

Reddit - Machine Learning · 1 min ·
Machine Learning

FLUX 2 Pro (2026) Sketch to Image

I sketched a cow and tested how different models interpret it into a realistic image for downstream 3D generation, turns out some models ...

Reddit - Artificial Intelligence · 1 min ·
Improving AI models’ ability to explain their predictions
Machine Learning

Improving AI models’ ability to explain their predictions

AI News - General · 9 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime