[2512.20007] Semiparametric KSD test: unifying score and distance-based approaches for goodness-of-fit testing

[2512.20007] Semiparametric KSD test: unifying score and distance-based approaches for goodness-of-fit testing

arXiv - Machine Learning 4 min read Article

Summary

The paper introduces the Semiparametric Kernelized Stein Discrepancy (SKSD) test, a novel goodness-of-fit test that unifies score-based and distance-based approaches, demonstrating efficiency and consistency in model adequacy assessments.

Why It Matters

Goodness-of-fit tests are critical for validating statistical models. The SKSD test offers a new, efficient method that can handle complex models with intractable likelihoods, making it relevant for researchers and practitioners in machine learning and statistics.

Key Takeaways

  • The SKSD test combines score-based and distance-based approaches for improved goodness-of-fit testing.
  • It is computationally efficient and can accommodate general nuisance-parameter estimators.
  • The test is universally consistent and achieves Pitman efficiency.
  • It provides effective GoF tests for models with intractable likelihoods.
  • The SKSD test demonstrates comparable power to traditional normality tests.

Statistics > Machine Learning arXiv:2512.20007 (stat) [Submitted on 23 Dec 2025 (v1), last revised 20 Feb 2026 (this version, v2)] Title:Semiparametric KSD test: unifying score and distance-based approaches for goodness-of-fit testing Authors:Zhihan Huang, Ziang Niu View a PDF of the paper titled Semiparametric KSD test: unifying score and distance-based approaches for goodness-of-fit testing, by Zhihan Huang and 1 other authors View PDF HTML (experimental) Abstract:Goodness-of-fit (GoF) tests are fundamental for assessing model adequacy. Score-based tests are appealing because they require fitting the model only once under the null. However, extending them to powerful nonparametric alternatives is difficult due to the lack of suitable score functions. Through a class of exponentially tilted models, we show that the resulting score-based GoF tests are equivalent to the tests based on integral probability metrics (IPMs) indexed by a function class. When the class is rich, the test is universally consistent. This simple yet insightful perspective enables reinterpretation of classical distance-based testing procedures-including those based on Kolmogorov-Smirnov distance, Wasserstein-1 distance, and maximum mean discrepancy-as arising from score-based constructions. Building on this insight, we propose a new nonparametric score-based GoF test through a special class of IPM induced by kernelized Stein's function class, called semiparametric kernelized Stein discrepancy (SKSD) t...

Related Articles

Machine Learning

[D] ICML reviewer making up false claim in acknowledgement, what to do?

In a rebuttal acknowledgement we received, the reviewer made up a claim that our method performs worse than baselines with some hyperpara...

Reddit - Machine Learning · 1 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
Machine Learning

[D] Budget Machine Learning Hardware

Looking to get into machine learning and found this video on a piece of hardware for less than £500. Is it really possible to teach auton...

Reddit - Machine Learning · 1 min ·
Machine Learning

Your prompts aren’t the problem — something else is

I keep seeing people focus heavily on prompt optimization. But in practice, a lot of failures I’ve observed don’t come from the prompt it...

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime