[2510.00463] On the Adversarial Robustness of Learning-based Conformal Novelty Detection

[2510.00463] On the Adversarial Robustness of Learning-based Conformal Novelty Detection

arXiv - Machine Learning 4 min read Article

Summary

This paper investigates the adversarial robustness of learning-based conformal novelty detection methods, revealing significant vulnerabilities that can increase false discovery rates under adversarial conditions.

Why It Matters

As machine learning models are increasingly deployed in real-world applications, understanding their vulnerabilities to adversarial attacks is crucial. This research highlights the limitations of current novelty detection methods, emphasizing the need for more robust alternatives to ensure reliability in critical applications.

Key Takeaways

  • Adversarial perturbations can significantly increase false discovery rates in novelty detection methods.
  • The study formulates an oracle attack setup to quantify the degradation of performance under adversarial conditions.
  • Two learning-based frameworks were evaluated, exposing their vulnerabilities and motivating the need for improved robustness.
  • The research provides a systematic evaluation using both synthetic and real-world datasets.
  • Findings suggest that existing error-controlled novelty detection methods have fundamental limitations.

Statistics > Machine Learning arXiv:2510.00463 (stat) [Submitted on 1 Oct 2025 (v1), last revised 20 Feb 2026 (this version, v3)] Title:On the Adversarial Robustness of Learning-based Conformal Novelty Detection Authors:Daofu Zhang, Mehrdad Pournaderi, Hanne M. Clifford, Yu Xiang, Pramod K. Varshney View a PDF of the paper titled On the Adversarial Robustness of Learning-based Conformal Novelty Detection, by Daofu Zhang and 4 other authors View PDF HTML (experimental) Abstract:This paper studies the adversarial robustness of conformal novelty detection. In particular, we focus on two powerful learning-based frameworks that come with finite-sample false discovery rate (FDR) control: one is AdaDetect (by Marandon et al., 2024) that is based on the positive-unlabeled classifier, and the other is a one-class classifier-based approach (by Bates et al., 2023). While they provide rigorous statistical guarantees under benign conditions, their behavior under adversarial perturbations remains underexplored. We first formulate an oracle attack setup, under the AdaDetect formulation, that quantifies the worst-case degradation of FDR, deriving an upper bound that characterizes the statistical cost of attacks. This idealized formulation directly motivates a practical and effective attack scheme that only requires query access to the output labels of both frameworks. Coupling these formulations with two popular and complementary black-box adversarial algorithms, we systematically evaluat...

Related Articles

Improving AI models’ ability to explain their predictions
Machine Learning

Improving AI models’ ability to explain their predictions

AI News - General · 9 min ·
Machine Learning

Auto agent - Self improving domain expertise agent

someone opensource an ai agent that autonomously upgraded itself to #1 across multiple domains in < 24 hours…. then open sourced the e...

Reddit - Artificial Intelligence · 1 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
Tuskegee University to host the 2026 Amazon Web Services–Machine Learning University Research & Teaching Symposium
Machine Learning

Tuskegee University to host the 2026 Amazon Web Services–Machine Learning University Research & Teaching Symposium

Tuskegee University will host the 2026 Amazon Web Services–Machine Learning University Spring AI/ML Teaching & Research Symposium on Febr...

AI News - General · 8 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime