[2602.15457] Benchmarking IoT Time-Series AD with Event-Level Augmentations

[2602.15457] Benchmarking IoT Time-Series AD with Event-Level Augmentations

arXiv - Machine Learning 4 min read Article

Summary

This paper presents a novel evaluation protocol for anomaly detection in IoT time-series data, emphasizing event-level assessments over traditional point-level metrics, and evaluates various models under realistic conditions.

Why It Matters

As IoT systems become increasingly integral to safety-critical applications, effective anomaly detection is crucial. This study addresses the limitations of existing methodologies by proposing a more realistic evaluation framework that enhances model selection and performance in real-world scenarios.

Key Takeaways

  • Introduces an evaluation protocol focusing on event-level augmentations for anomaly detection in IoT.
  • Evaluates 14 models across multiple datasets, revealing no single best-performing model.
  • Highlights the impact of realistic perturbations on model performance, guiding future research and application.

Computer Science > Machine Learning arXiv:2602.15457 (cs) [Submitted on 17 Feb 2026] Title:Benchmarking IoT Time-Series AD with Event-Level Augmentations Authors:Dmitry Zhevnenko, Ilya Makarov, Aleksandr Kovalenko, Fedor Meshchaninov, Anton Kozhukhov, Vladislav Travnikov, Makar Ippolitov, Kirill Yashunin, Iurii Katser View a PDF of the paper titled Benchmarking IoT Time-Series AD with Event-Level Augmentations, by Dmitry Zhevnenko and 8 other authors View PDF HTML (experimental) Abstract:Anomaly detection (AD) for safety-critical IoT time series should be judged at the event level: reliability and earliness under realistic perturbations. Yet many studies still emphasize point-level results on curated base datasets, limiting value for model selection in practice. We introduce an evaluation protocol with unified event-level augmentations that simulate real-world issues: calibrated sensor dropout, linear and log drift, additive noise, and window shifts. We also perform sensor-level probing via mask-as-missing zeroing with per-channel influence estimation to support root-cause analysis. We evaluate 14 representative models on five public anomaly datasets (SWaT, WADI, SMD, SKAB, TEP) and two industrial datasets (steam turbine, nuclear turbogenerator) using unified splits and event aggregation. There is no universal winner: graph-structured models transfer best under dropout and long events (e.g., on SWaT under additive noise F1 drops 0.804->0.677 for a graph autoencoder, 0.759-...

Related Articles

UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
Machine Learning

[D] Budget Machine Learning Hardware

Looking to get into machine learning and found this video on a piece of hardware for less than £500. Is it really possible to teach auton...

Reddit - Machine Learning · 1 min ·
Machine Learning

Your prompts aren’t the problem — something else is

I keep seeing people focus heavily on prompt optimization. But in practice, a lot of failures I’ve observed don’t come from the prompt it...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[R], 31 MILLIONS High frequency data, Light GBM worked perfectly

We just published a paper on predicting adverse selection in high-frequency crypto markets using LightGBM, and I wanted to share it here ...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime