[2602.22258] Poisoned Acoustics

[2602.22258] Poisoned Acoustics

arXiv - AI 3 min read Article

Summary

The paper 'Poisoned Acoustics' explores training-data poisoning attacks on deep neural networks, demonstrating significant vulnerabilities in acoustic vehicle classification systems.

Why It Matters

This research highlights critical security concerns in machine learning, particularly regarding the integrity of training data. As AI systems become more prevalent, understanding and mitigating these vulnerabilities is essential for ensuring reliable performance in real-world applications.

Key Takeaways

  • Training-data poisoning can lead to undetectable failures in neural networks.
  • A compact CNN achieved a 95.7% attack success rate with minimal data corruption.
  • Aggregate accuracy monitoring is insufficient for detecting such attacks.
  • Proposed defense mechanisms include cryptographic verification for data provenance.
  • Understanding the attack surface in ML training pipelines is crucial for developing robust defenses.

Computer Science > Cryptography and Security arXiv:2602.22258 (cs) [Submitted on 25 Feb 2026] Title:Poisoned Acoustics Authors:Harrison Dahme View a PDF of the paper titled Poisoned Acoustics, by Harrison Dahme View PDF HTML (experimental) Abstract:Training-data poisoning attacks can induce targeted, undetectable failure in deep neural networks by corrupting a vanishingly small fraction of training labels. We demonstrate this on acoustic vehicle classification using the MELAUDIS urban intersection dataset (approx. 9,600 audio clips, 6 classes): a compact 2-D convolutional neural network (CNN) trained on log-mel spectrograms achieves 95.7% Attack Success Rate (ASR) -- the fraction of target-class test samples misclassified under the attack -- on a Truck-to-Car label-flipping attack at just p=0.5% corruption (48 records), with zero detectable change in aggregate accuracy (87.6% baseline; 95% CI: 88-100%, n=3 seeds). We prove this stealth is structural: the maximum accuracy drop from a complete targeted attack is bounded above by the minority class fraction (beta). For real-world class imbalances (Truck approx. 3%), this bound falls below training-run noise, making aggregate accuracy monitoring provably insufficient regardless of architecture or attack method. A companion backdoor trigger attack reveals a novel trigger-dominance collapse: when the target class is a dataset minority, the spectrogram patch trigger becomes functionally redundant--clean ASR equals triggered ASR, ...

Related Articles

Llms

[R] Depth-first pruning transfers: GPT-2 → TinyLlama with stable gains and minimal loss

TL;DR: Removing the right layers (instead of shrinking all layers) makes transformer models ~8–12% smaller with only ~6–8% quality loss, ...

Reddit - Machine Learning · 1 min ·
Llms

Built a training stability monitor that detects instability before your loss curve shows anything — open sourced the core today

Been working on a weight divergence trajectory curvature approach to detecting neural network training instability. Treats weight updates...

Reddit - Artificial Intelligence · 1 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
Improving AI models’ ability to explain their predictions
Machine Learning

Improving AI models’ ability to explain their predictions

AI News - General · 9 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime