[2604.04681] Batch Loss Score for Dynamic Data Pruning
About this article
Abstract page for arXiv paper 2604.04681: Batch Loss Score for Dynamic Data Pruning
Computer Science > Machine Learning arXiv:2604.04681 (cs) [Submitted on 6 Apr 2026] Title:Batch Loss Score for Dynamic Data Pruning Authors:Qing Zhou, Bingxuan Zhao, Tao Yang, Hongyuan Zhang, Junyu Gao, Qi Wang View a PDF of the paper titled Batch Loss Score for Dynamic Data Pruning, by Qing Zhou and 5 other authors View PDF HTML (experimental) Abstract:Dynamic data pruning accelerates deep learning by selectively omitting less informative samples during training. While per-sample loss is a common importance metric, obtaining it can be challenging or infeasible for complex models or loss functions, often requiring significant implementation effort. This work proposes the Batch Loss Score (BLS), a computationally efficient alternative using an Exponential Moving Average (EMA) of readily available batch losses to assign scores to individual samples. We frame the batch loss, from the perspective of a single sample, as a noisy measurement of its scaled individual loss, with noise originating from stochastic batch composition. It is formally shown that the EMA mechanism functions as a first-order low-pass filter, attenuating high-frequency batch composition noise. This yields a score approximating the smoothed and persistent contribution of the individual sample to the loss, providing a theoretical grounding for BLS as a proxy for sample importance. BLS demonstrates remarkable code integration simplicity (\textbf{three-line injection}) and readily adapts existing per-sample los...