[2603.28644] Constructing Composite Features for Interpretable Music-Tagging
About this article
Abstract page for arXiv paper 2603.28644: Constructing Composite Features for Interpretable Music-Tagging
Computer Science > Sound arXiv:2603.28644 (cs) [Submitted on 30 Mar 2026] Title:Constructing Composite Features for Interpretable Music-Tagging Authors:Chenhao Xue, Weitao Hu, Joyraj Chakraborty, Zhijin Guo, Kang Li, Tianyu Shi, Martin Reed, Nikolaos Thomos View a PDF of the paper titled Constructing Composite Features for Interpretable Music-Tagging, by Chenhao Xue and 7 other authors View PDF HTML (experimental) Abstract:Combining multiple audio features can improve the performance of music tagging, but common deep learning-based feature fusion methods often lack interpretability. To address this problem, we propose a Genetic Programming (GP) pipeline that automatically evolves composite features by mathematically combining base music features, thereby capturing synergistic interactions while preserving interpretability. This approach provides representational benefits similar to deep feature fusion without sacrificing interpretability. Experiments on the MTG-Jamendo and GTZAN datasets demonstrate consistent improvements compared to state-of-the-art systems across base feature sets at different abstraction levels. It should be noted that most of the performance gains are noticed within the first few hundred GP evaluations, indicating that effective feature combinations can be identified under modest search budgets. The top evolved expressions include linear, nonlinear, and conditional forms, with various low-complexity solutions at top performance aligned with parsimony ...