[2603.21612] Towards Multimodal Time Series Anomaly Detection with Semantic Alignment and Condensed Interaction
About this article
Abstract page for arXiv paper 2603.21612: Towards Multimodal Time Series Anomaly Detection with Semantic Alignment and Condensed Interaction
Computer Science > Machine Learning arXiv:2603.21612 (cs) [Submitted on 23 Mar 2026] Title:Towards Multimodal Time Series Anomaly Detection with Semantic Alignment and Condensed Interaction Authors:Shiyan Hu, Jianxin Jin, Yang Shu, Peng Chen, Bin Yang, Chenjuan Guo View a PDF of the paper titled Towards Multimodal Time Series Anomaly Detection with Semantic Alignment and Condensed Interaction, by Shiyan Hu and 5 other authors View PDF HTML (experimental) Abstract:Time series anomaly detection plays a critical role in many dynamic systems. Despite its importance, previous approaches have primarily relied on unimodal numerical data, overlooking the importance of complementary information from other modalities. In this paper, we propose a novel multimodal time series anomaly detection model (MindTS) that focuses on addressing two key challenges: (1) how to achieve semantically consistent alignment across heterogeneous multimodal data, and (2) how to filter out redundant modality information to enhance cross-modal interaction effectively. To address the first challenge, we propose Fine-grained Time-text Semantic Alignment. It integrates exogenous and endogenous text information through cross-view text fusion and a multimodal alignment mechanism, achieving semantically consistent alignment between time and text modalities. For the second challenge, we introduce Content Condenser Reconstruction, which filters redundant information within the aligned text modality and performs cr...