[2604.06266] Attribution-Driven Explainable Intrusion Detection with Encoder-Based Large Language Models
About this article
Abstract page for arXiv paper 2604.06266: Attribution-Driven Explainable Intrusion Detection with Encoder-Based Large Language Models
Computer Science > Cryptography and Security arXiv:2604.06266 (cs) [Submitted on 7 Apr 2026] Title:Attribution-Driven Explainable Intrusion Detection with Encoder-Based Large Language Models Authors:Umesh Biswas, Shafqat Hasan, Syed Mohammed Farhan, Nisha Pillai, Charan Gudla View a PDF of the paper titled Attribution-Driven Explainable Intrusion Detection with Encoder-Based Large Language Models, by Umesh Biswas and 4 other authors View PDF HTML (experimental) Abstract:Software-Defined Networking (SDN) improves network flexibility but also increases the need for reliable and interpretable intrusion detection. Large Language Models (LLMs) have recently been explored for cybersecurity tasks due to their strong representation learning capabilities; however, their lack of transparency limits their practical adoption in security-critical environments. Understanding how LLMs make decisions is therefore essential. This paper presents an attribution-driven analysis of encoder-based LLMs for network intrusion detection using flow-level traffic features. Attribution analysis demonstrates that model decisions are driven by meaningful traffic behavior patterns, improving transparency and trust in transformer-based SDN intrusion detection. These patterns align with established intrusion detection principles, indicating that LLMs learn attack behavior from traffic dynamics. This work demonstrates the value of attribution methods for validating and trusting LLM-based security analysis. ...