[2602.12681] Fool Me If You Can: On the Robustness of Binary Code Similarity Detection Models against Semantics-preserving Transformations
Summary
This paper evaluates the robustness of binary code similarity detection models against semantics-preserving transformations, introducing asmFooler to assess model resilience.
Why It Matters
As cybersecurity threats evolve, understanding the limitations of binary code analysis models is crucial. This research highlights vulnerabilities in machine learning approaches, which can inform future developments in secure coding practices and model design.
Key Takeaways
- Model robustness is influenced by the processing pipeline and feature selection.
- Adversarial transformations can effectively mislead models with minimal changes.
- A diverse dataset of binary variants enhances the evaluation of model resilience.
Computer Science > Cryptography and Security arXiv:2602.12681 (cs) [Submitted on 13 Feb 2026] Title:Fool Me If You Can: On the Robustness of Binary Code Similarity Detection Models against Semantics-preserving Transformations Authors:Jiyong Uhm, Minseok Kim, Michalis Polychronakis, Hyungjoon Koo View a PDF of the paper titled Fool Me If You Can: On the Robustness of Binary Code Similarity Detection Models against Semantics-preserving Transformations, by Jiyong Uhm and 3 other authors View PDF HTML (experimental) Abstract:Binary code analysis plays an essential role in cybersecurity, facilitating reverse engineering to reveal the inner workings of programs in the absence of source code. Traditional approaches, such as static and dynamic analysis, extract valuable insights from stripped binaries, but often demand substantial expertise and manual effort. Recent advances in deep learning have opened promising opportunities to enhance binary analysis by capturing latent features and disclosing underlying code semantics. Despite the growing number of binary analysis models based on machine learning, their robustness to adversarial code transformations at the binary level remains underexplored. We evaluate the robustness of deep learning models for the task of binary code similarity detection (BCSD) under semantics-preserving transformations. The unique nature of machine instructions presents distinct challenges compared to the typical input perturbations found in other domains. ...