[2510.10066] OBsmith: LLM-Powered JavaScript Obfuscator Testing

[2510.10066] OBsmith: LLM-Powered JavaScript Obfuscator Testing

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2510.10066: OBsmith: LLM-Powered JavaScript Obfuscator Testing

Computer Science > Software Engineering arXiv:2510.10066 (cs) [Submitted on 11 Oct 2025 (v1), last revised 27 Feb 2026 (this version, v2)] Title:OBsmith: LLM-Powered JavaScript Obfuscator Testing Authors:Shan Jiang, Chenguang Zhu, Sarfraz Khurshid View a PDF of the paper titled OBsmith: LLM-Powered JavaScript Obfuscator Testing, by Shan Jiang and 2 other authors View PDF HTML (experimental) Abstract:JavaScript obfuscators are widely deployed to protect intellectual property and resist reverse engineering, yet their correctness has been largely overlooked compared to performance and resilience. Existing evaluations typically measure resistance to deobfuscation, leaving the critical question of whether obfuscators preserve program semantics unanswered. Incorrect transformations can silently alter functionality, compromise reliability, and erode security-undermining the very purpose of obfuscation. To address this gap, we present OBsmith, a novel framework to systematically test JavaScript obfuscators using large language models (LLMs). OBsmith leverages LLMs to generate program sketches abstract templates capturing diverse language constructs, idioms, and corner cases-which are instantiated into executable programs and subjected to obfuscation under different configurations. Besides LLM-powered sketching, OBsmith also employs a second source: automatic extraction of sketches from real programs. This extraction path enables more focused testing of project specific features an...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

Llms

I Accidentally Discovered a Security Vulnerability in AI Education — Then Submitted It To a $200K Competition

Last night I was testing Maestro University, the first fully AI-taught university. I walked into their enrollment chatbot and asked it to...

Reddit - Artificial Intelligence · 1 min ·
Llms

Is anyone else concerned with this blatant potential of security / privacy breach?

Recently, when sending a very sensitive email to my brother including my mother’s health information, I wondered what happens if a recipi...

Reddit - Artificial Intelligence · 1 min ·
Llms

An attack class that passes every current LLM filter - no payload, no injection signature, no log trace

https://shapingrooms.com/research I published a paper today on something I've been calling postural manipulation. The short version: ordi...

Reddit - Artificial Intelligence · 1 min ·
Llms

[R] An attack class that passes every current LLM filter - no payload, no injection signature, no log trace

https://shapingrooms.com/research I've been documenting what I'm calling postural manipulation: a specific class of language that install...

Reddit - Machine Learning · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime