[2601.12104] Powerful Training-Free Membership Inference Against Autoregressive Language Models

[2601.12104] Powerful Training-Free Membership Inference Against Autoregressive Language Models

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2601.12104: Powerful Training-Free Membership Inference Against Autoregressive Language Models

Computer Science > Computation and Language arXiv:2601.12104 (cs) [Submitted on 17 Jan 2026 (v1), last revised 13 Apr 2026 (this version, v2)] Title:Powerful Training-Free Membership Inference Against Autoregressive Language Models Authors:David Ilić, David Stanojević, Kostadin Cvejoski View a PDF of the paper titled Powerful Training-Free Membership Inference Against Autoregressive Language Models, by David Ili\'c and 2 other authors View PDF HTML (experimental) Abstract:Fine-tuned language models pose significant privacy risks, as they may memorize and expose sensitive information from their training data. Membership inference attacks (MIAs) provide a principled framework for auditing these risks, yet existing methods achieve limited detection rates, particularly at the low false-positive thresholds required for practical privacy auditing. We present EZ-MIA, a membership inference attack that exploits a key observation: memorization manifests most strongly at error positions, specifically tokens where the model predicts incorrectly yet still shows elevated probability for training examples. We introduce the Error Zone (EZ) score, which measures the directional imbalance of probability shifts at error positions relative to a pretrained reference model. This principled statistic requires only two forward passes per query and no model training of any kind. On WikiText with GPT-2, EZ-MIA achieves 3.8x higher detection than the previous state-of-the-art under identical condit...

Originally published on April 14, 2026. Curated by AI News.

Related Articles

Llms

20M+ Indian legal documents with citation graphs and vector embeddings – potential uses for legal NLP? [D]

been working on structuring India's legal corpus for the past 2 years and wanted to share what I've built and hear from people working on...

Reddit - Machine Learning · 1 min ·
Llms

openclaw ai agent vs just using chatgpt

I've been using AI tools pretty heavily for the past couple of years. ChatGPT, Claude, Perplexity, a few others. I thought I had a good m...

Reddit - Artificial Intelligence · 1 min ·
Llms

We benchmarked TranslateGemma against 5 other LLMs on subtitle translation across 6 languages. At first glance the numbers told a clean story, but then human QA added a chapter. [D]

We evaluated six models on English subtitle translation into Spanish, Japanese, Korean, Thai, Chinese Simplified, and Chinese Traditional...

Reddit - Machine Learning · 1 min ·
Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert
Llms

Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert

A new AI model could automate the process of searching for cybersecurity bugs and flaws – for better or worse.

AI Tools & Products · 5 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime