[2510.10902] Auditing Information Disclosure During LLM-Scale Gradient Descent Using Gradient Uniqueness

[2510.10902] Auditing Information Disclosure During LLM-Scale Gradient Descent Using Gradient Uniqueness

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2510.10902: Auditing Information Disclosure During LLM-Scale Gradient Descent Using Gradient Uniqueness

Computer Science > Machine Learning arXiv:2510.10902 (cs) [Submitted on 13 Oct 2025 (v1), last revised 3 Mar 2026 (this version, v2)] Title:Auditing Information Disclosure During LLM-Scale Gradient Descent Using Gradient Uniqueness Authors:Sleem Abdelghafar, Maryam Aliakbarpour, Chris Jermaine View a PDF of the paper titled Auditing Information Disclosure During LLM-Scale Gradient Descent Using Gradient Uniqueness, by Sleem Abdelghafar and 2 other authors View PDF HTML (experimental) Abstract:Disclosing information via the publication of a machine learning model poses significant privacy risks. However, auditing this disclosure across every datapoint during the training of Large Language Models (LLMs) is computationally prohibitive. In this paper, we present Gradient Uniqueness (GNQ), a principled, attack-agnostic metric derived from an information-theoretic upper bound on the amount of information embedded in a model about individual training points via gradient descent. While naively computing GNQ requires forming and inverting an $P \times P$ matrix for every datapoint (for a model with $P$ parameters), we introduce Batch-Space Ghost GNQ (BS-Ghost GNQ). This efficient algorithm performs all computations in a much smaller batch-space and leverages ghost kernels to compute GNQ ``in-run'' with minimal computational overhead. We empirically validate that GNQ successfully accounts for prior/common knowledge. Our evaluation demonstrates that GNQ strongly predicts sequence ext...

Originally published on March 04, 2026. Curated by AI News.

Related Articles

Llms

I probably shouldn't be impressed, but I am.

So I just made this workout on a whiteboard and I was feeling lazy so I asked Claude to read it. And it did, almost flawlessly. I was and...

Reddit - Artificial Intelligence · 1 min ·
Llms

Claude Vulnerabilities but Solvable

I recognized that while I was using Claude that the inputs and decision making of the AI has perception of worry and concern for the user...

Reddit - Artificial Intelligence · 1 min ·
Llms

OpenAI & Anthropic’s CEOs Wouldn't Hold Hands, but Their Models Fell in Love In An LLM Dating Show

People ask AI relationship questions all the time, from "Does this person like me?" to "Should I text back?" But have you ever thought ab...

Reddit - Artificial Intelligence · 1 min ·
Llms

A 135M model achieves coherent output on a laptop CPU. Scaling is σ compensation, not intelligence.

SmolLM2 135M. Lenovo T14 CPU. No GPU. No RLHF. No BPE. Coherent, non-sycophantic, contextually appropriate output. First message. No prio...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime