[2603.00140] Steering Away from Memorization: Reachability-Constrained Reinforcement Learning for Text-to-Image Diffusion

[2603.00140] Steering Away from Memorization: Reachability-Constrained Reinforcement Learning for Text-to-Image Diffusion

arXiv - Machine Learning 3 min read

About this article

Abstract page for arXiv paper 2603.00140: Steering Away from Memorization: Reachability-Constrained Reinforcement Learning for Text-to-Image Diffusion

Computer Science > Computer Vision and Pattern Recognition arXiv:2603.00140 (cs) [Submitted on 24 Feb 2026] Title:Steering Away from Memorization: Reachability-Constrained Reinforcement Learning for Text-to-Image Diffusion Authors:Sathwik Karnik, Juyeop Kim, Sanmi Koyejo, Jong-Seok Lee, Somil Bansal View a PDF of the paper titled Steering Away from Memorization: Reachability-Constrained Reinforcement Learning for Text-to-Image Diffusion, by Sathwik Karnik and 4 other authors View PDF HTML (experimental) Abstract:Text-to-image diffusion models often memorize training data, revealing a fundamental failure to generalize beyond the training set. Current mitigation strategies typically sacrifice image quality or prompt alignment to reduce memorization. To address this, we propose Reachability-Aware Diffusion Steering (RADS), an inference-time framework that prevents memorization while preserving generation fidelity. RADS models the diffusion denoising process as a dynamical system and applies concepts from reachability analysis to approximate the "backward reachable tube"--the set of intermediate states that inevitably evolve into memorized samples. We then formulate mitigation as a constrained reinforcement learning (RL) problem, where a policy learns to steer the trajectory away from memorization via minimal perturbations in the caption embedding space. Empirical evaluations show that RADS achieves a superior Pareto frontier between generation diversity (SSCD), quality (FID),...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

Machine Learning

[P] Unix philosophy for ML pipelines: modular, swappable stages with typed contracts

We built an open-source prototype that applies Unix philosophy to retrieval pipelines. Each stage (PII redaction, chunking, dedup, embedd...

Reddit - Machine Learning · 1 min ·
Machine Learning

Making an AI native sovereign computational stack

I’ve been working on a personal project that ended up becoming a kind of full computing stack: identity / trust protocol decentralized ch...

Reddit - Artificial Intelligence · 1 min ·
Llms

An attack class that passes every current LLM filter - no payload, no injection signature, no log trace

https://shapingrooms.com/research I published a paper today on something I've been calling postural manipulation. The short version: ordi...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

What tools are sr MLEs using? (clawdbot, openspec, wispr) [D]

I'm already blasting cursor, but I want to level up my output. I heard that these kind of AI tools and workflows are being asked in SF. W...

Reddit - Machine Learning · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime