[2506.01929] Image Generation from Contextually-Contradictory Prompts
About this article
Abstract page for arXiv paper 2506.01929: Image Generation from Contextually-Contradictory Prompts
Computer Science > Graphics arXiv:2506.01929 (cs) [Submitted on 2 Jun 2025 (v1), last revised 23 Mar 2026 (this version, v2)] Title:Image Generation from Contextually-Contradictory Prompts Authors:Saar Huberman, Or Patashnik, Omer Dahary, Ron Mokady, Daniel Cohen-Or View a PDF of the paper titled Image Generation from Contextually-Contradictory Prompts, by Saar Huberman and 4 other authors View PDF HTML (experimental) Abstract:Text-to-image diffusion models excel at generating high-quality, diverse images from natural language prompts. However, they often fail to produce semantically accurate results when the prompt contains concept combinations that contradict their learned priors. We define this failure mode as contextual contradiction, where one concept implicitly negates another due to entangled associations learned during training. To address this, we propose a stage-aware prompt decomposition framework that guides the denoising process using a sequence of proxy prompts. Each proxy prompt is constructed to match the semantic content expected to emerge at a specific stage of denoising, while ensuring contextual coherence. To construct these proxy prompts, we leverage a large language model (LLM) to analyze the target prompt, identify contradictions, and generate alternative expressions that preserve the original intent while resolving contextual conflicts. By aligning prompt information with the denoising progression, our method enables fine-grained semantic control an...