[2306.14685] DiffSketcher: Text Guided Vector Sketch Synthesis through Latent Diffusion Models
About this article
Abstract page for arXiv paper 2306.14685: DiffSketcher: Text Guided Vector Sketch Synthesis through Latent Diffusion Models
Computer Science > Computer Vision and Pattern Recognition arXiv:2306.14685 (cs) [Submitted on 26 Jun 2023 (v1), last revised 8 Apr 2026 (this version, v5)] Title:DiffSketcher: Text Guided Vector Sketch Synthesis through Latent Diffusion Models Authors:Ximing Xing, Chuang Wang, Haitao Zhou, Jing Zhang, Qian Yu, Dong Xu View a PDF of the paper titled DiffSketcher: Text Guided Vector Sketch Synthesis through Latent Diffusion Models, by Ximing Xing and 5 other authors View PDF HTML (experimental) Abstract:We demonstrate that pre-trained text-to-image diffusion models, despite being trained on raster images, possess a remarkable capacity to guide vector sketch synthesis. In this paper, we introduce DiffSketcher, a novel algorithm for generating vectorized free-hand sketches directly from natural language prompts. Our method optimizes a set of Bézier curves via an extended Score Distillation Sampling (SDS) loss, successfully bridging a raster-level diffusion prior with a parametric vector generator. To further accelerate the generation process, we propose a stroke initialization strategy driven by the diffusion model's intrinsic attention maps. Results show that DiffSketcher produces sketches across varying levels of abstraction while maintaining the structural integrity and essential visual details of the subject. Experiments confirm that our approach yields superior perceptual quality and controllability over existing methods. The code and demo are available at this https URL...