[2512.10685] Sharp Monocular View Synthesis in Less Than a Second
About this article
Abstract page for arXiv paper 2512.10685: Sharp Monocular View Synthesis in Less Than a Second
Computer Science > Computer Vision and Pattern Recognition arXiv:2512.10685 (cs) [Submitted on 11 Dec 2025 (v1), last revised 27 Feb 2026 (this version, v2)] Title:Sharp Monocular View Synthesis in Less Than a Second Authors:Lars Mescheder, Wei Dong, Shiwei Li, Xuyang Bai, Marcel Santos, Peiyun Hu, Bruno Lecouat, Mingmin Zhen, Amaël Delaunoy, Tian Fang, Yanghai Tsin, Stephan R. Richter, Vladlen Koltun View a PDF of the paper titled Sharp Monocular View Synthesis in Less Than a Second, by Lars Mescheder and 12 other authors View PDF HTML (experimental) Abstract:We present SHARP, an approach to photorealistic view synthesis from a single image. Given a single photograph, SHARP regresses the parameters of a 3D Gaussian representation of the depicted scene. This is done in less than a second on a standard GPU via a single feedforward pass through a neural network. The 3D Gaussian representation produced by SHARP can then be rendered in real time, yielding high-resolution photorealistic images for nearby views. The representation is metric, with absolute scale, supporting metric camera movements. Experimental results demonstrate that SHARP delivers robust zero-shot generalization across datasets. It sets a new state of the art on multiple datasets, reducing LPIPS by 25-34% and DISTS by 21-43% versus the best prior model, while lowering the synthesis time by three orders of magnitude. Code and weights are provided at this https URL Comments: Subjects: Computer Vision and Pattern...