[2602.24181] A Mixed Diet Makes DINO An Omnivorous Vision Encoder
About this article
Abstract page for arXiv paper 2602.24181: A Mixed Diet Makes DINO An Omnivorous Vision Encoder
Computer Science > Computer Vision and Pattern Recognition arXiv:2602.24181 (cs) [Submitted on 27 Feb 2026] Title:A Mixed Diet Makes DINO An Omnivorous Vision Encoder Authors:Rishabh Kabra, Maks Ovsjanikov, Drew A. Hudson, Ye Xia, Skanda Koppula, Andre Araujo, Joao Carreira, Niloy J. Mitra View a PDF of the paper titled A Mixed Diet Makes DINO An Omnivorous Vision Encoder, by Rishabh Kabra and 7 other authors View PDF HTML (experimental) Abstract:Pre-trained vision encoders like DINOv2 have demonstrated exceptional performance on unimodal tasks. However, we observe that their feature representations are poorly aligned across different modalities. For instance, the feature embedding for an RGB image and its corresponding depth map of the same scene exhibit a cosine similarity that is nearly identical to that of two random, unrelated images. To address this, we propose the Omnivorous Vision Encoder, a novel framework that learns a modality-agnostic feature space. We train the encoder with a dual objective: first, to maximize the feature alignment between different modalities of the same scene; and second, a distillation objective that anchors the learned representations to the output of a fully frozen teacher such as DINOv2. The resulting student encoder becomes "omnivorous" by producing a consistent, powerful embedding for a given scene, regardless of the input modality (RGB, Depth, Segmentation, etc.). This approach enables robust cross-modal understanding while retaining ...