[2603.04405] Lost in Translation: How Language Re-Aligns Vision for Cross-Species Pathology
About this article
Abstract page for arXiv paper 2603.04405: Lost in Translation: How Language Re-Aligns Vision for Cross-Species Pathology
Computer Science > Computer Vision and Pattern Recognition arXiv:2603.04405 (cs) [Submitted on 24 Jan 2026] Title:Lost in Translation: How Language Re-Aligns Vision for Cross-Species Pathology Authors:Ekansh Arora View a PDF of the paper titled Lost in Translation: How Language Re-Aligns Vision for Cross-Species Pathology, by Ekansh Arora View PDF HTML (experimental) Abstract:Foundation models are increasingly applied to computational pathology, yet their behavior under cross-cancer and cross-species transfer remains unspecified. This study investigated how fine-tuning CPath-CLIP affects cancer detection under same-cancer, cross-cancer, and cross-species conditions using whole-slide image patches from canine and human histopathology. Performance was measured using area under the receiver operating characteristic curve (AUC). Few-shot fine-tuning improved same-cancer (64.9% to 72.6% AUC) and cross-cancer performance (56.84% to 66.31% AUC). Cross-species evaluation revealed that while tissue matching enables meaningful transfer, performance remains below state-of-the-art benchmarks (H-optimus-0: 84.97% AUC), indicating that standard vision-language alignment is suboptimal for cross-species generalization. Embedding space analysis revealed extremely high cosine similarity (greater than 0.99) between tumor and normal prototypes. Grad-CAM shows prototype-based models remain domain-locked, while language-guided models attend to conserved tumor morphology. To address this, we int...