[2507.17343] Principled Multimodal Representation Learning
About this article
Abstract page for arXiv paper 2507.17343: Principled Multimodal Representation Learning
Computer Science > Computer Vision and Pattern Recognition arXiv:2507.17343 (cs) [Submitted on 23 Jul 2025 (v1), last revised 20 Mar 2026 (this version, v3)] Title:Principled Multimodal Representation Learning Authors:Xiaohao Liu, Xiaobo Xia, See-Kiong Ng, Tat-Seng Chua View a PDF of the paper titled Principled Multimodal Representation Learning, by Xiaohao Liu and 3 other authors View PDF HTML (experimental) Abstract:Multimodal representation learning seeks to create a unified representation space by integrating diverse data modalities to improve multimodal understanding. Traditional methods often depend on pairwise contrastive learning, which relies on a predefined anchor modality, restricting alignment across all modalities. Recent advances have investigated the simultaneous alignment of multiple modalities, yet several challenges remain, such as limitations imposed by fixed anchor points and instability arising from optimizing the product of singular values. To address the challenges, in this paper, we propose Principled Multimodal Representation Learning (PMRL), a novel framework that achieves simultaneous alignment of multiple modalities without anchor dependency in a more stable manner. Specifically, grounded in the theoretical insight that full alignment corresponds to a rank-1 Gram matrix, PMRL optimizes the dominant singular value of the representation matrix to align modalities along a shared leading direction. We propose a softmax-based loss function that treat...