[2512.00234] OmniFusion: Simultaneous Multilingual Multimodal Translations via Modular Fusion
About this article
Abstract page for arXiv paper 2512.00234: OmniFusion: Simultaneous Multilingual Multimodal Translations via Modular Fusion
Computer Science > Computation and Language arXiv:2512.00234 (cs) [Submitted on 28 Nov 2025 (v1), last revised 1 Apr 2026 (this version, v2)] Title:OmniFusion: Simultaneous Multilingual Multimodal Translations via Modular Fusion Authors:Sai Koneru, Matthias Huck, Jan Niehues View a PDF of the paper titled OmniFusion: Simultaneous Multilingual Multimodal Translations via Modular Fusion, by Sai Koneru and 2 other authors View PDF HTML (experimental) Abstract:There has been significant progress in open-source text-only translation large language models (LLMs) with better language coverage and quality. However, these models can be only used in cascaded pipelines for speech translation (ST), performing automatic speech recognition first followed by translation. This introduces additional latency, which is particularly critical in simultaneous ST (SimulST), and prevents the model from exploiting multimodal context, such as images, which can aid disambiguation. Pretrained multimodal foundation models (MMFMs) already possess strong perception and reasoning capabilities across multiple modalities, but generally lack the multilingual coverage and specialized translation performance of dedicated translation LLMs. To build an effective multimodal translation system, we propose an end-to-end approach that fuses MMFMs with translation LLMs. We introduce a novel fusion strategy that connects hidden states from multiple layers of a pretrained MMFM to a translation LLM, enabling joint end-...