[2603.05432] Ensembling Language Models with Sequential Monte Carlo
About this article
Abstract page for arXiv paper 2603.05432: Ensembling Language Models with Sequential Monte Carlo
Computer Science > Computation and Language arXiv:2603.05432 (cs) [Submitted on 5 Mar 2026] Title:Ensembling Language Models with Sequential Monte Carlo Authors:Robin Shing Moon Chan, Tianyu Liu, Samuel Kiegeland, Clemente Pasti, Jacob Hoover Vigly, Timothy J. O'Donnell, Ryan Cotterell, Tim Vieira View a PDF of the paper titled Ensembling Language Models with Sequential Monte Carlo, by Robin Shing Moon Chan and 7 other authors View PDF Abstract:Practitioners have access to an abundance of language models and prompting strategies for solving many language modeling tasks; yet prior work shows that modeling performance is highly sensitive to both choices. Classical machine learning ensembling techniques offer a principled approach: aggregate predictions from multiple sources to achieve better performance than any single one. However, applying ensembling to language models during decoding is challenging: naively aggregating next-token probabilities yields samples from a locally normalized, biased approximation of the generally intractable ensemble distribution over strings. In this work, we introduce a unified framework for composing $K$ language models into $f$-ensemble distributions for a wide range of functions $f\colon\mathbb{R}_{\geq 0}^{K}\to\mathbb{R}_{\geq 0}$. To sample from these distributions, we propose a byte-level sequential Monte Carlo (SMC) algorithm that operates in a shared character space, enabling ensembles of models with mismatching vocabularies and consis...