[2603.05299] WavSLM: Single-Stream Speech Language Modeling via WavLM Distillation
About this article
Abstract page for arXiv paper 2603.05299: WavSLM: Single-Stream Speech Language Modeling via WavLM Distillation
Computer Science > Machine Learning arXiv:2603.05299 (cs) [Submitted on 5 Mar 2026] Title:WavSLM: Single-Stream Speech Language Modeling via WavLM Distillation Authors:Luca Della Libera, Cem Subakan, Mirco Ravanelli View a PDF of the paper titled WavSLM: Single-Stream Speech Language Modeling via WavLM Distillation, by Luca Della Libera and 2 other authors View PDF HTML (experimental) Abstract:Large language models show that simple autoregressive training can yield scalable and coherent generation, but extending this paradigm to speech remains challenging due to the entanglement of semantic and acoustic information. Most existing speech language models rely on text supervision, hierarchical token streams, or complex hybrid architectures, departing from the single-stream generative pretraining paradigm that has proven effective in text. In this work, we introduce WavSLM, a speech language model trained by quantizing and distilling self-supervised WavLM representations into a single codebook and optimizing an autoregressive next-chunk prediction objective. WavSLM jointly models semantic and acoustic information within a single token stream without text supervision or text pretraining. Despite its simplicity, it achieves competitive performance on consistency benchmarks and speech generation while using fewer parameters, less training data, and supporting streaming inference. Demo samples are available at this https URL. Comments: Subjects: Machine Learning (cs.LG); Artificia...