[2604.02339] SIEVE: Sample-Efficient Parametric Learning from Natural Language
About this article
Abstract page for arXiv paper 2604.02339: SIEVE: Sample-Efficient Parametric Learning from Natural Language
Computer Science > Machine Learning arXiv:2604.02339 (cs) [Submitted on 2 Feb 2026] Title:SIEVE: Sample-Efficient Parametric Learning from Natural Language Authors:Parth Asawa, Alexandros G. Dimakis, Matei Zaharia View a PDF of the paper titled SIEVE: Sample-Efficient Parametric Learning from Natural Language, by Parth Asawa and 2 other authors View PDF HTML (experimental) Abstract:Natural language context-such as instructions, knowledge, or feedback-contains rich signal for adapting language models. While in-context learning provides adaptation via the prompt, parametric learning persists into model weights and can improve performance further, though is data hungry and heavily relies on either high-quality traces or automated verifiers. We propose SIEVE, a method for sample-efficient parametric learning from natural language context that requires as few as three query examples. SIEVE uses a novel synthetic data generation pipeline, SIEVE-GEN, that leverages the insight that context is decomposable. Decomposing context allows us to generate higher quality rollouts by pairing synthetic queries with only the applicable context rather than the entirety, then using context distillation to internalize context into the model. We evaluate in reasoning settings where context is necessary, including custom domains and the RuleArena and Machine Translation from One Book tasks. Our results show that SIEVE outperforms prior context distillation methods using just three query examples,...