[2604.02339] SIEVE: Sample-Efficient Parametric Learning from Natural Language

[2604.02339] SIEVE: Sample-Efficient Parametric Learning from Natural Language

arXiv - Machine Learning 3 min read

About this article

Abstract page for arXiv paper 2604.02339: SIEVE: Sample-Efficient Parametric Learning from Natural Language

Computer Science > Machine Learning arXiv:2604.02339 (cs) [Submitted on 2 Feb 2026] Title:SIEVE: Sample-Efficient Parametric Learning from Natural Language Authors:Parth Asawa, Alexandros G. Dimakis, Matei Zaharia View a PDF of the paper titled SIEVE: Sample-Efficient Parametric Learning from Natural Language, by Parth Asawa and 2 other authors View PDF HTML (experimental) Abstract:Natural language context-such as instructions, knowledge, or feedback-contains rich signal for adapting language models. While in-context learning provides adaptation via the prompt, parametric learning persists into model weights and can improve performance further, though is data hungry and heavily relies on either high-quality traces or automated verifiers. We propose SIEVE, a method for sample-efficient parametric learning from natural language context that requires as few as three query examples. SIEVE uses a novel synthetic data generation pipeline, SIEVE-GEN, that leverages the insight that context is decomposable. Decomposing context allows us to generate higher quality rollouts by pairing synthetic queries with only the applicable context rather than the entirety, then using context distillation to internalize context into the model. We evaluate in reasoning settings where context is necessary, including custom domains and the RuleArena and Machine Translation from One Book tasks. Our results show that SIEVE outperforms prior context distillation methods using just three query examples,...

Originally published on April 06, 2026. Curated by AI News.

Related Articles

Llms

We benchmarked TranslateGemma against 5 other LLMs on subtitle translation across 6 languages. At first glance the numbers told a clean story, but then human QA added a chapter. [D]

We evaluated six models on English subtitle translation into Spanish, Japanese, Korean, Thai, Chinese Simplified, and Chinese Traditional...

Reddit - Machine Learning · 1 min ·
Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert
Llms

Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert

A new AI model could automate the process of searching for cybersecurity bugs and flaws – for better or worse.

AI Tools & Products · 5 min ·
Gemini could take a 'proactive' approach with leaked 'Your Day' feature
Llms

Gemini could take a 'proactive' approach with leaked 'Your Day' feature

This feature could leverage your apps in a way that might feel familiar.

AI Tools & Products · 5 min ·
I ditched my paper planner for Gemini Live — and it solved the one professional problem I couldn't fix
Llms

I ditched my paper planner for Gemini Live — and it solved the one professional problem I couldn't fix

Can Gemini Live replace a physical planner? Tom's Guide AI Editor Amanda Caswell ditched her notebook for Google’s voice AI. Here’s how i...

AI Tools & Products · 8 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime