[2603.01683] Surgical Post-Training: Cutting Errors, Keeping Knowledge

[2603.01683] Surgical Post-Training: Cutting Errors, Keeping Knowledge

arXiv - AI 3 min read

About this article

Abstract page for arXiv paper 2603.01683: Surgical Post-Training: Cutting Errors, Keeping Knowledge

Computer Science > Computation and Language arXiv:2603.01683 (cs) [Submitted on 2 Mar 2026] Title:Surgical Post-Training: Cutting Errors, Keeping Knowledge Authors:Wenye Lin, Kai Han View a PDF of the paper titled Surgical Post-Training: Cutting Errors, Keeping Knowledge, by Wenye Lin and 1 other authors View PDF HTML (experimental) Abstract:Enhancing the reasoning capabilities of Large Language Models (LLMs) via post-training is often constrained by the trade-off between efficiency and catastrophic forgetting. While prior research emphasizes the role of on-policy data in mitigating forgetting, we uncover--and validate both theoretically and empirically--an overlooked yet critical mechanism: the implicit regularization inherent in Direct Preference Optimization's (DPO) reward estimate. This motivates our Surgical Post-Training (SPoT), a new paradigm designed to optimize reasoning efficiently while preserving learned prior knowledge. SPoT consists of: (1) a data rectification pipeline that employs an Oracle to surgically correct erroneous steps via minimal edits, generating data proximal to the model's distribution; and (2) a reward-based binary cross-entropy objective. Unlike the relative ranking in DPO, this objective treats reasoning correctness as a binary classification problem, enforcing decoupled supervision signals. Empirically, with only 4k rectified math data pairs, SPoT improves Qwen3-8B's accuracy by 6.2% on average across in-domain and OOD tasks, requiring mere...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

Llms

8 free AI courses from Anthropic’s Claude platform with certificates

AI News - General ·
Llms

How is mythos mythos ? [D]

Hello, I’ve been seeing discussions about “Mythos AI” showing behaviors that seem far beyond simple text prediction—like accessing inform...

Reddit - Machine Learning · 1 min ·
Llms

Claude developer hosts Christian leaders for AI summit

AI Tools & Products ·
CoreWeave stock pops 11% on deal to power Anthropic's Claude
Llms

CoreWeave stock pops 11% on deal to power Anthropic's Claude

AI Tools & Products · 3 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime