[2603.28858] OptiMer: Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training

[2603.28858] OptiMer: Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2603.28858: OptiMer: Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training

Computer Science > Computation and Language arXiv:2603.28858 (cs) [Submitted on 30 Mar 2026] Title:OptiMer: Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training Authors:Haiyue Song, Masao Utiyama View a PDF of the paper titled OptiMer: Optimal Distribution Vector Merging Is Better than Data Mixing for Continual Pre-Training, by Haiyue Song and Masao Utiyama View PDF HTML (experimental) Abstract:Continual pre-training is widely used to adapt LLMs to target languages and domains, yet the mixture ratio of training data remains a sensitive hyperparameter that is expensive to tune: they must be fixed before training begins, and a suboptimal choice can waste weeks of compute. In this work, we propose OptiMer, which decouples ratio selection from training: we train one CPT model per dataset, extract each model's distribution vector, which represents the parameter shift induced by that dataset, and search for optimal composition weights post-hoc via Bayesian optimization. Experiments on Gemma 3 27B across languages (Japanese, Chinese) and domains (Math, Code) show that OptiMer consistently outperforms data mixture and model averaging baselines with 15-35 times lower search cost. Key findings reveal that 1) the optimized weights can be interpreted as data mixture ratios, and retraining with these ratios improves data mixture CPT, and 2) the same vector pool can be re-optimized for a given objective without any retraining, producing target-tailor...

Originally published on April 01, 2026. Curated by AI News.

Related Articles

Can Claude Opus 4.7 and Ensemble AI Models Finally Make Code Review Reliable?
Llms

Can Claude Opus 4.7 and Ensemble AI Models Finally Make Code Review Reliable?

Ensemble AI models like Claude Opus 4.7 transform code review reliability. Discover how multi-model approaches catch subtle bugs human re...

AI Tools & Products · 9 min ·
Starbucks Tests AI-Driven Drink Discovery Through ChatGPT Integration |
Llms

Starbucks Tests AI-Driven Drink Discovery Through ChatGPT Integration |

Not long ago, the idea that a customer could describe a mood instead of a menu item and receive a tailored drink recommendation would hav...

AI Tools & Products · 7 min ·
AI XRP Price Prediction: ChatGPT and Claude Predict XRP Price After Hitting $1.45
Llms

AI XRP Price Prediction: ChatGPT and Claude Predict XRP Price After Hitting $1.45

XRP has seen recent gains due to Rakuten listing it as a payment method and Ripple's partnership with Kyobo Life. Bitcoin's rise also con...

AI Tools & Products · 6 min ·
I canceled ChatGPT Plus and 2 other AI subscriptions — here’s what I replaced them with
Llms

I canceled ChatGPT Plus and 2 other AI subscriptions — here’s what I replaced them with

I was paying for Adobe Firefly, ChatGPT Plus, and Perplexity Pro at the same time. Here's why I canceled all three, and what replaced them.

AI Tools & Products · 6 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime