[2506.05850] Cross-lingual Collapse: How Language-Centric Foundation Models Shape Reasoning in Large Language Models

[2506.05850] Cross-lingual Collapse: How Language-Centric Foundation Models Shape Reasoning in Large Language Models

arXiv - AI 4 min read Article

Summary

The paper investigates 'Cross-lingual Collapse' in large language models (LLMs), revealing how reasoning capabilities can revert to a dominant pre-training language, impacting multilingual performance.

Why It Matters

Understanding Cross-lingual Collapse is crucial for improving multilingual LLMs. It highlights the trade-off between reasoning depth and language fidelity, which is essential for developing more effective AI systems that can perform well across different languages.

Key Takeaways

  • Cross-lingual Collapse occurs when multilingual LLMs revert to their dominant pre-training language during reasoning.
  • The phenomenon is amplified by English-centric priors and long-CoT optimization.
  • Interventions can mitigate collapse but reveal a trade-off between performance and language fidelity.

Computer Science > Computation and Language arXiv:2506.05850 (cs) [Submitted on 6 Jun 2025 (v1), last revised 23 Feb 2026 (this version, v3)] Title:Cross-lingual Collapse: How Language-Centric Foundation Models Shape Reasoning in Large Language Models Authors:Cheonbok Park, Jeonghoon Kim, Joosung Lee, Sanghwan Bae, Jaegul Choo, Kang Min Yoo View a PDF of the paper titled Cross-lingual Collapse: How Language-Centric Foundation Models Shape Reasoning in Large Language Models, by Cheonbok Park and 5 other authors View PDF HTML (experimental) Abstract:Reinforcement learning with verifiable reward (RLVR) has been instrumental in eliciting strong reasoning capabilities from large language models (LLMs) via long chains of thought (CoT). During RLVR training, we formalize and systemically study an empirical phenomenon whereby a multilingual model's CoT reverts to its dominant pre-training language (e.g., English) even when prompted in another language, which we term Cross-lingual Collapse. Because the long-CoT regime magnifies exposure to linguistic priors, the underlying trade-off between maximizing reasoning depth and preserving target-language fidelity has remained under-characterized. To examine this trade-off, we train LLMs with Group-Relative Policy Optimization (GRPO) on translated versions of math datasets widely used to elicit long-CoT reasoning. Throughout training, we track both task accuracy and the language consistency of reasoning chains. Our experiments yield three ...

Related Articles

Llms

Building knowledge bases from YouTube data using LLMs -- my workflow after 52 guides

I've been building a system that turns YouTube channels into structured knowledge bases. Thought I'd share the workflow since Karpathy's ...

Reddit - Artificial Intelligence · 1 min ·
What is AI, how do apps like ChatGPT work and why are there concerns?
Llms

What is AI, how do apps like ChatGPT work and why are there concerns?

AI is transforming modern life, but some critics worry about its potential misuse and environmental impact.

AI News - General · 7 min ·
[2603.29957] Think Anywhere in Code Generation
Llms

[2603.29957] Think Anywhere in Code Generation

Abstract page for arXiv paper 2603.29957: Think Anywhere in Code Generation

arXiv - Machine Learning · 3 min ·
[2603.16880] NeuroNarrator: A Generalist EEG-to-Text Foundation Model for Clinical Interpretation via Spectro-Spatial Grounding and Temporal State-Space Reasoning
Llms

[2603.16880] NeuroNarrator: A Generalist EEG-to-Text Foundation Model for Clinical Interpretation via Spectro-Spatial Grounding and Temporal State-Space Reasoning

Abstract page for arXiv paper 2603.16880: NeuroNarrator: A Generalist EEG-to-Text Foundation Model for Clinical Interpretation via Spectr...

arXiv - Machine Learning · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime