[2604.04204] Which English Do LLMs Prefer? Triangulating Structural Bias Towards American English in Foundation Models

[2604.04204] Which English Do LLMs Prefer? Triangulating Structural Bias Towards American English in Foundation Models

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2604.04204: Which English Do LLMs Prefer? Triangulating Structural Bias Towards American English in Foundation Models

Computer Science > Computation and Language arXiv:2604.04204 (cs) [Submitted on 5 Apr 2026] Title:Which English Do LLMs Prefer? Triangulating Structural Bias Towards American English in Foundation Models Authors:Mir Tafseer Nayeem, Davood Rafiei View a PDF of the paper titled Which English Do LLMs Prefer? Triangulating Structural Bias Towards American English in Foundation Models, by Mir Tafseer Nayeem and 1 other authors View PDF HTML (experimental) Abstract:Large language models (LLMs) are increasingly deployed in high-stakes domains, yet they expose only limited language settings, most notably "English (US)," despite the global diversity and colonial history of English. Through a postcolonial framing to explain the broader significance, we investigate how geopolitical histories of data curation, digital dominance, and linguistic standardization shape the LLM development pipeline. Focusing on two dominant standard varieties, American English (AmE) and British English (BrE), we construct a curated corpus of 1,813 AmE--BrE variants and introduce DiAlign, a dynamic, training-free method for estimating dialectal alignment using distributional evidence. We operationalize structural bias by triangulating evidence across three stages: (i) audits of six major pretraining corpora reveal systematic skew toward AmE, (ii) tokenizer analyses show that BrE forms incur higher segmentation costs, and (iii) generative evaluations show a persistent AmE preference in model outputs. To our ...

Originally published on April 07, 2026. Curated by AI News.

Related Articles

AI isn’t just evolving, it’s taking over how we think — here’s my take on ChatGPT’s big upgrade, Google's plan to make AI invisible, DeepSeek’s return and more
Llms

AI isn’t just evolving, it’s taking over how we think — here’s my take on ChatGPT’s big upgrade, Google's plan to make AI invisible, DeepSeek’s return and more

This week, AI is all about "super" apps, invisible apps and a geopolitical AI race

AI Tools & Products · 10 min ·
Llms

Founding Engineer (Full-Stack / AI) – Build the Future of Personalized Healthcare (San Francisco, In-Person)

Hi everyone Galen AI is an early-stage, YC-backed healthtech startup building a personal AI doctor by combining clinical data, wearable d...

Reddit - ML Jobs · 1 min ·
Llms

Looking for Job Opportunities — Senior MLOps / LLMOps Engineer (Remote / Visa Sponsorship)

Hi Everyone 👋 I’m a Senior MLOps / LLMOps Engineer with ~5 years of experience building and operating production-scale ML & LLM platf...

Reddit - ML Jobs · 1 min ·
Llms

Early career / PhD (USA only) - $80-120/hr

Mercor is hiring Machine Learning Engineers to: Draft detailed natural-language plans and code implementations for machine learning tasks...

Reddit - ML Jobs · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime