Elon Musk confirms xAI used OpenAI’s models to train Grok | The Verge

Elon Musk confirms xAI used OpenAI’s models to train Grok | The Verge

The Verge - AI 5 min read

About this article

During the Musk v. Altman trial, while on the stand, Elon Musk said it was “partly” true that xAI had used model distillation of OpenAI’s models to improve Grok.

AINewsPolicyElon Musk confirms xAI used OpenAI’s models to train GrokHe said it was “partly” true that the company had used model distillation to improve xAI’s models.He said it was “partly” true that the company had used model distillation to improve xAI’s models.by Hayden FieldApr 30, 2026, 6:16 PM UTCLinkShareGiftIf you buy something from a Verge link, Vox Media may earn a commission. See our ethics statement.Image: Cath Virginia / The Verge, Getty ImagesPart OfLIVELive updates from Elon Musk and Sam Altman’s court battle over the future of OpenAIsee all updates Hayden Field is The Verge’s senior AI reporter. An AI beat reporter for more than five years, her work has also appeared in CNBC, MIT Technology Review, Wired UK, and other outlets.In a federal courtroom in California on Thursday, Elon Musk testified that his own AI startup, xAI, has used OpenAI’s models to improve its own.The matter at question is model distillation, a common industry practice by which one larger AI model acts as a “teacher” of sorts to pass on knowledge to a smaller AI model, the “student.” Although it’s often used legitimately within companies using one of their own AI models to train another, it’s also a practice that’s sometimes used by smaller AI labs to try to get their models to mimic the performance of a larger competitor’s model.Asked on the stand whether he knew what model distillation was, Musk said it’s to use one AI model to train another. When asked whether xAI has distilled Ope...

Originally published on April 30, 2026. Curated by AI News.

Related Articles

Llms

A Hackable ML Compiler Stack in 5,000 Lines of Python [P]

Hey r/MachineLearning, The modern ML (LLM) compiler stack is brutal. TVM is 500K+ lines of C++. PyTorch piles Dynamo, Inductor, and Trito...

Reddit - Machine Learning · 1 min ·
Llms

Applying Karpathy's autoresearch to a 33M-token public transit dataset (14% improvement, replication notes) [P]

Hello r/MachineLearning! I work in the US transit industry and I went all-in on learning AI & ML a few months ago. When I heard about...

Reddit - Machine Learning · 1 min ·
Machine Learning

Seems ICML is rejecting MANY unanimous positively rated papers [D]

My 4444 (4443 pre-rebuttal) got rejected (as expected). Just copying a reply I wrote a couple of days ago before decisions were out: Ther...

Reddit - Machine Learning · 1 min ·
Llms

Track real-time GPU and LLM pricing across all cloud and inference providers

Deploybase is a dashboard for tracking real-time GPU and LLM pricing across cloud and inference providers. You can view performance stats...

Reddit - Artificial Intelligence · 1 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime