Elon Musk testifies that xAI trained Grok on OpenAI models | TechCrunch
About this article
"Distillation" is a hot topic as frontier labs try to prevent smaller competitors from copying their models.
OpenAI and Anthropic have been on the warpath lately against third-party efforts to train new AI models by prompting their publicly-accessible chatbots and APIs, a process known as “distillation.” That conversation has focused on Chinese firms using distillation to create open-weight models that are nearly as capable as U.S. offerings, but available at a much lower cost. However, tech workers have widely assumed that American labs use these techniques on each other to avoid falling behind competitors. Now, we know it’s true in at least one case: on the stand in a California federal court on Thursday, Elon Musk was asked if xAI has used distillation techniques on OpenAI models to train Grok, and he asserted it was a general practice among AI companies. Asked if that meant “yes,” he said, “Partly.” Musk is in the process of suing OpenAI, CEO Sam Altman, and Greg Brockman, alleging they breached the original nonprofit mission for OpenAI by shifting the entity to a for-profit structure. That trial began this week, featuring testimony from the tech leader. Musk’s admission is notable because distillation threatens AI giants by undermining the advantage they’ve built by investing in compute infrastructure. This allows other software makers to create models that are nearly as capable on the cheap. There’s no small amount of irony here, given the bending and alleged breaking of copyright rules by frontier labs in their search for sufficient data to train their models. It’s no surp...