A new version of OpenAI's Codex is powered by a new dedicated chip | TechCrunch
Summary
OpenAI has launched Codex-Spark, a lightweight coding tool powered by Cerebras' dedicated chip, enhancing real-time collaboration and rapid prototyping capabilities for users.
Why It Matters
The introduction of Codex-Spark represents a significant advancement in AI-driven coding tools, enabling faster inference and real-time collaboration. This development not only improves productivity for developers but also highlights the growing integration of specialized hardware in AI applications, potentially reshaping how coding tasks are approached in the future.
Key Takeaways
- Codex-Spark is designed for rapid prototyping and real-time collaboration, enhancing productivity for developers.
- The tool is powered by Cerebras' Wafer Scale Engine 3, which boasts 4 trillion transistors for low-latency performance.
- OpenAI's partnership with Cerebras marks a significant step in integrating specialized hardware into AI solutions.
- Codex-Spark aims to support both rapid iteration and long-running tasks, offering flexibility in coding workflows.
- The launch reflects a broader trend of AI tools becoming more accessible and efficient for everyday use.
On Thursday, OpenAI announced the release of a light-weight version of its agentic coding tool Codex, the latest model of which OpenAI launched earlier this month. GPT-5.3-Codex-Spark is described by the company as a “smaller version” of that model, one that is designed for faster inference. To power that inference, OpenAI has brought in a dedicated chip from its hardware partner Cerebras, marking a new level of integration in the company’s physical infrastructure. The partnership between Cerebras and OpenAI was announced last month, when OpenAI said that it had reached a multi-year agreement with the firm worth over $10 billion. “Integrating Cerebras into our mix of compute solutions is all about making our AI respond much faster,” the company said at the time. Now, OpenAI calls Spark the “first milestone” in that relationship. Spark, which OpenAI says is designed for swift, real-time collaboration and “rapid iteration,” will be powered by Cerebras’ Wafer Scale Engine 3. The WSE-3 is Cerebras’ third-generation waferscale megachip, decked out with 4 trillion transistors. OpenAI describes the new lightweight tool as a “daily productivity driver, helping users with rapid prototyping” rather than the longer, heavier tasks that the original 5.3 is designed for. Spark is currently enjoying a research preview for ChatGPT Pro users in the Codex app. In a tweet in advance of the announcement, CEO Sam Altman seemed to hint at the new model. “We have a special thing launching to Cod...