Key Takeaways
- Cerebras’s Wafer-Scale Engine processors will be integrated into AWS data centers for AI inference workloads.
- Financial details of the multiyear agreement remain undisclosed.
- According to Cerebras, its chip technology can execute inference operations up to 25 times faster than Nvidia’s GPU solutions.
- In January 2026, OpenAI entered into a $10 billion-plus agreement with Cerebras.
- The chip startup secured $1 billion in funding during February 2026, pushing its valuation to around $23 billion.
Amazon Web Services has entered into a long-term collaboration with semiconductor startup Cerebras Systems, bringing the company’s Wafer-Scale Engine technology into AWS’s global data center infrastructure. The strategic focus centers on AI inference operationsāthe computational stage where artificial intelligence models generate responses to user prompts.
As the world’s leading cloud computing platform, AWS has traditionally relied on proprietary silicon solutions, particularly its Trainium processors designed by subsidiary Annapurna Labs. This latest arrangement sees AWS planning to merge Trainium capabilities with Cerebras technology to create a high-performance inference platform.
According to Cerebras, its Wafer-Scale Engine architecture excels during the “decode” stage of inferenceāwhere models actually produce their outputādelivering performance up to 25 times superior to Nvidia’s graphics processing units.
The offering will target the premium market segment. Cerebras CEO Andrew Feldman explained, “If you want slow inference, there will be cheaper ways to go.” AWS has indicated it will continue providing budget-friendly inference options powered exclusively by Trainium chips.
Cerebras Momentum Builds
This partnership follows closely on the heels of another major announcement from January 2026, when OpenAI signed its own massive deal with Cerebras valued at over $10 billion. That arrangement aims to power ChatGPT infrastructure using Cerebras hardware, with OpenAI targeting deployment of computing capacity reaching 750 megawatts.
One month later, in February 2026, Cerebras completed a $1 billion financing round, elevating total capital raised to $2.6 billion and establishing a company valuation near $23 billion. The investor roster includes prominent names such as Fidelity Management, Benchmark, Tiger Global, and Coatue.
Cerebras had initially pursued a public listing by filing for an IPO in September 2024, though it ultimately pulled that application approximately 12 months afterward.
Nvidia Faces Growing Competition
This AWS-Cerebras collaboration represents another competitive pressure point for Nvidia in the inference computing space. The artificial intelligence sector has been experiencing a strategic pivot from training-focused workloads, where Nvidia maintains market leadership, toward inference applications that prioritize processing velocity.
Nvidia hasn’t remained passive in response. During December 2025, the company finalized a $20 billion licensing arrangement with chip innovator Groq. Nvidia has also signaled plans to launch a new processing architecture incorporating Groq’s technological advances in the coming period.
Nafea Bshara, co-founder of Annapurna Labs at AWS, emphasized that the Cerebras collaboration targets simultaneous improvements in performance and economics. “Our job is to push the speed and lower the price,” he stated.
Shares of Amazon AMZN traded down 0.44% at the time of this report.





