Cerebras’s Strategy to Challenge Nvidia in AI Hardware
The AI vendor's director of product marketing says the company first focused on large chips and later turned attention to ramping up inferencing.
By
- Esther Shittu, News Writer
- Shaun Sutner, Senior News Director
Published: 12 Aug 2025
While Nvidia dominates the AI chip market, Cerebras Systems is working to be a differentiator.
Founded in 2015, the AI inference vendor started with the idea of creating the world's largest computer chip.
This mission led the vendor to create a chip in 2019 that was about the size of a dinner plate, said James Wang, director of product marketing at Cerebras, on the latest Targeting AI podcast from Informa TechTarget.
If you're just going to make small changes, Nvidia will catch up and beat you. James WangDirector of product marketing, Cerebras"Nothing like that had ever been done before," Wang said. He added that he followed the development as a technology analyst then and that, while vendors like Graphcore or SambaNova were trying to compete with Nvidia with smaller chips, only Cerebras went large.
"I thought that was probably the only chance anyone had of taking on Nvidia," Wang said. "If you're just going to make small changes, Nvidia will catch up and beat you."
Cerebras' approach to AI chips has since changed. Instead of trying to train the chips, it's now gone into AI inference.
In August 2024, the AI vendor launched Cerebras Inference, an AI inference product that delivers 1,800 tokens per second for Llama 3.1 8B and 450 tokens per second for Llama 3.1 70B.
The vendor has seen much growth since the launch of Cerebras Inference, Wang said.
"The amount of inbound interest, the amount of companies that can use our products, the amount of startups building on Cerebras have just exploded," he said.
He continued by saying that a vendor needs a large lead to compete against a technology giant like Nvidia; otherwise, it erodes its lead in one generation.
On Aug. 5, Cerebras Systems announced that it will help power OpenAI's open model gpt-oss-120B.
Esther Shittu is an Informa TechTarget news writer and podcast host covering artificial intelligence software and systems. Shaun Sutner is senior news director for Informa TechTarget's information management team, driving coverage of AI, analytics and data management technologies, and big tech and federal regulation. He is a veteran journalist with more than 30 years of news experience. Together, they host Informa TechTarget's Targeting AI podcast.
Dig Deeper on AI technologies
-
10 top AI hardware and chip-making companies in 2026
By: Kelly Richardson
-
Nvidia bets on Intel: What it means for IT leaders
By: Sean Kerner
-
FuriosaAI to fuel LG Exaone LLM: Is it a challenge to Nvidia?
By: Shane Snider
-
Cerebras launches Alibaba model, forms key AI partnerships
By: Esther Shittu
Internet of Things Technology
- Why Trustworthy Software Is Essential for Industrial Safety
- 8 Key Reasons Explainer Videos Drive Success in the IoT Industry
- Amphenol Expands Liquid Cooling Connectors for AI, ESS & EV Systems
- Discovery’s IoT Solution Transforms Supply Chain Management
- A Beginner’s Guide to Open‑Source Terminology
- Essential Security Practices for Fog Computing
- IoT Takes Center Stage at CES 2015: Trends, Security, and Low‑Power Connectivity
- 6 Powerful Ways IoT Transforms Your Business
- Telcos Turning to Value‑Added ICT Services for Revenue Growth – Are They Ready?
- Secure IoT: Best Practices for Building Trustworthy Connected Products