6 min read
|
Saved February 14, 2026
|
Copied!
Do you care about this?
Eric Vishria discusses Nvidia's dominance in AI but highlights a potential weakness in its chip architecture. He argues that new SRAM-based designs from companies like Groq and Cerebras show superior performance for AI inference, challenging Nvidia's lead.
If you do, here's more
Nvidia currently dominates the AI chip market, benefiting from high gross and operating margins. However, a conversation is brewing around the potential vulnerabilities in Nvidia's position, particularly concerning GenAI inference. While Nvidia’s chips excel at training AI models, the architecture designed for inference is gaining attention. Companies like Groq and Cerebras, which utilize SRAM memory architecture, are showing impressive performance metrics that far exceed those of Nvidia’s offerings. For instance, Groq recently showcased a capability of 1,200 tokens per second, while Nvidia's best models are significantly slower.
The debate about the ROI of AI infrastructure is shifting. Instead of questioning the value of AI, the focus is on identifying the winners in this space. The architecture of AI chips is essential; training chips are complex and harder to adopt, while inference chips are simpler and more flexible. This distinction is critical as more companies explore the potential of AI workloads. If SRAM architecture can deliver faster inference speeds, it could redefine how AI applications function, making them more efficient and responsive.
Moreover, Nvidia reported that 35-40% of its datacenter revenue came from inference over the past year, and this number is expected to rise. The question remains whether customers will continue to invest heavily in Nvidia's chips if alternatives provide significantly better performance at a fraction of the cost. The landscape of AI computing is in flux, and as the technology evolves, Nvidia's dominance could be challenged by these emerging players.
Questions about this article
No questions yet.