Click any tag below to further narrow down your results
Links
Eric Vishria discusses Nvidia's dominance in AI but highlights a potential weakness in its chip architecture. He argues that new SRAM-based designs from companies like Groq and Cerebras show superior performance for AI inference, challenging Nvidia's lead.
This article discusses the unique difficulties in hardware design for large language model inference, particularly during the autoregressive Decode phase. It identifies memory and interconnect issues as primary challenges and proposes four research directions to improve performance, focusing on datacenter AI but also considering mobile applications.