Click any tag below to further narrow down your results
Links
NVIDIA's new GB200 NVL72 AI cluster has increased the performance of Mixture of Experts (MoE) models by ten times compared to its previous generation. This boost is attributed to a co-design approach that enhances parallel processing and optimizes resource allocation for AI tasks. The Kimi K2 Thinking model, tested on this architecture, showcases significant improvements in efficiency and capability.
The launch of Gemini 3 has demonstrated significant performance improvements over its predecessor, Gemini 2.5, despite having the same parameter count. This, along with Nvidia's strong earnings report, suggests that pre-training scaling laws remain effective when combined with algorithmic advancements and improved compute power. Together, these developments challenge the notion that AI model performance has plateaued.
The NVIDIA HGX B200, now available in the Cirrascale AI Innovation Cloud, offers an 8-GPU configuration that significantly enhances AI performance, achieving up to 15X faster inference compared to the previous generation. With advanced features such as the second-generation Transformer Engine and NVLink interconnect, it is designed for demanding AI and HPC workloads, ensuring efficient scalability and lower operational costs.