Google unveiled its latest tensor processing unit (TPU), Ironwood, at Hot Chips 2025, designed specifically for large-scale AI inference with a performance capacity of 42.5 Exaflops. Featuring up to 9,216 chips per node and significant improvements in power efficiency and reliability, Ironwood aims to excel in handling advanced AI models such as LLMs and mixture-of-expert systems. The architecture incorporates advanced memory subsystems and enhanced RAS features, making it a pivotal development for Google's AI computing capabilities.
google ✓
+ tpu
hot-chips ✓
ai-inference ✓
power-efficiency ✓