3 links
tagged with all of: architecture + machine-learning
Click any tag below to further narrow down your results
Links
TPUs, or Tensor Processing Units, are Google's custom ASICs designed for high throughput and energy efficiency, particularly in AI applications. They utilize a unique architecture featuring systolic arrays and a co-design with the XLA compiler to achieve scalability and performance, contrasting significantly with traditional GPUs. The article explores the TPU's design philosophy, internal architecture, and their role in powering Google's AI services.
Google Cloud has expanded its collection of generative AI use cases to over 600 examples, providing 101 architectural blueprints to guide developers and business leaders in implementing AI solutions. The blueprints address real-world challenges across various industries, illustrating how Google Cloud technologies can streamline operations, enhance customer experiences, and improve decision-making processes.
The article explores the architecture and functionality of NVIDIA GPUs, detailing their compute cores, memory hierarchy, and comparison with TPUs. It emphasizes the importance of Tensor Cores for matrix multiplication in modern machine learning tasks and outlines the evolution of GPU specifications across generations. The content builds on previous chapters, providing a comprehensive understanding of GPU capabilities in the context of large language models.