Collaborating Authors


Google Cloud's New TPU v4 ML Hub Packs 9 Exaflops of AI


Almost exactly a year ago, Google launched its Tensor Processing Unit (TPU) v4 chips at Google I/O 2021, promising twice the performance compared to the TPU v3. At the time, Google CEO Sundar Pichai said that Google's datacenters would "soon have dozens of TPU v4 Pods, many of which will be operating at or near 90 percent carbon-free energy." Now, at Google I/O 2022, Pichai revealed the blue-ribbon fruit of those labors: a TPU v4-powered datacenter in Mayes County, Oklahoma, that Google says is the world's largest publicly available machine learning hub. "This machine learning hub has eight Cloud TPU v4 Pods, custom-built on the same networking infrastructure that powers Google's largest neural models," Pichai said. Google's TPU v4 Pods consist of 4,096 TPU v4 chips, each of which delivers 275 teraflops of ML-targeted bfloat16 ("brain floating point") performance.