To optimize cloud-based collaborative computing and configure AI infrastructure, Google announced that its next-generation AI computing accelerator, TPU v5p, unveiled at the end of last year, is now available globally. Starting in May, it will also incorporate NVIDIA's H100 accelerator, codenamed "Hopper," to create a computing device called A3 Mega. Additionally, it plans to introduce NVIDIA's recently unveiled next-generation accelerator, codenamed "Blackwell," with the GB200 NVL72 computing system expected to be deployed in early 2025. ▲Google will begin incorporating NVIDIA's H100 accelerator, codenamed "Hopper," in May to create a computing device called A3 Mega. Previously, when introducing the new TPU v5p, Google touted its scalability and flexible deployment capabilities, highlighting it as Google's most powerful tensor accelerator to date. It can achieve twice the computing power, more than three times the memory bandwidth, and near-linear data throughput on a single processor. It can also support four times the size of next-generation AI models while simultaneously reducing training time for existing models by 2.8 times. A single TPU v5p pod will consist of 8960 chips, more than double the number of chips used in the TPU v4 pod, thus meeting the demands of large-scale artificial intelligence computing. ▲Google's most powerful Tensor Accelerator to date, the TPU v5p. In addition to driving artificial intelligence computing, Google emphasizes that the new generation of TPUs will accelerate more cloud tasks and further improve the performance of Google's services such as Search, YouTube, Gmail, Google Maps, and online services including the Google Play Store. It will also enable faster performance for many Android devices combined with cloud-based collaborative computing applications, and provide a more convenient user experience when combined with on-device computing. Besides incorporating its own TPUs and NVIDIA accelerated computing components, Google will continue to collaborate with AMD to integrate its accelerated computing products, thereby providing more diverse options for artificial intelligence acceleration. ▲The GB200 NVL72 computing system is expected to be introduced in early 2025. In addition to continuing to accelerate AI computing efficiency with its own TPUs, Google also plans to accelerate overall computing efficiency with the custom Arm Neoverse V2 core architecture processor "Axion". Compared to traditional x86 architecture processors, this represents a 50% improvement in execution efficiency and a 60% reduction in power consumption. It also emphasizes a 30% increase in execution performance compared to current Arm architecture processors used for cloud collaborative computing, thereby significantly reducing carbon emissions. ▲The custom Arm Neoverse V2 core architecture processor "Axion" accelerates overall computing efficiency, representing a 50% improvement in execution efficiency and a 60% reduction in power consumption compared to traditional x86 architecture processors. On the other hand, Google continues its collaboration with Intel to introduce the fifth-generation Xeon Scalable server processor, thereby creating the previewable C4 and the N4 virtual machine currently available globally. Additionally, the C3 virtual machine for bare-metal use will also be launched. ...