As we’ve seen LLMs and generative AI come screaming into our consciousness in recent months, it’s clear that these models take enormous amounts of compute power to train and run. Recognizing this, Google Cloud announced a new A3 supercomputer virtual machine today at Google I/O.
The A3 has been purpose-built to handle the considerable demands of these resource-hungry use cases.
“A3 GPU VMs were purpose-built to deliver the highest-performance training for today’s ML workloads, complete with modern CPU, improved host memory, next-generation NVIDIA GPUs and major network upgrades,” the company wrote in an announcement.
Specifically, the company is arming these machines with NVIDIA’s H100 GPUs and combining that with a specialized data center to derive immense computational power with high throughput and low latency, all at what they suggest is a more reasonable ...