Plato Data Intelligence.
Vertical Search & Ai.

Dell exec reveals Nvidia has a 1,000-watt GPU in the works

Date:

If you thought Nvidia’s 700W H100s were hot and power-hungry machines, just wait until the GPU slinger’s B100 arrives later this year.

According to Dell Technologies COO Jeff Clarke, Nvidia’s latest AI accelerator will consume 1,000 watts – 42 percent more than its predecessor. But don’t worry, he’s pretty sure liquid cooling won’t be required to tame the beast.

“We’re excited about what’s happening with the H200 and its performance improvement,” Clarke told investors on Dell’s earnings call [PDF] last week, before adding he feels the same emotion about Nvidia’s forthcoming B100 accelerator and another he referred to as the B200.

He opined that direct liquid cooling won’t be needed to handle GPUs that consume 1,000 watts apiece – a level he said “happens next year with the B200.”

It’s not entirely clear what card Clarke is referring to with the “B200,” since no chip by that moniker appears on the roadmap Nvidia shared with investors last fall. However, we suspect Clarke is actually referring to the GB200 Superchip which, like the GH200, is expected to combine Nvidia’s Grace CPU with its B100 GPU.

According to an investor presentation released this month, Nvidia plans to shift to a One-Year release cadence.

According to an investor presentation released this month, Nvidia plans to shift to a one-year release cadence – Click to enlarge

Based on what we know of the Grace CPU in the GH200, and assuming no major changes in power consumption, that would put the GB200’s thermal design power (TDP) somewhere in the neighborhood of 1,300 watts – 30 percent higher than its predecessor.

It’s also possible that Nvidia has another card up its sleeve that we don’t know about yet. Details of the GPU giant’s next-gen Blackwell architecture remain scanty.

Nomenclature aside, Clarke suggested the forthcoming chip would provide an opportunity to showcase Dell’s expertise in other forms of liquid cooling at scale. He referred to “things in fluid chemistry and performance, our interconnect work, the telemetry we are doing, the power management work we’re doing” as steps toward alternatives to direct liquid cooling, even for very dense chips.

Nvidia declined to comment – as you’d expect, given its annual GTC conference is only a few weeks away. The Register will be onsite at the event to bring you all the details when they drop.

The B100 isn’t expected to launch until late 2024 after Nvidia’s bandwidth juiced H200 GPUs debut in the first half of the year.

Announced in late 2023, the H200 is a refresh of the H100 with up to 141GB of HBM3e memory that’s good for a whopping 4.8TB/sec of bandwidth. Nvidia claims the device can double the performance of large language models including Llama 70B, thanks to the chip’s HBM3e memory stacks.

Even with two new accelerators slated to hit the market this year, analysts warn Nvidia’s supply of GPUs will remain supply constrained. That’s despite reports predicting Nvidia could move more than triple shipments of GPUs in 2024.

Beyond its new accelerators, Nvidia’s roadmap also calls for faster, more capable InfiniBand and Ethernet NICs and switches capable of 800Gb/sec of bandwidth per port before the year is out. ®

spot_img

Latest Intelligence

spot_img

Chat with us

Hi there! How can I help you?