NVIDIA spent $10 billion on developing its next-generation Blackwell GPU

NVIDIA spent $10 billion on the development of its next-gen Blackwell AI GPU, with the company spending big on the future of AI GPUs.

Published
Updated
2 minutes & 29 seconds read time

The R&D budget of NVIDIA's next-generation Blackwell GPU architecture was "something like" $10 billion, according to NVIDIA CEO Jensen Huang.

Jensen was on CNBC this week, where he said NVIDIA's next-gen Blackwell AI GPU "will cost $30-$40K USD. The very first one, the R&D budget of this generation is probably something like $10 billion USD". Rumor has it that NVIDIA's new Blackwell B200 AI GPU will cost over $6000 to make, so they can expect some truly healthy profits from Blackwell GPU sales.

Blackwell costs $30,000 to $40,000 per AI GPU, which means at a cost of, let's say, $8000, NVIDIA is in the green, pun intended. If we thought that record-breaking share prices pushing the company over the $1 trillion market cap was good, Blackwell will hit another few home runs for NVIDIA.

Now, as for NVIDIA's new Blackwell B200 AI GPU: the new NVIDIA B200 AI GPU features a whopping 208 billion transistors made on TSMC's new N4P process node. It also has 192GB of ultra-fast HBM3E memory with 8TB/sec of memory bandwidth. NVIDIA is not using a single GPU die here, but a multi-GPU die with a small line between the dies differentiating the two dies, a first for NVIDIA.

The two chips think they're a single chip, with 10TB/sec of bandwidth between the GPU dies, which have no idea they're separate. The two B100 GPU dies think they're a single chip, with no memory locality issues and no cache issues... it just thinks it's a single GPU and does its (AI) thing at blistering speeds, which is thanks to NV-HBI (NVIDIA High Bandwidth Interface).

NVIDIA spent $10 billion on developing its next-generation Blackwell GPU 76

NVIDIA's new B200 AI GPU has 20 petaflops of AI performance from a single GPU, compared to just 4 petaflops of AI performance from the current H100 AI GPU. Impressive. Note: NVIDIA is using a new FP4 number format for these numbers, with H100 using the FP8 format, which means that B200 has 2.5x theoretical FP8 compute than H100. Still, very impressive.

Each of the B200 GPU dies two full reticle-size chips, with 4 x HBM3E stacks of 24GB each, along with 1TB/sec of memory bandwidth on a 1024-bit memory interface. The total of 192GB of HBM3E memory, with 8TB/sec of memory bandwidth, is a huge upgrade over the H100 AI GPU, which had 6 x HBM3 stacks of 16GB each (at first, H200 kicked that up to 24GB per stack).

NVIDIA is using an all-new NVLink chip design that has 1.8TB/sec of bi-directional bandwidth and packing support for a 576 GPU NVLink domain. This NVLink chip itself features 50 billion transistors, manufactured by TSMC on the same N4P process node.

Buy at Amazon

NVIDIA H100 80 GB Graphic Card PCIe HBM2e Memory 350W

TodayYesterday7 days ago30 days ago
Buy at Newegg
$139.99$139.99$139.99
$29949.95$30099.99$30099.99
* Prices last scanned on 4/30/2024 at 9:45 pm CDT - prices may not be accurate, click links above for the latest price. We may earn an affiliate commission.

Anthony joined the TweakTown team in 2010 and has since reviewed 100s of graphics cards. Anthony is a long time PC enthusiast with a passion of hate for games built around consoles. FPS gaming since the pre-Quake days, where you were insulted if you used a mouse to aim, he has been addicted to gaming and hardware ever since. Working in IT retail for 10 years gave him great experience with custom-built PCs. His addiction to GPU tech is unwavering and has recently taken a keen interest in artificial intelligence (AI) hardware.

Newsletter Subscription

Related Tags