NVIDIA Steps Up Its Game: Unveiling the Blackwell Architecture and the Powerhouse B200/B100 Accelerators
April 08, 2024
NVIDIA Steps Up Its Game: Unveiling the Blackwell Architecture and the Powerhouse B200/B100 Accelerators

NVIDIA, a titan in the world of generative AI accelerators, is not one to rest on its laurels. Despite already dominating the accelerator market, the tech giant is determined to push the envelope further. With the unveiling of its next-generation Blackwell architecture and the B200/B100 accelerators, NVIDIA is set to redefine what's possible in AI computing yet again.

As we witnessed the return of the in-person GTC for the first time in five years, NVIDIA's CEO, Jensen Huang, took center stage to introduce an array of new enterprise technologies. However, it was the announcement of the Blackwell architecture that stole the spotlight. This move marks a significant leap forward, building upon the success of NVIDIA's H100/H200/GH200 series.

Named after the American statistical and mathematical pioneer Dr. David Harold Blackwell, this architecture embodies NVIDIA's commitment to innovation. Blackwell aims to elevate the performance of NVIDIA's datacenter and high-performance computing (HPC) accelerators by integrating more features, flexibility, and transistors. This approach is a testament to NVIDIA's strategy of blending hardware advancements with software optimization to tackle the evolving needs of high-performance accelerators.

At the heart of the Blackwell architecture is the notion of going big—literally. The B200 modules, powered by the Blackwell GPU, will boast two GPU dies in a single package, signaling NVIDIA's foray into chiplet designs. Despite constraints in die space, NVIDIA is leveraging a new TSMC 4NP node, optimizing architectural efficiency to achieve substantial performance gains without the benefits of a major node shrink.


With an impressive 208 billion transistors across the complete accelerator, NVIDIA's first multi-die chip represents a bold step in unified GPU performance. The Blackwell GPUs are designed to function as a single CUDA GPU, thanks to the NV-High Bandwidth Interface (NV-HBI) facilitating an unprecedented 10TB/second of bandwidth. This architectural marvel is complemented by up to 192GB of HBM3E memory, significantly enhancing both the memory capacity and bandwidth.

However, it's not just about packing more power into the hardware. The Blackwell architecture is engineered to dramatically boost AI training and inference performance while achieving remarkable energy efficiency. This ambition is evident in NVIDIA's projections of a 4x increase in training performance and a staggering 30x surge in inference performance at the cluster level.

The Blackwell lineup features three key variants: the flagship B200 accelerator, the GB200 Grace Blackwell Superchip, and the lower-tier B100 accelerator. Each model targets different performance and power consumption needs, with the GB200 representing the pinnacle of Blackwell's capabilities.


Moreover, NVIDIA is pushing the boundaries of precision with its second-generation transformer engine, capable of handling computations down to FP4 precision. This advancement is crucial for optimizing inference workloads, offering a significant leap in throughput for AI models.

On the connectivity front, NVIDIA introduces NVLink 5, doubling the interconnect bandwidth to 1.8TB/second per GPU. This enhancement, coupled with the launch of the fifth-generation NVLink Switch, underscores NVIDIA's commitment to scalability and networking efficiency, essential for training large AI models.


In summary, NVIDIA's Blackwell architecture and the B200/B100 accelerators represent a formidable advancement in AI accelerator technology. By pushing the limits of architectural efficiency, memory capacity, and computational precision, NVIDIA is not just maintaining its leadership in the AI space but setting new benchmarks for performance and efficiency. As we await the rollout of these groundbreaking products, the tech world watches with anticipation, eager to see how NVIDIA's latest innovations will shape the future of AI computing.