Understanding NVIDIA’s Product Ecosystem and Naming Conventions

Compute Chips—V100, A100, H100, B200, etc.

These terms are among the most commonly encountered in discussions about artificial intelligence. They refer to AI compute cards, specifically GPU models.

GPU models
GPU Cards (6 Models)

NVIDIA releases a new GPU architecture every few years, each named after a renowned scientist. Cards based on a particular architecture typically start with the first letter of the architecture’s name, except for gaming GPUs. For instance:

  • V100 is based on the Volta architecture (named after Alessandro Volta).
  • A100 derives from the Ampere architecture (named after André-Marie Ampère).
  • H100/H200 originates from Hopper (named after Grace Hopper).
  • B100/B200 come from Blackwell (named after David Blackwell).
NVIDIA releases a new GPU architecture every few years

A model ending in “200” is typically an upgraded version of its “100” predecessor. For example, the H200 is an enhanced version of the H100, featuring improvements such as HBM3e memory.

L40 & L40s. These two models are slightly different in naming. They are based on the Ada Lovelace architecture (honoring the first female programmer). The L40s is an upgraded version of the L40. Both are designed for data center markets, focusing on cost efficiency and performance.

H20 & Export Restrictions The H20 is a modified version of NVIDIA’s GPUs created in response to U.S. export restrictions. Similarly, reports suggest that the B200 may have a restricted variant called B20.

Future Architectures NVIDIA’s next-generation AI platform, Rubin, is scheduled for release in 2026. Initially, many expected GPUs to be labeled R100/R200, but NVIDIA’s roadmap instead suggests an X100 series, leaving room for speculation. By 2028, the company plans to launch the subsequent platform, Feynman.

Superchips—GH200, GB200, etc.

While GPUs form the core of NVIDIA’s computing power, the company also develops complementary solutions beyond GPUs.

Early Partnerships & CPU Development Initially, NVIDIA collaborated with IBM’s POWER CPUs. However, due to performance concerns, the company began developing its own CPUs, such as the Grace CPU (ARM-based), with the Vera CPU also in development.

Using NVLink technology, NVIDIA pairs GPUs and CPUs to create Superchip platforms. Examples include:

  • GH200 (Grace CPU + Hopper GPU)
  • GB200 (Grace CPU + two Blackwell B200 GPUs)
  • GB300 (Blackwell Ultra)

The GB200 is particularly powerful, boasting approximately seven times the performance of the H100.

GB200

Supercomputer Platforms—DGX, EGX, IGX, etc.

At a higher computing tier, NVIDIA has developed supercomputer platforms based on these chip architectures, including DGX, EGX, IGX, HGX, and MGX.

Supercomputer Platforms

The DGX series is well known, with Jensen Huang gifting the first-generation DGX-1 to OpenAI.

the first-generation DGX-1

Today, modern DGX systems typically feature a signature gold design and premium pricing.

For desktop applications, NVIDIA offers DGX Spark and DGX Station, functioning as workstation-class machines.

DGX Spark and DGX Station
DGX Spark and DGX Station

Intra-Node Connectivity—Scale Up (Supernodes)—DGX GB200 NVL72

At the communication level, NVIDIA has developed NVLink, a technology designed to interconnect GPUs, effectively replacing PCIe. It also connects CPUs and GPUs via NVLink.

NVLink

For multi-node setups, direct NVLink connections become impractical, requiring NVLink Switch chips, also known as NVSwitch. Over time, these chips evolved into standalone devices.

NVSwitch

NVLink enables the connection of numerous GPUs into a unified logical node—an ultra-scale computing platform. A frequently mentioned setup is DGX GB200 NVL72, utilizing NVLink5 with:

  • 18 Compute Trays (each containing two GB200 Superchips)
  • 9 NVLink Network Switch Trays

Each Compute Tray houses: 2 GB200 Superchips (totaling 36 Grace CPUs and 72 B200 GPUs across the full system).

DGX GB200 NVL72

Additionally, eight DGX GB200 NVL72 units can be combined to create a SuperPod node with 576 GPUs.

This structured product ecosystem defines NVIDIA’s positioning in AI and high-performance computing.

SuperPod node

External Node Interconnection: Scale-Out (IB and Ethernet)

Within a single node, continuously adding GPUs is referred to as Scale-Up (vertical expansion). However, once a node reaches a certain size, further expansion becomes challenging. At this point, increasing the number of nodes and interconnecting them becomes necessary—this is known as Scale-Out (horizontal expansion).

External Node Interconnection

NVIDIA provides solutions for Scale-Out, primarily through InfiniBand (IB) technology. Originally developed by Mellanox, InfiniBand became a proprietary NVIDIA technology after NVIDIA acquired Mellanox in 2019. InfiniBand is a technical term rather than a product name; NVIDIA’s product platform based on InfiniBand is NVIDIA Quantum.

For instance, in March 2024, NVIDIA unveiled the Quantum-X800 network switch platform, which achieves an end-to-end throughput of 800 Gbps. This platform includes hardware such as the Quantum Q3400 switch and ConnectX-8 SuperNIC network cards. These components belong to a broader series—Quantum-X800 follows its predecessor Quantum-2, while ConnectX-8 was preceded by ConnectX-6 and ConnectX-7.

nvidia blackwell platform

The ConnectX high-speed network card also originates from Mellanox. InfiniBand is one of the two primary Scale-Out solutions; the other is Ethernet. NVIDIA has also developed products in this space, notably the Spectrum-X800 platform. The Spectrum-X800 lineup includes Spectrum SN5600 switches and BlueField-3 SuperNIC network cards, with a similarly high 800Gbps throughput.

Spectrum-X800

BlueField has gained significant attention as a Data Processing Unit (DPU). NVIDIA combined Mellanox’s ConnectX network card technology with its own innovations, officially launching the BlueField-2 DPU and BlueField-2X DPU in 2020. Since then, the technology has evolved into BlueField-3.

Furthermore, NVIDIA has recently introduced CPO (Co-Packaged Optics) integrated network switches, including Spectrum-X Photonics and Quantum-X Photonics.

nvidia photonics switch systems

NVIDIA offers additional networking accessories, including various network cards, connectors, and cables, but a detailed discussion of each is beyond the scope here.

Recently, Jensen Huang hinted that the upcoming Rubin platform release would include NVLink 6, ConnectX-9 SuperNIC, and Quantum (Spectrum)-X1600—something worth anticipating.

Development Framework: CUDA

After discussing NVIDIA’s compute hardware platforms and networking solutions, it’s time to explore CUDA, a crucial software component.

While NVIDIA excels in hardware and networking, its CUDA platform is widely regarded as its core competitive advantage. CUDA (Compute Unified Device Architecture) is a parallel computing platform and programming model introduced by NVIDIA in 2006. It enables developers to write code directly for GPUs, significantly accelerating computational performance.

Today, CUDA serves as the operating system for intelligent computing, integrating a programming model, compiler, APIs, libraries, and tools, helping users maximize NVIDIA hardware capabilities.

Beyond being a tool, CUDA fosters a powerful AI development ecosystem, functioning as the central nervous system of NVIDIA’s entire business framework. Many AI development projects heavily rely on NVIDIA’s hardware and CUDA, while switching hardware is relatively straightforward, migrating the ecosystem as a whole presents a much greater challenge.

Compute Unified Device Architecture

Leave a Comment

Scroll to Top