Top Podcasts
Health & Wellness
Personal Growth
Social & Politics
Technology
AI
Personal Finance
Crypto
Explainers
YouTube SummarySee all latest Top Podcasts summaries
Watch on YouTube
Publisher thumbnail
Linus Tech Tips
23:049/25/25

Nvidia Wouldn't Send Me This Graphics Card - H200 Holy $H!T

TLDR

Nvidia's H200 GPU, priced at $30,000, is a powerhouse for AI training and deep learning due to its massive HBM3E memory and specialized architecture, significantly outperforming consumer GPUs and even high-end CPUs in these tasks.

Takeways

Nvidia's $30,000 H200 GPU is purpose-built for AI, not graphics, featuring 141GB of HBM3E memory and 4.8 TB/s bandwidth.

The H200 vastly outperforms high-end server CPUs and consumer GPUs like the RTX 5090 in AI inference and training, especially with large language models.

HBM memory's density and on-package integration contribute to the H200's high cost and exceptional power efficiency in large-scale AI deployments.

Nvidia's H200 NVL, an enterprise-grade GPU costing $30,000, was rigorously tested to showcase its capabilities in AI applications. Despite its low clock speeds and lack of graphics processing support, its 141GB of HBM3E memory and specialized architecture make it vastly superior for AI inference and especially training compared to consumer GPUs like the RTX 5090 or high-end server CPUs, emphasizing the importance of memory bandwidth and capacity in large-scale AI deployments.

H200 Hardware Overview

00:01:32 The H200 NVL is a PCIe version featuring an 80 billion transistor Nvidia GH100 die and 141GB of HBM3E memory on a 6,144-bit bus, delivering 4.8 terabytes per second of bandwidth. Built on TSMC's 5-nanometer node, it has nearly 17,000 shading units and over 500 tensor cores, but it lacks support for standard graphics APIs like Vulkan or DirectX, indicating its sole focus on compute tasks. Its dense design includes a large vapor chamber cooler that relies on server fans for airflow, drawing 600 Watts of power, primarily due to its vast amount of high-bandwidth memory.

CPU vs. H200 in AI

00:08:24 In a direct comparison, a single H200 GPU significantly outclasses a dual AMD EPYC 9965 server with 400 CPU cores and 1.7 terabytes of RAM in AI inference tasks, specifically with a 120 billion parameter GPT-OSS model. The H200 achieved 122 tokens per second, while the CPU setup struggled, producing only 21 tokens per second. This demonstrates that GPUs have effectively displaced CPUs for AI workloads due to their specialized architecture and parallel processing capabilities, despite the CPU setup's substantial cost and resources.

H200 vs. RTX 5090

00:16:00 While the H200 is an enterprise card, it is significantly outmatched by a consumer RTX 5090 in traditional rendering applications like Blender. However, for AI inference using large language models, the H200 excels due to its immense memory capacity and bandwidth, allowing it to load and process larger models without overflowing to slower system memory. The RTX 5090, with less VRAM, caps out and becomes considerably slower when its memory is insufficient for the model size.

H200 in AI Training

00:19:09 The H200's true strength lies in AI training, particularly for large language models, where it leverages its vast HBM3E memory for efficient Low-Rank Adaptation (LoRa) fine-tuning. In training tasks like Tiny Llama and 8 billion parameter models, the H200 NVL performs substantially faster than the RTX 5090, which struggles to fine-tune larger models at full precision due to VRAM limitations. The efficiency gained by the H200 in large-scale training operations translates into significant cost savings in power and cooling for deployments with thousands of GPUs.