|
| 1 | +--- |
| 2 | +date: 2026-01-16 |
| 3 | +title: So You Need Enterprise GPUs - A No-BS Guide to H100, H200, and B200 |
| 4 | +authors: |
| 5 | + - bet0x |
| 6 | +description: > |
| 7 | + Confused about which NVIDIA GPU to pick for your AI workloads? H100 SXM vs NVL, H200, B200... the naming is a mess. Here's what actually matters when you're spending serious money on compute. |
| 8 | +categories: |
| 9 | + - GPU |
| 10 | + - Infrastructure |
| 11 | + - Hardware |
| 12 | + |
| 13 | +--- |
| 14 | + |
| 15 | +# So You Need Enterprise GPUs: A No-BS Guide to H100, H200, and B200 |
| 16 | + |
| 17 | +Let's be honest—NVIDIA's naming conventions are designed to confuse procurement teams. H100 SXM5, H100 NVL, H200 SXM, B200... it sounds like someone spilled alphabet soup on a product roadmap. |
| 18 | + |
| 19 | +I've spent way too many hours explaining these differences to engineering teams, so here's everything you actually need to know before signing that hardware purchase order. |
| 20 | + |
| 21 | +<!-- more --> |
| 22 | + |
| 23 | +## The TL;DR Table |
| 24 | + |
| 25 | +Before we dive in, here's the cheat sheet: |
| 26 | + |
| 27 | +| Spec | H100 SXM | H100 NVL | H200 SXM | B200 | |
| 28 | +|------|----------|----------|----------|------| |
| 29 | +| **Memory** | 80GB HBM3 | 94GB × 2 (188GB) | 141GB HBM3e | 192GB HBM3e | |
| 30 | +| **Bandwidth** | 3.35 TB/s | 3.9 TB/s × 2 | 4.8 TB/s | 8.0 TB/s | |
| 31 | +| **NVLink** | 900 GB/s | 600 GB/s | 900 GB/s | 1.8 TB/s | |
| 32 | +| **TDP** | 700W | 400W × 2 | 700W | 1000W | |
| 33 | +| **Sweet Spot** | Training | Inference | Both | Next-gen training | |
| 34 | + |
| 35 | +Now let's talk about *why* these numbers matter. |
| 36 | + |
| 37 | +## H100: Still the Workhorse |
| 38 | + |
| 39 | +The H100 has been the backbone of AI infrastructure since 2023. Two variants, completely different use cases. |
| 40 | + |
| 41 | +### H100 SXM5 - The Training Beast |
| 42 | + |
| 43 | +This is what goes into DGX systems and large-scale training clusters. The SXM form factor means it slots directly into specialized motherboards—no PCIe bus bottlenecks. |
| 44 | + |
| 45 | +**Why it matters:** |
| 46 | + |
| 47 | +- 900 GB/s NVLink bandwidth lets you scale to 8 GPUs without choking on data transfer |
| 48 | +- The 700W TDP is manageable in purpose-built data centers |
| 49 | +- 80GB is tight for modern LLMs, but fine for training with gradient checkpointing |
| 50 | + |
| 51 | +The SXM5 assumes you're building dedicated AI infrastructure. If you're thinking "we'll just add it to our existing servers," stop. This isn't that kind of hardware. |
| 52 | + |
| 53 | +### H100 NVL - The Inference Play |
| 54 | + |
| 55 | +Here's where things get interesting. The NVL is two H100s on a single board, connected via NVLink bridge. You get 188GB of combined memory in a PCIe form factor. |
| 56 | + |
| 57 | +**The real story:** |
| 58 | + |
| 59 | +That 188GB memory pool is the selling point. A 70B parameter model in FP16 needs ~140GB just for weights. Single H100 SXM? Won't fit. H100 NVL? No problem. |
| 60 | + |
| 61 | +But here's what the spec sheets don't tell you: the NVLink bridge between the two dies gives you 600 GB/s, not the 900 GB/s you get with SXM. For inference, this rarely matters. For training, it's a bottleneck. |
| 62 | + |
| 63 | +**When to pick NVL:** |
| 64 | + |
| 65 | +- Running inference on 70B+ models |
| 66 | +- Your facility isn't built for 10KW+ racks |
| 67 | +- You want to deploy in standard PCIe servers |
| 68 | + |
| 69 | +## H200: The Memory Upgrade |
| 70 | + |
| 71 | +The H200 is basically "what if we put better memory on an H100?" Same Hopper architecture, but with HBM3e instead of HBM3. |
| 72 | + |
| 73 | +**What changed:** |
| 74 | + |
| 75 | +- 141GB vs 80GB memory |
| 76 | +- 4.8 TB/s vs 3.35 TB/s bandwidth |
| 77 | +- Everything else? Virtually identical |
| 78 | + |
| 79 | +That 76% memory increase is significant. Models that needed 2× H100s now fit on a single H200. The bandwidth bump helps with memory-bound workloads—which is basically everything in inference. |
| 80 | + |
| 81 | +**The catch:** |
| 82 | + |
| 83 | +H200 availability has been... challenging. If you can actually get them, they're a meaningful upgrade. But don't hold up your project waiting. |
| 84 | + |
| 85 | +## B200: Welcome to Blackwell |
| 86 | + |
| 87 | +Blackwell is a generational leap, not an incremental upgrade. The B200 rewrites what's possible. |
| 88 | + |
| 89 | +**The headline numbers:** |
| 90 | + |
| 91 | +- 192GB HBM3e |
| 92 | +- 8.0 TB/s memory bandwidth (2.4× the H100) |
| 93 | +- 1.8 TB/s NVLink 5 (2× the H100) |
| 94 | +- FP4 tensor support (new) |
| 95 | + |
| 96 | +Fifth-generation tensor cores with dual transformer engines. In MLPerf benchmarks, B200 hits roughly 2× the performance of H200 on LLM training. That's not marketing fluff—it's measured. |
| 97 | + |
| 98 | +**The infrastructure problem:** |
| 99 | + |
| 100 | +B200 draws 1000W per GPU. An 8-GPU node needs 8KW+ just for compute, before you count networking, storage, and cooling. Your existing data center probably can't handle it without upgrades. |
| 101 | + |
| 102 | +Also: CUDA 12.4+ required. If you're running older software stacks, expect a migration project. |
| 103 | + |
| 104 | +## The Memory Bandwidth Story |
| 105 | + |
| 106 | +Here's the thing nobody talks about enough: memory bandwidth is often the real bottleneck, not compute. |
| 107 | + |
| 108 | +``` |
| 109 | +H100 SXM: 3.35 TB/s |
| 110 | +H100 NVL: 3.9 TB/s (per die) |
| 111 | +H200: 4.8 TB/s |
| 112 | +B200: 8.0 TB/s |
| 113 | +``` |
| 114 | + |
| 115 | +During inference, you're constantly shuffling weights and KV cache in and out of HBM. More bandwidth = more tokens per second, period. The B200's 8 TB/s is why it demolishes everything else in inference benchmarks despite "only" having 2× the raw compute. |
| 116 | + |
| 117 | +## NVLink: When Interconnect Matters |
| 118 | + |
| 119 | +NVLink bandwidth determines how well multi-GPU setups scale: |
| 120 | + |
| 121 | +``` |
| 122 | +H100/H200: 900 GB/s (NVLink 4) |
| 123 | +B200: 1.8 TB/s (NVLink 5) |
| 124 | +``` |
| 125 | + |
| 126 | +For single-GPU inference, this is irrelevant. For 8-GPU training runs, it's everything. |
| 127 | + |
| 128 | +Tensor parallelism means shipping activation tensors between GPUs every single layer. Slow interconnect = GPUs waiting on data = expensive hardware sitting idle. |
| 129 | + |
| 130 | +The jump from 900 GB/s to 1.8 TB/s isn't just "better"—it enables training paradigms that were impractical before. |
| 131 | + |
| 132 | +## What Should You Actually Buy? |
| 133 | + |
| 134 | +Cut through the noise: |
| 135 | + |
| 136 | +**"We're deploying inference for a 70B model"** |
| 137 | +→ H100 NVL or wait for H200 availability. The memory is what matters. |
| 138 | + |
| 139 | +**"We're training models up to 30B parameters"** |
| 140 | +→ H100 SXM is still cost-effective if you can get good pricing. 8-GPU nodes work well. |
| 141 | + |
| 142 | +**"We're training 100B+ parameter models"** |
| 143 | +→ H200 if available, B200 if you can wait and your facility supports the power draw. |
| 144 | + |
| 145 | +**"We need the absolute best performance and budget isn't the constraint"** |
| 146 | +→ B200. Not close. |
| 147 | + |
| 148 | +**"We're a startup figuring things out"** |
| 149 | +→ Rent H100 SXM time from cloud providers. Don't buy hardware until you know your workload. |
| 150 | + |
| 151 | +## The Hidden Costs |
| 152 | + |
| 153 | +Everyone focuses on GPU price. The real costs: |
| 154 | + |
| 155 | +1. **Power infrastructure**: B200 nodes need serious electrical work. Budget $50-100K per rack for upgrades. |
| 156 | + |
| 157 | +2. **Cooling**: 8KW+ of heat per node. Traditional air cooling won't cut it for dense deployments. |
| 158 | + |
| 159 | +3. **Networking**: You need 400GbE minimum between nodes for distributed training. That's $10K+ per port. |
| 160 | + |
| 161 | +4. **Software stack**: Blackwell needs CUDA 12.4+. Older frameworks may need updates. |
| 162 | + |
| 163 | +5. **Lead times**: Enterprise GPU lead times are still measured in months, not weeks. |
| 164 | + |
| 165 | +## Final Thoughts |
| 166 | + |
| 167 | +The GPU landscape moves fast. By the time you read this, there might be new variants I haven't covered. But the fundamentals stay the same: |
| 168 | + |
| 169 | +- Memory size determines what models fit |
| 170 | +- Memory bandwidth determines inference speed |
| 171 | +- NVLink bandwidth determines training scale |
| 172 | +- TDP determines infrastructure requirements |
| 173 | + |
| 174 | +Match your workload to these constraints, not to marketing materials. |
| 175 | + |
| 176 | +And if someone tells you "just get the newest thing"—ask them if they've actually priced out the infrastructure upgrades. Hardware is the easy part. Everything around it is where projects die. |
| 177 | + |
| 178 | +--- |
| 179 | + |
| 180 | +*Have questions about GPU selection for your specific workload? That's literally what we do at Rackspace Private Cloud AI. [Reach out](https://www.rackspace.com/cloud/private/ai) and let's talk specifics.* |
0 commit comments