Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 0 additions & 14 deletions pages/gpu/reference-content/choosing-gpu-instance-type.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -126,17 +126,3 @@ Remember that there is no one-size-fits-all answer, and the right GPU Instance t
| Storage | 64TB of a3i DDN low latency storage | 1.8 PB of a3i DDN low latency storage |
| [MIG compatibility](/gpu/how-to/use-nvidia-mig-technology/) | Yes | Yes |
| Inter-GPU bandwidth | Infiniband 400 Gb/s | Infiniband 400 Gb/s |

### NVIDIA GH200 Superchip

| | **[GH200 Grace Hopper™](https://www.scaleway.com/en/contact-gh200/)** |
|---------------------------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| GPU Type | NVIDIA [GH200 Grace Hopper™ Superchip](https://www.nvidia.com/en-us/data-center/grace-hopper-superchip/) |
| NVIDIA architecture | GH200 Grace Hopper™ Architecture |
| Performance | 990 TFLops (in FP166 Tensor Core) |
| Specifications | - GH200 SuperChip with 72 ARM Neoverse V2 cores<br />- 480 GB of LPDDR5X DRAM<br />- 96GB of HBM3 GPU memory<br />(Memory is fully merged for up to 576GB of global usable memory) |
| [MIG compatibility](/gpu/how-to/use-nvidia-mig-technology/) | Yes |
| Inter-GPU bandwidth (for clusters up to 256 GH200) | NVlink Switch System 900 GB/s |
| Format & Features | Single chip up to GH200 clusters. (For larger setup needs, [contact us](https://www.scaleway.com/en/contact-ai-supercomputers/)) |
| Use cases | - Extra large LLM and DL model inference<br />- HPC |
| What they are not made for | - Graphism<br /> - (Training) |
6 changes: 3 additions & 3 deletions pages/gpu/reference-content/understanding-nvidia-nvlink.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ dates:
posted: 2025-03-13
---

NVLink is NVIDIA's high-bandwidth, low-latency GPU-to-GPU interconnect with built-in resiliency features, available on Scaleway's [H100-SGX Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table). It was designed to significantly improve the performance and efficiency when connecting GPUs, CPUs, and other components within the same node.
NVLink is NVIDIA's high-bandwidth, low-latency GPU-to-GPU interconnect with built-in resiliency features, available on Scaleway's [H100-SXM Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table). It was designed to significantly improve the performance and efficiency when connecting GPUs, CPUs, and other components within the same node.
It provides much higher bandwidth (up to 900 GB/s total GPU-to-GPU bandwidth in an 8-GPU configuration) and lower latency compared to traditional PCIe Gen 4 (up to 32 GB/s per link).
This allows more data to be transferred between GPUs in less time while also reducing latency.

Expand All @@ -21,7 +21,7 @@ Unified Memory Access allows GPUs to access each other's memory directly without
### Comparison: NVLink vs. PCIe
NVLink and PCI Express (PCIe) are both used for GPU communication, but NVLink is specifically designed to address the bandwidth and latency bottlenecks of PCIe in multi-GPU setups.

| Feature | NVLink 4.0 (H100-SGX) | PCIe 5.0 |
| Feature | NVLink 4.0 (H100-SXM) | PCIe 5.0 |
|-------------------|---------------------------|------------------------------------|
| **Use case** | High-performance computing, deep learning | General-purpose computing, graphics |
| **Bandwidth** | Up to 900 GB/s (aggregate, multi-GPU) | 128 GB/s (x16 bidirectional) |
Expand All @@ -31,4 +31,4 @@ NVLink and PCI Express (PCIe) are both used for GPU communication, but NVLink is
| **Scalability** | Multi-GPU direct connection via NVSwitch | Limited by PCIe lanes |
| **Efficiency** | Optimized for GPU workloads | More general-purpose |

In summary, NVLink, available on [H100-SGX Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table), is **superior** for **multi-GPU AI and HPC** workloads due to its **higher bandwidth, lower latency, and memory-sharing capabilities**, while PCIe remains essential for broader system connectivity and general computing.
In summary, NVLink, available on [H100-SXM Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table), is **superior** for **multi-GPU AI and HPC** workloads due to its **higher bandwidth, lower latency, and memory-sharing capabilities**, while PCIe remains essential for broader system connectivity and general computing.
Loading