diff --git a/pages/gpu/reference-content/choosing-gpu-instance-type.mdx b/pages/gpu/reference-content/choosing-gpu-instance-type.mdx index e173a1695e..cd2c3de64d 100644 --- a/pages/gpu/reference-content/choosing-gpu-instance-type.mdx +++ b/pages/gpu/reference-content/choosing-gpu-instance-type.mdx @@ -126,17 +126,3 @@ Remember that there is no one-size-fits-all answer, and the right GPU Instance t | Storage | 64TB of a3i DDN low latency storage | 1.8 PB of a3i DDN low latency storage | | [MIG compatibility](/gpu/how-to/use-nvidia-mig-technology/) | Yes | Yes | | Inter-GPU bandwidth | Infiniband 400 Gb/s | Infiniband 400 Gb/s | - -### NVIDIA GH200 Superchip - -| | **[GH200 Grace Hopper™](https://www.scaleway.com/en/contact-gh200/)** | -|---------------------------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| -| GPU Type | NVIDIA [GH200 Grace Hopper™ Superchip](https://www.nvidia.com/en-us/data-center/grace-hopper-superchip/) | -| NVIDIA architecture | GH200 Grace Hopper™ Architecture | -| Performance | 990 TFLops (in FP166 Tensor Core) | -| Specifications | - GH200 SuperChip with 72 ARM Neoverse V2 cores
- 480 GB of LPDDR5X DRAM
- 96GB of HBM3 GPU memory
(Memory is fully merged for up to 576GB of global usable memory) | -| [MIG compatibility](/gpu/how-to/use-nvidia-mig-technology/) | Yes | -| Inter-GPU bandwidth (for clusters up to 256 GH200) | NVlink Switch System 900 GB/s | -| Format & Features | Single chip up to GH200 clusters. (For larger setup needs, [contact us](https://www.scaleway.com/en/contact-ai-supercomputers/)) | -| Use cases | - Extra large LLM and DL model inference
- HPC | -| What they are not made for | - Graphism
- (Training) | diff --git a/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx b/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx index b108b8cea1..f3bf24c443 100644 --- a/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx +++ b/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx @@ -7,7 +7,7 @@ dates: posted: 2025-03-13 --- -NVLink is NVIDIA's high-bandwidth, low-latency GPU-to-GPU interconnect with built-in resiliency features, available on Scaleway's [H100-SGX Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table). It was designed to significantly improve the performance and efficiency when connecting GPUs, CPUs, and other components within the same node. +NVLink is NVIDIA's high-bandwidth, low-latency GPU-to-GPU interconnect with built-in resiliency features, available on Scaleway's [H100-SXM Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table). It was designed to significantly improve the performance and efficiency when connecting GPUs, CPUs, and other components within the same node. It provides much higher bandwidth (up to 900 GB/s total GPU-to-GPU bandwidth in an 8-GPU configuration) and lower latency compared to traditional PCIe Gen 4 (up to 32 GB/s per link). This allows more data to be transferred between GPUs in less time while also reducing latency. @@ -21,7 +21,7 @@ Unified Memory Access allows GPUs to access each other's memory directly without ### Comparison: NVLink vs. PCIe NVLink and PCI Express (PCIe) are both used for GPU communication, but NVLink is specifically designed to address the bandwidth and latency bottlenecks of PCIe in multi-GPU setups. -| Feature | NVLink 4.0 (H100-SGX) | PCIe 5.0 | +| Feature | NVLink 4.0 (H100-SXM) | PCIe 5.0 | |-------------------|---------------------------|------------------------------------| | **Use case** | High-performance computing, deep learning | General-purpose computing, graphics | | **Bandwidth** | Up to 900 GB/s (aggregate, multi-GPU) | 128 GB/s (x16 bidirectional) | @@ -31,4 +31,4 @@ NVLink and PCI Express (PCIe) are both used for GPU communication, but NVLink is | **Scalability** | Multi-GPU direct connection via NVSwitch | Limited by PCIe lanes | | **Efficiency** | Optimized for GPU workloads | More general-purpose | -In summary, NVLink, available on [H100-SGX Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table), is **superior** for **multi-GPU AI and HPC** workloads due to its **higher bandwidth, lower latency, and memory-sharing capabilities**, while PCIe remains essential for broader system connectivity and general computing. +In summary, NVLink, available on [H100-SXM Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table), is **superior** for **multi-GPU AI and HPC** workloads due to its **higher bandwidth, lower latency, and memory-sharing capabilities**, while PCIe remains essential for broader system connectivity and general computing.