diff --git a/pages/gpu/reference-content/choosing-gpu-instance-type.mdx b/pages/gpu/reference-content/choosing-gpu-instance-type.mdx
index e173a1695e..cd2c3de64d 100644
--- a/pages/gpu/reference-content/choosing-gpu-instance-type.mdx
+++ b/pages/gpu/reference-content/choosing-gpu-instance-type.mdx
@@ -126,17 +126,3 @@ Remember that there is no one-size-fits-all answer, and the right GPU Instance t
| Storage | 64TB of a3i DDN low latency storage | 1.8 PB of a3i DDN low latency storage |
| [MIG compatibility](/gpu/how-to/use-nvidia-mig-technology/) | Yes | Yes |
| Inter-GPU bandwidth | Infiniband 400 Gb/s | Infiniband 400 Gb/s |
-
-### NVIDIA GH200 Superchip
-
-| | **[GH200 Grace Hopper™](https://www.scaleway.com/en/contact-gh200/)** |
-|---------------------------------------------|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
-| GPU Type | NVIDIA [GH200 Grace Hopper™ Superchip](https://www.nvidia.com/en-us/data-center/grace-hopper-superchip/) |
-| NVIDIA architecture | GH200 Grace Hopper™ Architecture |
-| Performance | 990 TFLops (in FP166 Tensor Core) |
-| Specifications | - GH200 SuperChip with 72 ARM Neoverse V2 cores
- 480 GB of LPDDR5X DRAM
- 96GB of HBM3 GPU memory
(Memory is fully merged for up to 576GB of global usable memory) |
-| [MIG compatibility](/gpu/how-to/use-nvidia-mig-technology/) | Yes |
-| Inter-GPU bandwidth (for clusters up to 256 GH200) | NVlink Switch System 900 GB/s |
-| Format & Features | Single chip up to GH200 clusters. (For larger setup needs, [contact us](https://www.scaleway.com/en/contact-ai-supercomputers/)) |
-| Use cases | - Extra large LLM and DL model inference
- HPC |
-| What they are not made for | - Graphism
- (Training) |
diff --git a/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx b/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx
index b108b8cea1..f3bf24c443 100644
--- a/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx
+++ b/pages/gpu/reference-content/understanding-nvidia-nvlink.mdx
@@ -7,7 +7,7 @@ dates:
posted: 2025-03-13
---
-NVLink is NVIDIA's high-bandwidth, low-latency GPU-to-GPU interconnect with built-in resiliency features, available on Scaleway's [H100-SGX Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table). It was designed to significantly improve the performance and efficiency when connecting GPUs, CPUs, and other components within the same node.
+NVLink is NVIDIA's high-bandwidth, low-latency GPU-to-GPU interconnect with built-in resiliency features, available on Scaleway's [H100-SXM Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table). It was designed to significantly improve the performance and efficiency when connecting GPUs, CPUs, and other components within the same node.
It provides much higher bandwidth (up to 900 GB/s total GPU-to-GPU bandwidth in an 8-GPU configuration) and lower latency compared to traditional PCIe Gen 4 (up to 32 GB/s per link).
This allows more data to be transferred between GPUs in less time while also reducing latency.
@@ -21,7 +21,7 @@ Unified Memory Access allows GPUs to access each other's memory directly without
### Comparison: NVLink vs. PCIe
NVLink and PCI Express (PCIe) are both used for GPU communication, but NVLink is specifically designed to address the bandwidth and latency bottlenecks of PCIe in multi-GPU setups.
-| Feature | NVLink 4.0 (H100-SGX) | PCIe 5.0 |
+| Feature | NVLink 4.0 (H100-SXM) | PCIe 5.0 |
|-------------------|---------------------------|------------------------------------|
| **Use case** | High-performance computing, deep learning | General-purpose computing, graphics |
| **Bandwidth** | Up to 900 GB/s (aggregate, multi-GPU) | 128 GB/s (x16 bidirectional) |
@@ -31,4 +31,4 @@ NVLink and PCI Express (PCIe) are both used for GPU communication, but NVLink is
| **Scalability** | Multi-GPU direct connection via NVSwitch | Limited by PCIe lanes |
| **Efficiency** | Optimized for GPU workloads | More general-purpose |
-In summary, NVLink, available on [H100-SGX Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table), is **superior** for **multi-GPU AI and HPC** workloads due to its **higher bandwidth, lower latency, and memory-sharing capabilities**, while PCIe remains essential for broader system connectivity and general computing.
+In summary, NVLink, available on [H100-SXM Instances](/gpu/reference-content/choosing-gpu-instance-type/#gpu-instances-and-ai-supercomputer-comparison-table), is **superior** for **multi-GPU AI and HPC** workloads due to its **higher bandwidth, lower latency, and memory-sharing capabilities**, while PCIe remains essential for broader system connectivity and general computing.