Skip to content

Commit 354c6a4

Browse files
authored
Fix broken links (#271)
1 parent 56258bb commit 354c6a4

13 files changed

+20
-19
lines changed

gpu-operator/amazon-eks.rst

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -102,11 +102,10 @@ without any limitations, you perform the following high-level actions:
102102
the instance type to meet your needs:
103103

104104
* Table of accelerated computing
105-
`instance types <https://aws.amazon.com/ec2/instance-types/#Accelerated_Computing>`_
105+
`instance types <https://aws.amazon.com/ec2/instance-types/accelerated-computing/>`_
106106
for information about GPU model and count, RAM, and storage.
107107

108-
* Table of
109-
`maximum network interfaces <https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/using-eni.html#enis-acceleratedcomputing>`_
108+
* `Maximum IP addresses per network interface <https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/AvailableIpPerENI.html>`_
110109
for accelerated computing instance types.
111110
Make sure the instance type supports enough IP addresses for your workload.
112111
For example, the ``g4dn.xlarge`` instance type supports ``29`` IP addresses for pods on the node.
@@ -132,7 +131,7 @@ Prerequisites
132131
and `Configuring the AWS CLI <https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-configure.html>`_
133132
in the AWS CLI documentation.
134133
* You installed the ``eksctl`` CLI if you prefer it as your client application.
135-
The CLI is available from https://eksctl.io/introduction/#installation.
134+
The CLI is available from https://docs.aws.amazon.com/eks/latest/userguide/install-kubectl.html#eksctl-install-update.
136135
* You have the AMI value from https://cloud-images.ubuntu.com/aws-eks/.
137136
* You have the EC2 instance type to use for your nodes.
138137

gpu-operator/dra-cds.rst

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,8 @@ For more detail on the security properties of a ComputeDomain, see `Security <dr
4949
A deeper dive: related resources
5050
================================
5151

52-
For more background on how ComputeDomains facilitate orchestrating MNNVL workloads on Kubernetes, see `this doc <https://docs.google.com/document/d/1PrdDofsPFVJuZvcv-vtlI9n2eAh-YVf_fRQLIVmDwVY/edit?tab=t.0#heading=h.qkogm924v5so>`_ and `this slide deck <https://docs.google.com/presentation/d/1Xupr8IZVAjs5bNFKJnYaK0LE7QWETnJjkz6KOfLu87E/edit?pli=1&slide=id.g28ac369118f_0_1647#slide=id.g28ac369118f_0_1647>`_.
52+
For more background on how ComputeDomains facilitate orchestrating MNNVL workloads on Kubernetes, refer to the `Kubernetes support for GH200 / GB200 <https://docs.google.com/document/d/1PrdDofsPFVJuZvcv-vtlI9n2eAh-YVf_fRQLIVmDwVY/edit?tab=t.0#heading=h.nfp9friarxam>`_ document
53+
and the `Supporting GB200 on Kubernetes <https://docs.google.com/presentation/d/1Xupr8IZVAjs5bNFKJnYaK0LE7QWETnJjkz6KOfLu87E/edit?pli=1&slide=id.g373e0ebfa8e_1_142#slide=id.g373e0ebfa8e_1_142>`_ slide deck.
5354
For an outlook on planned improvements on the ComputeDomain concept, please refer to `this document <https://github.com/NVIDIA/k8s-dra-driver-gpu/releases/tag/v25.3.0-rc.3>`_.
5455

5556
Details about IMEX and its relationship to NVLink may be found in `NVIDIA's IMEX guide <https://docs.nvidia.com/multi-node-nvlink-systems/imex-guide/overview.html>`_, and in `NVIDIA's NVLink guide <https://docs.nvidia.com/multi-node-nvlink-systems/mnnvl-user-guide/overview.html#internode-memory-exchange-service>`_.

gpu-operator/dra-gpus.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ NVIDIA DRA Driver for GPUs
1212
GPU allocation
1313
**************
1414

15-
Compared to `traditional GPU allocation <https://kubernetes.io/docs/tasks/manage-gpus/scheduling-gpus/#using-device-plugins/>`_ using coarse-grained count-based requests, the GPU allocation side of this driver enables fine-grained control and powerful features long desired by the community, such as:
15+
Compared to `traditional GPU allocation <https://kubernetes.io/docs/tasks/manage-gpus/scheduling-gpus/#using-device-plugins>`_ using coarse-grained count-based requests, the GPU allocation side of this driver enables fine-grained control and powerful features long desired by the community, such as:
1616

1717
#. Controlled sharing of individual GPUs between multiple pods and/or containers.
1818
#. GPU selection via complex constraints expressed via `CEL <https://kubernetes.io/docs/reference/using-api/cel/>`_.

gpu-operator/dra-intro-install.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ Prerequisites
4848
=============
4949

5050
- Kubernetes v1.32 or newer.
51-
- DRA and corresponding API groups must be enabled (`see Kubernetes docs <https://kubernetes.io/docs/concepts/scheduling-eviction/dynamic-resource-allocation/#enabling-dynamic-resource-allocation>`_).
51+
- DRA and corresponding API groups must be enabled (`see Kubernetes docs <https://kubernetes.io/docs/tasks/configure-pod-container/assign-resources/set-up-dra-cluster/#enable-dra>`_).
5252
- `CDI <https://github.com/cncf-tags/container-device-interface?tab=readme-ov-file#how-to-configure-cdi>`_ must be enabled in the underlying container runtime (such as containerd or CRI-O).
5353
- NVIDIA GPU Driver 565 or later.
5454

gpu-operator/getting-started.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -168,7 +168,7 @@ To view all the options, run ``helm show values nvidia/gpu-operator``.
168168
- ``true``
169169

170170
* - ``dcgmExporter.service.internalTrafficPolicy``
171-
- Specifies the `internalTrafficPolicy <https://kubernetes.io/docs/concepts/services-networking/service/#internal-traffic-policy>`_ for the DCGM Exporter service.
171+
- Specifies the `internalTrafficPolicy <https://kubernetes.io/docs/concepts/services-networking/service/#traffic-policies>`_ for the DCGM Exporter service.
172172
Available values are ``Cluster`` (default) or ``Local``.
173173
- ``Cluster``
174174

gpu-operator/gpu-operator-kubevirt.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ Assumptions, constraints, and dependencies
7070

7171
* The GPU Operator will not automate the installation of NVIDIA drivers inside KubeVirt virtual machines with GPUs/vGPUs attached.
7272

73-
* Users must manually add all passthrough GPU and vGPU resources to the ``permittedDevices`` list in the KubeVirt CR before assigning them to KubeVirt virtual machines. Refer to the `KubeVirt documentation <https://kubevirt.io/user-guide/virtual_machines/host-devices/#listing-permitted-devices>`_ for more information.
73+
* Users must manually add all passthrough GPU and vGPU resources to the ``permittedDevices`` list in the KubeVirt CR before assigning them to KubeVirt virtual machines. Refer to the `KubeVirt documentation <https://kubevirt.io/user-guide/compute/host-devices/#listing-permitted-devices>`_ for more information.
7474

7575
* MIG-backed vGPUs are not supported.
7676

@@ -512,7 +512,7 @@ Building the NVIDIA vGPU Manager image
512512

513513
This section covers building the NVIDIA vGPU Manager container image and pushing it to a private registry.
514514

515-
Download the vGPU Software from the `NVIDIA Licensing Portal <https://nvid.nvidia.com/dashboard/#/dashboard>`_.
515+
Download the vGPU Software from the `NVIDIA Licensing Portal <https://stg.ui.licensing.nvidia.com/>`_.
516516

517517
* Login to the NVIDIA Licensing Portal and navigate to the **Software Downloads** section.
518518
* The NVIDIA vGPU Software is located in the **Software Downloads** section of the NVIDIA Licensing Portal.

gpu-operator/gpu-operator-rdma.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -99,7 +99,7 @@ The prerequisites for configuring GPUDirect RDMA or GPUDirect Storage depend on
9999
* ``pciPassthru.64bitMMIOSizeGB = 128``
100100

101101
For information about configuring the settings, refer to the
102-
`Deploy an AI-Ready Enterprise Platform on vSphere 7 <https://core.vmware.com/resource/deploy-ai-ready-vsphere-7#vm-settings-A>`_
102+
`Deploy an AI-Ready Enterprise Platform on vSphere 7 <https://www.vmware.com/docs/deploy-an-ai-ready-enterprise-platform-on-vsphere-7-update-2#vm-settings-A>`_
103103
document from VMWare.
104104

105105
**************************

gpu-operator/install-gpu-operator-nvaie.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -82,7 +82,7 @@ Prerequisites
8282
in the *NVIDIA License System User Guide* for more information.
8383
- An NGC CLI API key that is used to create an image pull secret.
8484
The secret is used to pull the prebuilt vGPU driver image from NVIDIA NGC.
85-
Refer to `Generating Your NGC API Key <https://docs.nvidia.com/ngc/gpu-cloud/ngc-private-registry-user-guide/index.html#generating-api-key>`__
85+
Refer to `Generating Your NGC API Key <https://docs.nvidia.com/ngc/latest/ngc-private-registry-user-guide.html#prug-generating-personal-api-key>`__
8686
in the *NVIDIA NGC Private Registry User Guide* for more information.
8787

8888
Procedure
@@ -179,7 +179,7 @@ The following list summarizes the driver branches for each release.
179179

180180
For newer releases, you can confirm the the supported driver branch by performing the following steps:
181181

182-
#. Refer to the `release documentation <https://docs.nvidia.com/ai-enterprise/#release-documentation>`__
182+
#. Refer to the `NVIDIA AI Enterprise Infra Release Branches <https://docs.nvidia.com/ai-enterprise/#infrastructure-software>`__
183183
for NVIDIA AI Enterprise and access the documentation for your release.
184184

185185
#. In the release notes, identify the supported NVIDIA Data Center GPU Driver branch.

gpu-operator/life-cycle-policy.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -167,7 +167,7 @@ Refer to :ref:`Upgrading the NVIDIA GPU Operator` for more information.
167167
.. note::
168168

169169
- Driver version could be different with NVIDIA vGPU, as it depends on the driver
170-
version downloaded from the `NVIDIA vGPU Software Portal <https://nvid.nvidia.com/dashboard/#/dashboard>`_.
170+
version downloaded from the `NVIDIA Licensing Portal <https://ui.licensing.nvidia.com>`_.
171171
- The GPU Operator is supported on all active NVIDIA data center production drivers.
172-
Refer to `Supported Drivers and CUDA Toolkit Versions <https://docs.nvidia.com/datacenter/tesla/drivers/index.html#cuda-drivers>`_
172+
Refer to `Supported Drivers and CUDA Toolkit Versions <https://docs.nvidia.com/datacenter/tesla/drivers/index.html#supported-drivers-and-cuda-toolkit-versions>`_
173173
for more information.

gpu-operator/microsoft-aks.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -48,8 +48,8 @@ When you follow this approach, you can install the Operator without any special
4848
considerations or arguments.
4949
Refer to :ref:`Install NVIDIA GPU Operator`.
5050

51-
For more information about this preview feature, see
52-
`Skip GPU driver installation (preview) <https://learn.microsoft.com/en-us/azure/aks/gpu-cluster?source=recommendations&tabs=add-ubuntu-gpu-node-pool#skip-gpu-driver-installation-preview>`__
51+
For more information about this feature, see
52+
`Skip GPU driver installation <https://learn.microsoft.com/en-us/azure/aks/use-nvidia-gpu?source=recommendations&tabs=add-ubuntu-gpu-node-pool#skip-gpu-driver-installation>`__
5353
in the Azure Kubernetes Service documentation.
5454

5555

0 commit comments

Comments
 (0)