For the best performance with NVIDIA GPUs, we recommend using cuGraph. Refer to our installation guide for setup instructions and to the cuGraph-PyG examples for ready-to-run training scripts covering single-node, multi-node, and link-prediction workloads.
| Example | Scalability | Description |
|---|---|---|
distributed_batching.py |
single-node | Graph-level prediction on many small graphs (ogbg-molhiv) using DataLoader with DistributedSampler. |
distributed_sampling.py |
single-node | Node-level classification on a single large graph (Reddit) using NeighborLoader for multi-hop subgraph sampling. |
distributed_sampling_multinode.py |
multi-node | Training GNNs on a homogeneous graph with neighbor sampling on multiple nodes. |
distributed_sampling_multinode.sbatch |
multi-node | Submitting a training job to a Slurm cluster using distributed_sampling_multinode.py. |
papers100m_gcn.py |
single-node | Training GNNs on the ogbn-papers100M homogeneous graph w/ ~1.6B edges. |
papers100m_gcn_multinode.py |
multi-node | Training GNNs on a homogeneous graph on multiple nodes. |
pcqm4m_ogb.py |
single-node | Training GNNs for a graph-level regression task. |
mag240m_graphsage.py |
single-node | Training GNNs on a large heterogeneous graph. |
taobao.py |
single-node | Training link prediction GNNs on a heterogeneous graph. |
model_parallel.py |
single-node | Model parallelism by manually placing layers on each GPU. |
| Example | Scalability | Description |
|---|---|---|
distributed_sampling_xpu.py |
single-node, multi-gpu | Training GNNs on a homogeneous graph with neighbor sampling. |