Skip to content

Latest commit

 

History

History

README.md

Examples for Distributed Training

Examples with NVIDIA GPUs

Examples with cuGraph

For the best performance with NVIDIA GPUs, we recommend using cuGraph. Refer to our installation guide for setup instructions and to the cuGraph-PyG examples for ready-to-run training scripts covering single-node, multi-node, and link-prediction workloads.

Examples with Pure PyTorch

Example Scalability Description
distributed_batching.py single-node Graph-level prediction on many small graphs (ogbg-molhiv) using DataLoader with DistributedSampler.
distributed_sampling.py single-node Node-level classification on a single large graph (Reddit) using NeighborLoader for multi-hop subgraph sampling.
distributed_sampling_multinode.py multi-node Training GNNs on a homogeneous graph with neighbor sampling on multiple nodes.
distributed_sampling_multinode.sbatch multi-node Submitting a training job to a Slurm cluster using distributed_sampling_multinode.py.
papers100m_gcn.py single-node Training GNNs on the ogbn-papers100M homogeneous graph w/ ~1.6B edges.
papers100m_gcn_multinode.py multi-node Training GNNs on a homogeneous graph on multiple nodes.
pcqm4m_ogb.py single-node Training GNNs for a graph-level regression task.
mag240m_graphsage.py single-node Training GNNs on a large heterogeneous graph.
taobao.py single-node Training link prediction GNNs on a heterogeneous graph.
model_parallel.py single-node Model parallelism by manually placing layers on each GPU.

Examples with Intel GPUs (XPUs)

Example Scalability Description
distributed_sampling_xpu.py single-node, multi-gpu Training GNNs on a homogeneous graph with neighbor sampling.