Replies: 1 comment 2 replies
-
Hey. I'm not a developer of pyg, but I submitted a very similar question a while ago. You should take a look on sampling techniques. Pyg has some implemented sampling interfaces in the sampler module. If none of them are exactly what you need, you can always implement it yourself using the BaseSampler abstract class. Hope this helps. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
Thank you for this excellent library! I'm working with a large graph dataset that exceeds available memory. While the distributed training documentation (https://pytorch-geometric.readthedocs.io/en/latest/tutorial/distributed_pyg.html) is helpful, I have a question about dataset partitioning:
Since the documentation indicates partitioning occurs on in-memory data, how can I effectively leverage the distributed training system when my entire dataset doesn't fit into memory? Are there recommended strategies or examples for this scenario?
Beta Was this translation helpful? Give feedback.
All reactions