Replies: 2 comments 2 replies
-
|
Hey. I'm not a developer of pyg, but I submitted a very similar question a while ago. You should take a look on sampling techniques. Pyg has some implemented sampling interfaces in the sampler module. If none of them are exactly what you need, you can always implement it yourself using the BaseSampler abstract class. Hope this helps. |
Beta Was this translation helpful? Give feedback.
-
|
Hello All, Thank you for this implementation ! I wish I could have a distributed training like that during my thesis. If only it only existed already a few years ago it could have spared me some headaches! I tried to read the documentations and I am not sure whether it exists communication between the partitions ? I saw in the partitioning there were local and remote nodes automatically created but I am not sure if for instance some communication were made between the remotes nodes for every partitions ? I wish like for CFD applications, your partitions can send their information to their neighbors, like the ghost cells/nodes so in the end we will have a smooth field without discontinuities between the partitions. Thank you in advance Edit: Sorry, I found it with |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
Thank you for this excellent library! I'm working with a large graph dataset that exceeds available memory. While the distributed training documentation (https://pytorch-geometric.readthedocs.io/en/latest/tutorial/distributed_pyg.html) is helpful, I have a question about dataset partitioning:
Since the documentation indicates partitioning occurs on in-memory data, how can I effectively leverage the distributed training system when my entire dataset doesn't fit into memory? Are there recommended strategies or examples for this scenario?
Beta Was this translation helpful? Give feedback.
All reactions