-
Notifications
You must be signed in to change notification settings - Fork 15
Open
Description
When running with dataset DIGINETICA, I found this program is really memory demanding, not to speak of YOOCHOOSE. For my case, one TITAN XP is just not enough. Then I was trying to apply multi graph cards to cuda and I just failed to do so.
I try to use Dataparallel from torch:
os.environ['CUDA_VISIBLE_DEVICES'] = '0,1'
model = torch.nn.Dataparallel(SessionGraph())
model = model.cuda()
Since model is now a DataParallel object the parameters in train_test is modified accordingly
model.loss_function → model.module.loss_function
...
And that didn't work. I still got runtime error about not enough memory in GPU0 while GPU1 was barely utilized. So I want to know how you guys done it. Great thanks in advance!
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels