Replies: 2 comments
-
Hi @marksgraham, could you please take a look at this discussion and share some comments? Thanks. |
Beta Was this translation helpful? Give feedback.
-
Hi @Aman0phy Each level of the UNet halves the spatial dimensions, so if your images are |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I tried running the following two architectures for the DiffusionModelUNet
Model1:
model = DiffusionModelUNet( spatial_dims=2, in_channels=1, out_channels=1, num_channels=(128, 256,512,1024), attention_levels=(False, False,True, True), num_res_blocks=3, num_head_channels=256, )
Model2:
model = DiffusionModelUNet( spatial_dims=2, in_channels=1, out_channels=1, num_channels=(64,128, 256,512,1024), attention_levels=(False,False, True, True,True), num_res_blocks=3, num_head_channels=256, )
With everything else the same, 1 epoch training time for the two models are
Given the depth of Model 2 is more (64,128,256,512,1024) vs Model 1 (128,256,512,1024)
Can anyone please explain why Model 1 occupies more GPU memory and is slower to train ?
Beta Was this translation helpful? Give feedback.
All reactions