Replies: 2 comments
-
Hi @marksgraham, could you please take a look at this discussion and share some comments? Thanks. |
Beta Was this translation helpful? Give feedback.
-
Hi @Aman0phy Each level of the UNet halves the spatial dimensions, so if your images are |
Beta Was this translation helpful? Give feedback.
-
I tried running the following two architectures for the DiffusionModelUNet
Model1:
model = DiffusionModelUNet( spatial_dims=2, in_channels=1, out_channels=1, num_channels=(128, 256,512,1024), attention_levels=(False, False,True, True), num_res_blocks=3, num_head_channels=256, )
Model2:
model = DiffusionModelUNet( spatial_dims=2, in_channels=1, out_channels=1, num_channels=(64,128, 256,512,1024), attention_levels=(False,False, True, True,True), num_res_blocks=3, num_head_channels=256, )
With everything else the same, 1 epoch training time for the two models are
Given the depth of Model 2 is more (64,128,256,512,1024) vs Model 1 (128,256,512,1024)
Can anyone please explain why Model 1 occupies more GPU memory and is slower to train ?
Beta Was this translation helpful? Give feedback.
All reactions