-
Notifications
You must be signed in to change notification settings - Fork 3
Debug performance #108
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Debug performance #108
Conversation
Vesicle SegmentationTrying to debug the problems with vesicle segmentation performance reported by @SarahMuth . To reproduce:
srun -p grete:shared -G A100:1 --pty bash
synapse_net.run_segmentation -i /mnt/lustre-emmy-hdd/projects/nim00007/data/synaptic-reconstruction/cooper/original_imod_data/20240909_cp_datatransfer/10_tem_single_release/TEM_250nm_Culture5_SingleRelease/ -o out -m vesicles_3dThe segmentation runs in ca. 36 minutes (= ca. 2:15 min / tomogram): For a single tomogram, the individual steps take this long:
|
Mito SegmentationTrying to debug the problems / out of memory error in mito segmentation reported by @lufre1 . To reproduce:
srun -p grete:shared -G A100:1 --pty bash
synapse_net.run_segmentation -i /scratch-grete/projects/nim00007/data/mitochondria/cooper/20250212_test_I_h5_s2 -o out -m mitochondria --tile_shape 48 512 512 --halo 4 128 128 --verbose -s raw --data_ext .h5The segmentation runs in ca. 6:30 minutes (= ca 1:10 min / tomogram) without further issues: For an individual tomogram, the individual steps take this long:
(Note that the h5s don't contain voxel sizes, so the resizing step is skipped) |
Mito Segmentation@constantinpape The problem occurs only with .mrc files. The .h5 files are downscaled by a factor of two and i created those to cirumvent the problem. what i used: and then (pulled some mrc files from fidi we can use) and that is the output: EDITwith no mem specified: the output: and with 128GB mem specified: output: and with 256 mem specified: the ouput: there might still be issues with the environemnt - i will have a look into that... |
Why are you setting it to 32G ? That's tiny. Either set to at least 128G (or better 256G) or don't constraint it (not sure what the default is, but for me everything worked with it.) |
i tested with no mem specified: then i found 30GB of RAM: and got: |
This is because you specify But you can also just ask for more memory (128G or 256G) and the problem should be solved. |
with this srun command i get: ok, with this configuration I do not run into any issues |
|
Thanks for reporting @lufre1 . So it's consistent that this issue is due to memory. I will go ahead and merge this issue, as we don't need to address anything else. If you find a volume where we run into issues even with 256GB let me know, and then we can check how to fix memory for it. |
Trying to reproduce / debug the performance issues reported by @SarahMuth and @lufre1 .