Hi,
I am obtaining an OOM error when using ndmeasure.label
with a large array.
E.g.,
nx = 5120 # things are OK < 2500
arr = da.random.random(size=(nx,nx,nx))
darr_bin = arr > 0.8
# The next line will fail
label_image, num_labels = dask_image.ndmeasure.label(darr_bin)
Note that this problem occurs already in the last line, and not when executing the computation via, e.g., num_labels.compute()
.
This also means that I have the same problem when using a (large) cluster as the OOM always occurs on node 1.
I am not sure why this happens and how to prevent this (splitting the array manually is a workaround but of course one misses the overlapping structures), so any help will be appreciated.
Thanks!