Hi
Very sorry again for the late reply.
- For the memory part, yes it is quit a lot. Strange thing, when I set memory size to something around 10 GB, it starts complaining about memory spilling and worker exceeding %95 of memory budget or something like that. I don’t know why. I didn’t receive these problems when running the distributed scheduler on a local machine.
- The
LocalCluster
is implemented as a distributed dask scheduler running on a single machine with number of processes, and number of threads per process. - For the
performance_report
when running dask.distributed integrated withSLURMCluster
, I activate my Python environment first before running the main code. But as a precaution, I also activate the same python environment in theSLURMCluster
configuration to be sure the workers are using the same Python environment.
Regards