I’ve been exploring the integration of Dask distributed computing with Celery for asynchronous processing of large CSV files. Could you confirm if this is feasible, or if you have any alternative ideas for achieving this?
I’ve tried to computing inside the celery task. But it throws the error,
File "/home/julia/conda/envs/dask-dev/lib/python3.9/asyncio/base_events.py", line 814, in run_in_executor
executor.submit(func, *args), loop=self)
File "/home/julia/conda/envs/dask-dev/lib/python3.9/concurrent/futures/thread.py", line 167, in submit
raise RuntimeError('cannot schedule new futures after shutdown')
RuntimeError: cannot schedule new futures after shutdown