Hi,
I have a use case where I want to take all rows in a large Postgres database table, do a bit of computation for each row separately, and write something back to the database. I am new to Dask and to me it looks like it would be a good use case for it as this is something that will require a lot of memory and that is massively parallel.
However, I wanted to confirm that this is indeed something Dask would be good for? And if so, should I do everything in Dask Dataframes or maybe first load the data “manually” and use Dask only for the map operation?
Thanks