I hit an error in some code a while back where I was accidentally passing a pandas dataframe into dask’s “to_datetime” function. I’d have expected it to either run as normal or throw an error, but the output seemed to be an assortment of duplicated rows.
I realise this isn’t an error on dask’s part at all, but in my bad implementatinon (and the fix is simple enough) but since the output seems counterintuitive to what I’d expect, I’m curious about why it is that this happens? Does anyone know what’s going on under the hood for this to outuput?