You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Running Prefect /w Dask on Kubernetes results in huge memory usage as reported by Dask the list object is only ~47MB however when submitted to the workers it increases by 300x resulting in 13.7G object regardless to say this is expressive and doesn't scale well. please note that I tried methods mentioned at #3966 and dask/distributed#4091
Hi, sorry to hear you're having issues with this. I suspect this has less to do with the size of your input object and more to do with the overhead of creating a million orchestrated tasks. This is on our roadmap to investigate, but it's likely to require some low level optimizations. Can you share more information about what objects are consuming the memory?
Description
Running Prefect /w Dask on Kubernetes results in huge memory usage as reported by Dask the list object is only ~47MB however when submitted to the workers it increases by 300x resulting in 13.7G object regardless to say this is expressive and doesn't scale well. please note that I tried methods mentioned at #3966 and dask/distributed#4091
Expected Behavior
Dask scheduler mentioned using
scatter
method I'm not really familiar with how Prefect's handle DDG so I don't think it's a matter of usingscatter
before submitting at https://github.com/PrefectHQ/prefect/blob/master/src/prefect/executors/dask.py#L421Reproduction
Environment
The text was updated successfully, but these errors were encountered: