Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Dask-CuDF: use default Dask Dataframe optimizer #8581

Merged

Conversation

madsbk
Copy link
Member

@madsbk madsbk commented Jun 22, 2021

In order to use the new HighLevelGraph optimization work in Dask/Distributed, this PR makes dask_cudf.Dataframes use the default Dask optimizer.
Previously, we have been explicitly materialized the HighLevelGraphs when calling submit() and compute() on dask_cudf.Dataframes.

Overall, this should improve performance but by default low-level task optimizations are disabled, which might have a negative impact. High-level optimizations are done in any case and we are working on moving all low-level optimization to high-level but currently low-level optimization such as array slicing is only supported by the low-level.

I don't think we will be missing any low-level optimizations related to Dataframes so I think we should follow Dask on this one and disable low-level optimizations by default.
It is possible to enable low-level optimizations explicitly by setting the Dask config like:

dask.config.set({"optimization.fuse.active": True})

cc. @jakirkham, @quasiben, @beckernick, @VibhuJawa

@madsbk madsbk added 2 - In Progress Currently a work in progress Performance Performance related issue improvement Improvement / enhancement to an existing function non-breaking Non-breaking change labels Jun 22, 2021
@github-actions github-actions bot added the Python Affects Python cuDF API. label Jun 22, 2021
@jakirkham
Copy link
Member

cc @rjzamora

@jakirkham
Copy link
Member

Interestingly this optimize function has been around since Jim initial setup Dask-cuDF (called Dask-GDF at the time). Am curious why it was needed. @jcrist do you happen to remember? If not, no worries 🙂

@madsbk
Copy link
Member Author

madsbk commented Jun 23, 2021

rerun tests

@codecov
Copy link

codecov bot commented Jun 23, 2021

Codecov Report

❗ No coverage uploaded for pull request base (branch-21.08@aa67b13). Click here to learn what that means.
The diff coverage is n/a.

❗ Current head 75194d7 differs from pull request most recent head 3fc25c8. Consider uploading reports for the commit 3fc25c8 to get more accurate results
Impacted file tree graph

@@               Coverage Diff               @@
##             branch-21.08    #8581   +/-   ##
===============================================
  Coverage                ?   82.61%           
===============================================
  Files                   ?      109           
  Lines                   ?    17850           
  Branches                ?        0           
===============================================
  Hits                    ?    14747           
  Misses                  ?     3103           
  Partials                ?        0           

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update aa67b13...3fc25c8. Read the comment docs.

@madsbk madsbk marked this pull request as ready for review June 23, 2021 10:03
@madsbk madsbk requested a review from a team as a code owner June 23, 2021 10:03
@madsbk madsbk added 3 - Ready for Review Ready for review by team and removed 2 - In Progress Currently a work in progress labels Jun 23, 2021
dsk,
keys,
dependencies=dependencies,
ave_width=_globals.get("fuse_ave_width", 1),
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Was wondering if we should be setting this somehow to ensure it is 1. However that appears to be the default anyways. So that doesn't seem needed

The other culling steps happen naturally as part of optimize anyways so no need to reproduce them otherwise either

@jakirkham
Copy link
Member

@gpucibot merge

@rapids-bot rapids-bot bot merged commit 99808ab into rapidsai:branch-21.08 Jun 23, 2021
@jakirkham
Copy link
Member

Thanks Mads! 😄

@madsbk madsbk deleted the dask_cudf_use_default_dask_optimizer branch April 5, 2022 10:35
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
3 - Ready for Review Ready for review by team improvement Improvement / enhancement to an existing function non-breaking Non-breaking change Performance Performance related issue Python Affects Python cuDF API.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants