-
Notifications
You must be signed in to change notification settings - Fork 94
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[REVIEW] Pin dask
& distributed
for release
#916
Conversation
The failures seem to arise from inconsistent versions being picked from conda packages:
@Ethyling @ajschmidt8 @charlesbluca @jakirkham would be good to get your expertise here in what would be a safe way to prevent gpuCI from picking an inconsistent set of |
To be more specific, I think those issues are related to the changes from #897 . |
I think the issue here is that when Lines 69 to 74 in 63529e8
With something like this (used in cuDF's CI): if [[ "${INSTALL_DASK_MAIN}" == 1 ]]; then
gpuci_logger "gpuci_mamba_retry update dask"
gpuci_mamba_retry update dask
conda list
else
gpuci_logger "gpuci_mamba_retry install conda-forge::dask>=2022.03.0 conda-forge::distributed>=2022.03.0 conda-forge::dask-core>=2022.03.0 --force-reinstall"
gpuci_mamba_retry install conda-forge::dask>=2022.03.0 conda-forge::distributed>=2022.03.0 conda-forge::dask-core>=2022.03.0 --force-reinstall
fi |
Could we instead just update the Dask versions in https://github.com/rapidsai/integration/blob/branch-22.06/conda/recipes/versions.yaml and then eliminate the |
We could try waiting for new images & packages to be published after rapidsai/integration#476 is merged. If that works we wouldn't need else condition. |
Co-authored-by: Peter Andreas Entschev <[email protected]>
Yes, that would be good. But perhaps we can get this merged right away and then change it afterwards. Also added a few suggestions. |
Hmm it was my understanding that even after updating the version pinning in the integration axis file and removing the Dask channel from our conda config, we weren't guaranteed to have stable packages, and would want a step to explicitly ensure they are installed? At least that's what I thought the conclusion of rapidsai/cudf#10182 was |
How so? If that's the case, it then sounds to me that it's not guaranteed that pinning works for the release process which would be a major issue IMO. My hope was always that the integration repo versions MUST be respected, if not that feels like an enormous flaw to me. @Ethyling @ajschmidt8 could you comment on this? |
Codecov Report
@@ Coverage Diff @@
## branch-22.06 #916 +/- ##
==============================================
Coverage ? 0.00%
==============================================
Files ? 22
Lines ? 3075
Branches ? 0
==============================================
Hits ? 0
Misses ? 3075
Partials ? 0 Continue to review full report at Codecov.
|
Think this is more of a consequence of the fact that the Dask channel is placed at a higher priority than conda-forge during gpuCI image builds? In that case, installing We could potentially remove the
EDIT: Alternatively, since it seems like we would always want to run |
We do this for release builds I suppose. The |
Maybe I'm misunderstanding something, but this still seems problematic with the release process. When we release, won't a non-stable Dask version be picked then? Additionally, if anyone has |
Maybe we should follow up on this discussion after the release? |
rerun tests |
Nope, at least in cudf & docker image creations for release build we drop those channels:
Yea, that is possible. I don't think we have a mechanism to version pin based on channels. |
Got it. It seems a bit different in Dask-CUDA, where Dask and Distributed Lines 64 to 65 in 63529e8
I don't mean to hold off on this PR, but I want to make sure that we're not making any mistakes for the release process and that nightly packages after this PR is merged -- and ultimately the release packages -- indeed pin to the correct Dask version. If there's a mistake with that, it can be critical in the future. If there are reasons to believe that I'm overcomplicating things, please let me know, but I would like to confirm who/what/where is the actual pinning coming from/happening. |
dask>=2022.03.0 | ||
distributed>=2022.03.0 | ||
dask==2022.05.1 | ||
distributed==2022.05.1 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
These will ensure dask
& distributed
are pinned correctly in the Conda packages. In PR ( #893 ), we ensure these are pulled into the Conda packages
Thanks @galipremsagar for working on this, and everyone for clarifications. It seems like release process is correct, I will keep an eye on this just to be safe, but I think we're good here. |
@gpucibot merge |
Pinnings to be in-line with: rapidsai/cudf#10965