Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[CI] support multiple CUDA versions #1234

Closed

Conversation

shintaro-iwasaki
Copy link
Contributor

Summary:
This patch adds multiple CUDA versions for OSS CI (build only). Previously on the OSS side, we tested only 11.3, but this patch expands it to 11.3, 11.5, and 11.6.

This CI is useful. For example, it revealed an issue in cub_namespace_postfix.cuh: this change was cherry-picked for CUB 1.13.1 (https://github.com/NVIDIA/cub/blob/1.13.X/cub/util_namespace.cuh), so the CUB_VERSION condition should be CUB_VERSION >= 101301.

Q&A

  • Why don't we test multiple ROCm versions?
    • PyTorch provides two versions for ROCm at this point: 5.0 and 5.1.1, but FBGEMM does not support ROCm 5.0.
  • Why don't we test CUDA 11.2 or older?
    • I am not sure if FBGEMM supports CUDA <= 11.2. If needed, we can add it.
  • Why don't we test CUDA 11.4?
    • PyTorch does not have nightly CUDA 11.4 build for some reasons at present.
  • Why don't we test CUDA 11.7?
    • It has some compilation issues.

Differential Revision: D38359867

Summary:
This patch adds multiple CUDA versions for OSS CI (build only). Previously on the OSS side, we tested only 11.3, but this patch expands it to 11.3, 11.5, and 11.6.

This CI is useful. For example, it revealed an issue in `cub_namespace_postfix.cuh`: this change was cherry-picked  for CUB 1.13.1 (https://github.com/NVIDIA/cub/blob/1.13.X/cub/util_namespace.cuh), so the `CUB_VERSION` condition should be `CUB_VERSION >= 101301`.

Q&A
 - Why don't we test multiple ROCm versions?
    - PyTorch provides two versions for ROCm at this point: 5.0 and 5.1.1, but FBGEMM does not support ROCm 5.0.
 - Why don't we test CUDA 11.2 or older?
   - I am not sure if FBGEMM supports CUDA <= 11.2. If needed, we can add it.
 - Why don't we test CUDA 11.4?
   - PyTorch does not have nightly CUDA 11.4 build for some reasons at present.
 - Why don't we test CUDA 11.7?
   - It has some compilation issues.

Differential Revision: D38359867

fbshipit-source-id: faf6d8321269da72b92cf2af3ee1f940531ff8b8
@netlify
Copy link

netlify bot commented Aug 2, 2022

Deploy Preview for eclectic-stroopwafel-199537 canceled.

Name Link
🔨 Latest commit ba80ccf
🔍 Latest deploy log https://app.netlify.com/sites/eclectic-stroopwafel-199537/deploys/62e99cab34bc9a0008aa1fda

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D38359867

@shintaro-iwasaki shintaro-iwasaki changed the title support multiple CUDA versions [CI] support multiple CUDA versions Aug 2, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants