-
Notifications
You must be signed in to change notification settings - Fork 915
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEA] [Java] Add a way to allocate via cudaMalloc for device memory buffers #9270
Comments
Why bypass RMM? You can use more than resource at a time. So just use |
Everything we have right now assumes the RMM pool is a singleton, and on top of that, that there's a single resource. Because of that it seemed simpler to limit the RMM pool to the amount of free memory - the amount needed for GPUDirect RDMA, and allocating directly. That said, we do want to make |
Fixes #9270 Authors: - Rong Ou (https://github.com/rongou) Approvers: - Jason Lowe (https://github.com/jlowe) - Alessandro Bellina (https://github.com/abellina) URL: #9311
Note the way this was implemented was via the |
Given #9201, we are starting to experiment with the CUDA Async Allocator.
cudaMallocAsync
does not support GPUDirect RDMA, so we would like to be able to allocate some memory (i.e. bounce buffers) directly usingcudaMalloc
.One approach to get this done is to expose a
DeviceMemoryBuffer.cudaAllocate
orDeviceMemoryBuffer.cudaMalloc
, that implies bypassingRMM
, in order to retain the GPUDirect RDMA functionality in certain cases (RAPIDS Spark with UCX)The text was updated successfully, but these errors were encountered: