Best practices for distributing memory? #16618
Unanswered
rkruegs123
asked this question in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have a memory-intensive computation that I want to scale up by using multiple GPUs. I have been experimenting with
xmap
,pmap
, andshmap
. My understanding is that these would typically be use to distribute batches in a training set to parallelize computation. Note that I am not training a neural network, and therefore the computation I am trying to parallelize is not analogous -- instead, the computation I am trying to vectorize (and distribute across GPUs) is in ascan
.I have found two things:
scan
/jitted function isshmap
. This is fine.So, does anybody know how to distribute memory across GPUs when the computation to be vectorized is within another function? A minimal example would be extremely helpful.
Thank you in advance.
Beta Was this translation helpful? Give feedback.
All reactions