This repository has been archived by the owner on Jul 31, 2024. It is now read-only.
comm.py
should maybe consider backend-specific support of different devices
#30
Labels
enhancement
New feature or request
Depending on the backend, distributed communication may only be supported on either CPU or GPU, see table here.
Right now, in
comm.py
communication is always done on the GPU, see here e.g.:torchpack/torchpack/distributed/comm.py
Lines 32 to 34 in d3fda52
I would suggest considering the backend-specific device support for both
allgather()
andbroadcast()
to ensure the functions are usable across multiple backends.torch.distributed.broadcast_object_list
andtorch.distributed.all_gather_object
might be a useful starting points for this.The text was updated successfully, but these errors were encountered: