-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Strange behavior using PyTorch DDP #32
Comments
@burchim |
Hi @snakers4! I also recently experimented replacing it with the official torchaudio.transforms.RNNTLoss loss from torchaudio 0.10.0. |
Thanks for the heads up about the |
@burchim |
Yes, this means that logits / target lengths tensors do not match the logits / target tensors. |
Because I used the targets lengths instead of logits lengths, stupid error |
@snakers4 |
@1ytic
Hi,
So far I have been able to use the loss with DDP on a single GPU , it behaves more or less as expected.
But when I use more than 1 device, the following happens:
GPU-0
loss is calculated properlyGPU-1
loss is close to zero for each batchI checked the input tensors, devices, tensor values, etc - so far everything seems to be identical for
GPU-0
and other GPUs.The text was updated successfully, but these errors were encountered: