Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Federated Learning | Concept 24 FL for MNIST #403

Open
LeonMac opened this issue Mar 4, 2022 · 0 comments
Open

Federated Learning | Concept 24 FL for MNIST #403

LeonMac opened this issue Mar 4, 2022 · 0 comments
Labels
Type: Bug 🐛 Some functionality not working in the codebase as intended

Comments

@LeonMac
Copy link

LeonMac commented Mar 4, 2022

Description

When the DS launch up a remote training, on DO side, report "TypeError: can't convert cuda:0 device type tensor to numpy. Use Tensor.cpu() to copy the tensor to host memory first."

How to Reproduce

  1. run the code line-by-line, everything works fine, until arriving to PART 3: Training. (I have a GPU and CUDA )
  2. Training will stop at epoch 1 and no progress anymore.
  3. on DO side I can see the error report as above "TypeError: can't convert cuda:0 device type tensor to numpy. Use Tensor.cpu() to copy the tensor to host memory first."

Expected Behavior

This is a classic issue for general ML and I can find solution, but how to handle this by using FL lib (by which the training happen on DO side actually)

System Information

  • OS: ubuntu18.04
  • Language Version: Python:3.7.10, torch:1.8.1, torchvision:0.9.1
  • Package Manager Version: [e.g. conda 4.11.0, pip 21.2.2 ]
@LeonMac LeonMac added the Type: Bug 🐛 Some functionality not working in the codebase as intended label Mar 4, 2022
@LeonMac LeonMac changed the title Federated Learning | Concept 24 FL for MINST Federated Learning | Concept 24 FL for MNIST Mar 4, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Type: Bug 🐛 Some functionality not working in the codebase as intended
Projects
None yet
Development

No branches or pull requests

1 participant