Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Features/584 data parallel #660

Merged
merged 223 commits into from
Feb 26, 2021
Merged

Features/584 data parallel #660

merged 223 commits into from
Feb 26, 2021

Conversation

coquelin77
Copy link
Member

@coquelin77 coquelin77 commented Aug 31, 2020

Description

Implementation of Data Parallel Neural Networks. Documentation about how to use in still being written. This is still not a final iteration, there are many things which will be tuned in the background.

Issue/s resolved: #584 #585 #603 #604 #605 #606

Changes proposed:

  • Implement DataParallel class
  • Implement DataLoader class
  • Implement Dataset class (template)
  • call through to torch.nn modules and torch.nn.functional routines
  • Added MNIST example as a working data parallel NN to examples/nn/mnist
  • MNISTDataset added for loading data and preparing for the network architecture
  • Image-net example added
  • Partial data loading from an HDF5 file is made available with an iterator and dataset class.

Type of change

  • New feature (non-breaking change which adds functionality)
  • Documentation update

Due Diligence

  • All split configurations tested
  • Multiple dtypes tested in relevant functions
  • Documentation updated (if needed)
  • Updated changelog.md under the title "Pending Additions"

Does this change modify the behaviour of other functions? If so, which?

no

Notes

There will be backend changes for efficiency coming in the future. however, this feature is becoming rather large and a review/merge would make it more manageable in the future

  • changelog still to be fixed!

coquelin77 and others added 30 commits June 16, 2020 13:51
@coquelin77
Copy link
Member Author

This branch is updated and the tests run clean. there is no way to increase the coverage for this branch without overhauling how coverage is reported.

It is also important to note that since this development, there have been many improvements to the data parallel neural networks which builds on these addition. If there are questions / comments for this PR I will give it until the end of the year, then it will be merged into the master. I apologize for the length of this PR as well. The length was unavoidable for this feature

@coquelin77
Copy link
Member Author

rerun tests

1 similar comment
@coquelin77
Copy link
Member Author

rerun tests

@coquelin77
Copy link
Member Author

this is going to be forced in. Other changes are waiting and can be made on the fly

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
testing Implementation of tests, or test-related issues
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Feature: DataParallel Class
3 participants