Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enable non-CUDA device for CIFAR10 and HelloDeepSpeed training example #651

Merged
merged 13 commits into from
Aug 15, 2023

Conversation

delock
Copy link
Contributor

@delock delock commented Jul 18, 2023

This PR enables BF16 data type support for CIFAR10 training example. The following changes are added:

  1. Support the case that bfloat16 is turned on in ds_config.json
  2. Fix compatibility issues with latest Pytorch, where dataiter.next() is no longer supported by Pytorch
  3. Use accelerator abstration interface to allow run this example on other accelerators.

@delock delock changed the title Enable BF16 datatype for CIFAR10 training example Enable non-CUDA device for CIFAR10 and HelloDeepSpeed training example Jul 25, 2023
@delock
Copy link
Contributor Author

delock commented Aug 7, 2023

Hi @conglongli , can you review this PR? This PR enables CIFAR10 and HelloDeepSpeed training for non-CUDA devices i.e. CPU training. Thanks!

@tjruwase
Copy link
Contributor

@delock, is this ready to merge?

@delock
Copy link
Contributor Author

delock commented Aug 15, 2023

@delock, is this ready to merge?

Hi @tjruwase , yes this is ready to merge. Thanks for reviewing this PR!

@tjruwase tjruwase merged commit c94321a into deepspeedai:master Aug 15, 2023
LeetJoe pushed a commit to LeetJoe/DeepSpeedExamples that referenced this pull request Sep 15, 2023
deepspeedai#651)

* support bf16 and CPU accelerator

* support both bfloat16 and fp16 data type

* change default data type to bf16 to help run this demo on both CPU and GPU

* enable HelloDeepSpeed for non-CUDA device

* revert changes for sh output

* allow select bf16/fp16 datatype

* revert unnecessary changes

* seperate bf16 and fp16 config

---------

Co-authored-by: Olatunji Ruwase <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants