Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MP ZeRO++ #3954

Merged
merged 51 commits into from
Aug 21, 2023
Merged

MP ZeRO++ #3954

merged 51 commits into from
Aug 21, 2023

Conversation

HeyangQin
Copy link
Contributor

@HeyangQin HeyangQin commented Jul 13, 2023

As a follow-up and extension of the ZeRO++ release, the mixed precision ZeRO++ PR grants users the option to permanently keep the non-trainable weights quantized, which is very useful for LoRA. Compared with the standard weights quantization in ZeRO++, it allows for reduced memory usage and even better throughput. Many thanks to Sam for helping with this implementation.

HeyangQin and others added 30 commits June 21, 2023 11:51
* fix conv_flops_compute when padding is a str when stride=1

* fix error

* change type of paddings to tuple

* fix padding calculation

* apply formatting check

---------

Co-authored-by: Cheng Li <[email protected]>
Co-authored-by: Olatunji Ruwase <[email protected]>
* Update profiler.py

* pre-commit run --all-files

* Delete .DS_Store

* Delete .DS_Store

* Delete .DS_Store

---------

Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Cheng Li <[email protected]>
* zeropp chinese blog

* try better quality images

* make title larger

* even larger...

* various fix

* center captions

* more fixes

* fix format
Co-authored-by: Stephen Youn <[email protected]>
Co-authored-by: Arash Bakhtiari <[email protected]>
Co-authored-by: Cheng Li <[email protected]>
Co-authored-by: Ethan Doe <[email protected]>
Co-authored-by: yidoe <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: HeyangQin <[email protected]>
Co-authored-by: GuanhuaWang <[email protected]>
Co-authored-by: cmikeh2 <[email protected]>
Co-authored-by: Ammar Ahmad Awan <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Michael Wyatt <[email protected]>
Co-authored-by: Olatunji Ruwase <[email protected]>
Co-authored-by: Reza Yazdani <[email protected]>
* zeropp chinese blog

* try better quality images

* make title larger

* even larger...

* various fix

* center captions

* more fixes

* fix format

* add ZeRO++ Japanese blog

* add links

---------

Co-authored-by: HeyangQin <[email protected]>
Co-authored-by: Conglong Li <[email protected]>
* fix autotuner when backward is not called

* fix format

---------

Co-authored-by: Olatunji Ruwase <[email protected]>
Co-authored-by: Ammar Ahmad Awan <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Logan Adams <[email protected]>
* Bug fix

* Fixed formatting error

---------

Co-authored-by: Logan Adams <[email protected]>
Co-authored-by: Stephen Youn <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>
@HeyangQin HeyangQin changed the title Mixed precision LoRA release Mixed precision ZeRO++ release Aug 17, 2023
@HeyangQin HeyangQin changed the title Mixed precision ZeRO++ release MP ZeRO++ Aug 17, 2023
@HeyangQin HeyangQin enabled auto-merge August 18, 2023 18:54
@HeyangQin HeyangQin added this pull request to the merge queue Aug 20, 2023
@github-merge-queue github-merge-queue bot removed this pull request from the merge queue due to failed status checks Aug 21, 2023
@HeyangQin HeyangQin added this pull request to the merge queue Aug 21, 2023
Merged via the queue into master with commit 7711bdb Aug 21, 2023
@jeffra jeffra deleted the HeyangQin/mixed_precision_lora branch August 31, 2023 16:15
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.