Skip to content

Conversation

@ShadenSmith
Copy link
Contributor

No description provided.

@ShadenSmith ShadenSmith requested a review from jeffra March 18, 2020 01:22
@ShadenSmith ShadenSmith merged commit 4d73594 into deepspeedai:master Mar 18, 2020
@ShadenSmith ShadenSmith deleted the web-edits branch March 18, 2020 01:24
kouml pushed a commit to kouml/DeepSpeed that referenced this pull request Apr 3, 2020
* Add coming soon to posts

* Add what's new section to main page
samyam added a commit that referenced this pull request Mar 8, 2021
Co-authored-by: Samyam <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Samyam Rajbhandari <[email protected]>
Co-authored-by: Olatunji Ruwase <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: eltonzheng <[email protected]>
samyam added a commit that referenced this pull request Mar 8, 2021
* Squash stage3 v1 (#146)

Co-authored-by: Samyam <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Samyam Rajbhandari <[email protected]>
Co-authored-by: Olatunji Ruwase <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: eltonzheng <[email protected]>

* Fix correctness bug (#147)

* formatting fix (#150)

* stage3 bugfix (API) update and simplified FP16 Z3 tests (#151)

* fp16 Z3 API update and bugfix

* revert debug change

* ZeRO-3 detach and race condition bugfixes (#149)

* trying out ZeRO-3 race condition fix

* CUDA sync instead of stream

* reduction stream sync

* remove commented code

* Fix optimizer state_dict KeyError (#148)

Co-authored-by: Jeff Rasley <[email protected]>

* fix for smaller SGS sizes, ensures each grad is backed by unique tensors (#152)

* Simplifying the logic for getting averaged gradients (#153)

* skip for now

* Z3 Docs redux (#154)

* removing some TODOs and commented code (#155)

* New Z3 defaults (#156)

Co-authored-by: Jeff Rasley <[email protected]>

* formatting

* megatron external params

Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Olatunji Ruwase <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: eltonzheng <[email protected]>
jeffra added a commit to jeffra/DeepSpeed that referenced this pull request Aug 25, 2021
* set adamw_mode default true (follows FusedAdam and < 0.3.11 logic) (deepspeedai#844)

* less scary overflow notice (deepspeedai#833)

Co-authored-by: Jeff Rasley <[email protected]>

* Add optimizers and schedules to RTD and updated the corresponding part in the website (deepspeedai#799)

* add optimizers and schedules to rtd

* update ds website and fix links

* add optimizers and schedules to rtd

* update ds website and fix links

* add flops profiler to rtd

* fix

Co-authored-by: Shaden Smith <[email protected]>

* small tweaks (deepspeedai#839)

* Control ZeRO wall clock timers (deepspeedai#849)

* Control ZeRO wall clock timers

* Disable more ZeRO3 debug prints

Co-authored-by: Jeff Rasley <[email protected]>

* [WarmupDecayLR] fix log(0) & 1/log(1) bugs (deepspeedai#772)

* fix log(0) & 1/log(1) bugs

* simplify

Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Reza Yazdani <[email protected]>
Co-authored-by: Cheng Li <[email protected]>

* bump to v0.3.12

* Bug fix: Remove client optimizer param_group list item that does not have 'params' (deepspeedai#827)

Co-authored-by: Jeff Rasley <[email protected]>

* [doc] pipeline doc typos/improvements (deepspeedai#659)

Admin merging for pure-doc PR that does not trigger build.

* Samyamr/inference hook fix (deepspeedai#851)

* Fix mis-aligned-grad

When a parameter is not divisible by world size, the partitioned gradients are mis-aligned due to incorrect padding handling. This PR should fix for that.

* Formatting fix

* Adding static_scale test back for Z3, and also changing hidden size to be not divisile by world_size

* also removing alignment from flat fp16 buffers

* Testing for hidden dim alignment

* inference hook fix

* Update stage3.py

* formatting

* [bug-fix] move params to gpu if offload params is turned off

Co-authored-by: Samyam Rajbhandari <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>

* ZeRO Stage 2: Clear reduced gradients (deepspeedai#856)

* Ensure gradients of other partitions are cleared after reduction

* Remove redundant code

Co-authored-by: Jeff Rasley <[email protected]>

* Squash stage3 v1 (deepspeedai#146)

Co-authored-by: Samyam <[email protected]>
Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Samyam Rajbhandari <[email protected]>
Co-authored-by: Olatunji Ruwase <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: Shaden Smith <[email protected]>
Co-authored-by: eltonzheng <[email protected]>

* formatting fix (deepspeedai#150)

* stage3 bugfix (API) update and simplified FP16 Z3 tests (deepspeedai#151)

* fp16 Z3 API update and bugfix

* revert debug change

* docs

* filling in allocation docs

* better assumption docs

* doc progress

* config json

* major docs edits

* auto registration works for accessed cases

* working on small models.

* debugging large-model discovery?

* fix discovery to first forward pass?

* return obj ext param

* support None parameters in auto-discovery

Co-authored-by: Jeff Rasley <[email protected]>
Co-authored-by: Stas Bekman <[email protected]>
Co-authored-by: Cheng Li <[email protected]>
Co-authored-by: Olatunji Ruwase <[email protected]>
Co-authored-by: Reza Yazdani <[email protected]>
Co-authored-by: Samyam Rajbhandari <[email protected]>
Co-authored-by: eltonzheng <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants