pytorch/torch/distributed/algorithms
Rohan Varma 281663955f [Opt Overlap] Create Optimizer Hook State directly from functional optim (#71602)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/71602

The design in https://github.com/pytorch/pytorch/issues/67570 requires
`_OptimizerHookState` to be created directly from a functional optimizer. Add
support and tests for this. Also refactor a few tests.
ghstack-source-id: 147577175

Test Plan: CI

Reviewed By: cbalioglu

Differential Revision: D33687477

fbshipit-source-id: f3c789aa77773f918e01a8d0cf08739b2edf07b3
(cherry picked from commit 4851e1c6d4)
2022-01-25 18:32:13 +00:00
..
_checkpoint [FSDP/Checkpoint] Activation offload support in checkpoint_wrapper (#70165) 2021-12-21 10:08:18 -08:00
ddp_comm_hooks [Opt Overlap] Create Optimizer Hook State directly from functional optim (#71602) 2022-01-25 18:32:13 +00:00
model_averaging [LocalSGD] Move feature to Beta, clean up some docs (#71621) 2022-01-21 21:10:42 +00:00
quantization [BE] minor improvement to dist quantization (#67401) 2021-11-21 23:31:59 -08:00
__init__.py Make _Join, _Joinable, _JoinHook public (#62605) 2021-08-03 12:20:11 -07:00
join.py Make _Join, _Joinable, _JoinHook public (#62605) 2021-08-03 12:20:11 -07:00