pytorch/torch/distributed
Rohan Varma 5d4e170d58 [Optim in backward] API to retrieve in-backward optimizers (#105991)
API to retrieve in backward optimizer for checkpointing purposes

Differential Revision: [D47782225](https://our.internmc.facebook.com/intern/diff/D47782225/)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/105991
Approved by: https://github.com/awgu
2023-07-29 01:36:25 +00:00
..
_composable [Composable FSDP] Enable HSDP (#105206) 2023-07-26 21:03:55 +00:00
_shard [BE]: Update Ruff to 0.0.280 (#105724) 2023-07-22 23:03:34 +00:00
_sharded_tensor
_sharding_spec
_spmd Merge and improve torch optim optimizer type stubs (#102593) 2023-07-26 11:56:42 +00:00
_tensor [device_mesh][BE] reduce_scatter fallback to funcol and remove from DM (#105642) 2023-07-27 01:33:05 +00:00
_tools Fix typos under torch/distributed directory (#95638) 2023-03-27 21:13:44 +00:00
algorithms [CheckpointWrapper] Test for kwarg propagation, remove checkpoint_fn_arg support (#102679) 2023-07-28 21:18:35 +00:00
autograd
benchmarks [BE] Enable ruff's UP rules and autoformat distributed/ (#105433) 2023-07-19 14:27:11 +00:00
checkpoint Add option to disable rewriting index hints in default global save plan (#105861) 2023-07-25 06:00:13 +00:00
elastic Format: fixing multiple string concatenation in single line (#106013) 2023-07-26 18:39:18 +00:00
examples Fix typos under torch/distributed directory (#95638) 2023-03-27 21:13:44 +00:00
fsdp [FSDP] Optimize away intermediate div_ for HSDP (#106034) 2023-07-28 18:36:26 +00:00
launcher Convert logging f-strings to use % format, part four (#98705) 2023-04-11 13:17:59 +00:00
nn [BE] Enable ruff's UP rules and autoformat distributed/ (#105433) 2023-07-19 14:27:11 +00:00
optim [Optim in backward] API to retrieve in-backward optimizers (#105991) 2023-07-29 01:36:25 +00:00
pipeline Format: fixing multiple string concatenation in single line (#106013) 2023-07-26 18:39:18 +00:00
rpc [BE] Enable ruff's UP rules and autoformat distributed/ (#105433) 2023-07-19 14:27:11 +00:00
tensor Clean up unsed MHA code to avoid confusion (#105956) 2023-07-27 17:10:17 +00:00
__init__.py [BE] remove deprecated attributes from distributed_c10d (#105753) 2023-07-24 16:35:08 +00:00
_composable_state.py
_functional_collectives_impl.py [device_mesh][BE] reduce_scatter fallback to funcol and remove from DM (#105642) 2023-07-27 01:33:05 +00:00
_functional_collectives.py Refactor funcol for readability and dynamo tracing (#104387) 2023-07-06 23:29:49 +00:00
argparse_util.py
c10d_logger.py [c10d] Record time spent for init_process_group, new_group, _store_based_barrier (#101912) 2023-05-24 09:36:34 +00:00
collective_utils.py [Reland] Update mypy to 1.4.1 (#105227) 2023-07-15 20:30:20 +00:00
constants.py
CONTRIBUTING.md
distributed_c10d.py [PyTorch][ET] Collect Process Groups Mapping Info (#104373) 2023-07-25 03:34:53 +00:00
launch.py Fix typos under torch/distributed directory (#95638) 2023-03-27 21:13:44 +00:00
logging_handlers.py
remote_device.py
rendezvous.py [BE] Enable ruff's UP rules and autoformat distributed/ (#105433) 2023-07-19 14:27:11 +00:00
run.py Convert logging f-strings to use % format, part four (#98705) 2023-04-11 13:17:59 +00:00
utils.py [FSDP/DDP] Unify _cast_forward_inputs (#102680) 2023-06-04 18:31:21 +00:00