pytorch/torch/distributed
2024-06-04 19:44:30 +00:00
..
_composable [FSDP2] Fixed variable shadowing of module (#127776) 2024-06-04 10:27:34 +00:00
_cuda_p2p Introduce Inductor passes to micro-pipeline all-gather-matmul and matmul-reduce-scatter in certain cases (#126598) 2024-06-04 09:06:56 +00:00
_shard [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
_sharded_tensor [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
_sharding_spec [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
_spmd [dtensor] move some modules to private namespace (#127339) 2024-05-29 05:18:47 +00:00
_tensor [dtensor][debug] added c10d alltoall_ and alltoall_base_ to CommDebugMode (#127360) 2024-06-04 18:29:48 +00:00
_tools [BE]: Try TCH autofixes on torch/ (#125536) 2024-05-05 23:13:59 +00:00
algorithms [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
autograd
benchmarks Move doc links to point to main (#121823) 2024-03-15 19:49:37 +00:00
checkpoint [FSDP] keep paras in torch.distributed.checkpoint.state_dict.set_optimizer_state_dict (#127644) 2024-06-04 03:32:22 +00:00
elastic Reapply "distributed debug handlers (#126601)" (#127805) 2024-06-04 19:44:30 +00:00
examples
fsdp [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
launcher torchelastic: change monitor_interval default to 0.1 (#124692) 2024-04-24 01:44:41 +00:00
nn Fix get_rank under a non-default group. (#120481) 2024-03-11 05:40:54 +00:00
optim [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
pipeline Revert "Retire torch.distributed.pipeline (#127354)" 2024-06-04 18:19:31 +00:00
pipelining [pipelining] Simple 1F1B schedule (#127673) 2024-06-04 06:09:51 +00:00
rpc
tensor [BE] enable ruff rule Q from flake8-quotes (#127713) 2024-06-02 23:25:26 +00:00
__init__.py Reapply "c10d: add Collectives abstraction (#125978)" (#126695) 2024-05-21 18:00:09 +00:00
_composable_state.py
_functional_collectives_impl.py Make c10d_functional ops call into _c10d_functional ops (#124979) 2024-04-27 08:08:02 +00:00
_functional_collectives.py [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
_state_dict_utils.py [DSD] Fixes various bugs for broadcast_from_rank0 (#127635) 2024-06-03 06:35:21 +00:00
argparse_util.py
c10d_logger.py [DCP] Adds better handling in logging of specific kwargs (#123658) 2024-04-11 21:09:38 +00:00
collective_utils.py
constants.py
CONTRIBUTING.md
device_mesh.py [FSDP2] Fix submesh slicing to enable 3D parallelism (#127585) 2024-06-04 04:24:09 +00:00
distributed_c10d.py [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
launch.py [BE] wrap deprecated function/class with typing_extensions.deprecated (#127689) 2024-06-02 12:30:43 +00:00
logging_handlers.py
remote_device.py
rendezvous.py Fix public binding to actually traverse modules (#126103) 2024-05-15 19:36:03 +00:00
run.py [BE]: Improve exception typing. Remove NOQAs (#125535) 2024-05-08 14:07:13 +00:00
utils.py [BE] Enabled lintrunner on torch/distributed/utils.py (#127771) 2024-06-04 06:10:33 +00:00