| .. |
|
_composable
|
[FSDP2] Fixed variable shadowing of module (#127776)
|
2024-06-04 10:27:34 +00:00 |
|
_cuda_p2p
|
Introduce Inductor passes to micro-pipeline all-gather-matmul and matmul-reduce-scatter in certain cases (#126598)
|
2024-06-04 09:06:56 +00:00 |
|
_shard
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
_sharded_tensor
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
_sharding_spec
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
_spmd
|
[dtensor] move some modules to private namespace (#127339)
|
2024-05-29 05:18:47 +00:00 |
|
_tensor
|
[dtensor][debug] added c10d alltoall_ and alltoall_base_ to CommDebugMode (#127360)
|
2024-06-04 18:29:48 +00:00 |
|
_tools
|
[BE]: Try TCH autofixes on torch/ (#125536)
|
2024-05-05 23:13:59 +00:00 |
|
algorithms
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
autograd
|
|
|
|
benchmarks
|
Move doc links to point to main (#121823)
|
2024-03-15 19:49:37 +00:00 |
|
checkpoint
|
[FSDP] keep paras in torch.distributed.checkpoint.state_dict.set_optimizer_state_dict (#127644)
|
2024-06-04 03:32:22 +00:00 |
|
elastic
|
Reapply "distributed debug handlers (#126601)" (#127805)
|
2024-06-04 19:44:30 +00:00 |
|
examples
|
|
|
|
fsdp
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
launcher
|
torchelastic: change monitor_interval default to 0.1 (#124692)
|
2024-04-24 01:44:41 +00:00 |
|
nn
|
Fix get_rank under a non-default group. (#120481)
|
2024-03-11 05:40:54 +00:00 |
|
optim
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
pipeline
|
Revert "Retire torch.distributed.pipeline (#127354)"
|
2024-06-04 18:19:31 +00:00 |
|
pipelining
|
[pipelining] Simple 1F1B schedule (#127673)
|
2024-06-04 06:09:51 +00:00 |
|
rpc
|
|
|
|
tensor
|
[BE] enable ruff rule Q from flake8-quotes (#127713)
|
2024-06-02 23:25:26 +00:00 |
|
__init__.py
|
Reapply "c10d: add Collectives abstraction (#125978)" (#126695)
|
2024-05-21 18:00:09 +00:00 |
|
_composable_state.py
|
|
|
|
_functional_collectives_impl.py
|
Make c10d_functional ops call into _c10d_functional ops (#124979)
|
2024-04-27 08:08:02 +00:00 |
|
_functional_collectives.py
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
_state_dict_utils.py
|
[DSD] Fixes various bugs for broadcast_from_rank0 (#127635)
|
2024-06-03 06:35:21 +00:00 |
|
argparse_util.py
|
|
|
|
c10d_logger.py
|
[DCP] Adds better handling in logging of specific kwargs (#123658)
|
2024-04-11 21:09:38 +00:00 |
|
collective_utils.py
|
|
|
|
constants.py
|
|
|
|
CONTRIBUTING.md
|
|
|
|
device_mesh.py
|
[FSDP2] Fix submesh slicing to enable 3D parallelism (#127585)
|
2024-06-04 04:24:09 +00:00 |
|
distributed_c10d.py
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
launch.py
|
[BE] wrap deprecated function/class with typing_extensions.deprecated (#127689)
|
2024-06-02 12:30:43 +00:00 |
|
logging_handlers.py
|
|
|
|
remote_device.py
|
|
|
|
rendezvous.py
|
Fix public binding to actually traverse modules (#126103)
|
2024-05-15 19:36:03 +00:00 |
|
run.py
|
[BE]: Improve exception typing. Remove NOQAs (#125535)
|
2024-05-08 14:07:13 +00:00 |
|
utils.py
|
[BE] Enabled lintrunner on torch/distributed/utils.py (#127771)
|
2024-06-04 06:10:33 +00:00 |