pytorch/torch/distributed
shibo19 c2095af3f8 make funcs argument type from torch.cuda.stream as torch.Stream (#104156)
Fixes #ISSUE_NUMBER
1. we want to support fsdp for custom device, so we make funcs argument type from torch.cuda.stream as torch.Stream
Pull Request resolved: https://github.com/pytorch/pytorch/pull/104156
Approved by: https://github.com/awgu
2023-06-28 06:02:56 +00:00
..
_composable Fix composable checkpoint(use_reentrant=True) with multi args (#103590) 2023-06-14 21:53:30 +00:00
_shard enforce dtype (reland) (#102996) 2023-06-06 00:35:04 +00:00
_sharded_tensor
_sharding_spec
_spmd [distributed] implemented find_all_descendants (#102138) 2023-05-24 21:47:59 +00:00
_tensor [DTensor][Random] Introduce CudaRNGStateTracker to maintain parallel RNG state for DTensor (#103235) 2023-06-27 19:00:25 +00:00
_tools Fix typos under torch/distributed directory (#95638) 2023-03-27 21:13:44 +00:00
algorithms [CheckpointWrapper] Warn on reentrant use (#102890) 2023-06-04 18:31:22 +00:00
autograd
benchmarks
checkpoint [DCP] Rewrite read slicing to use a wrapper. (#99167) 2023-06-08 13:52:13 +00:00
elastic [BE] Fix flake8 B027 errors - missing abstractmethod decorator (#100715) 2023-05-09 17:28:48 +00:00
examples Fix typos under torch/distributed directory (#95638) 2023-03-27 21:13:44 +00:00
fsdp make funcs argument type from torch.cuda.stream as torch.Stream (#104156) 2023-06-28 06:02:56 +00:00
launcher Convert logging f-strings to use % format, part four (#98705) 2023-04-11 13:17:59 +00:00
nn Add assign kwarg to module.load_state_dict (#102212) 2023-06-15 18:41:00 +00:00
optim [FSDP][optim_state_dict] Cleanup the unused optimizer state_dict APIs (#103781) 2023-06-21 05:38:48 +00:00
pipeline Enable ruff in lintrunner (#99785) 2023-04-24 16:18:44 +00:00
rpc [BE]: enable PLE error codes in ruff and fix bugs (#101079) 2023-05-11 23:57:25 +00:00
tensor [TP] Add an input resharding wrapper for TP and unit test for 2D + AC (#103334) 2023-06-23 04:05:01 +00:00
__init__.py [c10d] Record time spent for init_process_group, new_group, _store_based_barrier (#101912) 2023-05-24 09:36:34 +00:00
_composable_state.py
_functional_collectives.py [C10D] Add functional collective reduce_scatter_into_tensor_coalesced. (#101023) 2023-06-23 19:24:11 +00:00
argparse_util.py
c10d_logger.py [c10d] Record time spent for init_process_group, new_group, _store_based_barrier (#101912) 2023-05-24 09:36:34 +00:00
collective_utils.py Initial commit of collective_utils (#101037) 2023-06-27 02:15:16 +00:00
constants.py
CONTRIBUTING.md
distributed_c10d.py Implement coalesced reduce_scatter_tensor (#103561) 2023-06-15 20:11:12 +00:00
launch.py Fix typos under torch/distributed directory (#95638) 2023-03-27 21:13:44 +00:00
logging_handlers.py
remote_device.py
rendezvous.py Revisit torch._six.string_classes removal (#94709) (#97863) 2023-03-30 17:02:45 +00:00
run.py Convert logging f-strings to use % format, part four (#98705) 2023-04-11 13:17:59 +00:00
utils.py [FSDP/DDP] Unify _cast_forward_inputs (#102680) 2023-06-04 18:31:21 +00:00