William Wen
433e43cbec
[dynamo, docs] programming model dynamo core concepts ( #157985 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/157985
Approved by: https://github.com/svekars , https://github.com/anijain2305
2025-07-29 01:53:34 +00:00
PyTorch MergeBot
fe0ff12dab
Revert "[Inductor] Support native Inductor as backend for MTIA ( #158526 )"
...
This reverts commit cd68559d04 .
Reverted https://github.com/pytorch/pytorch/pull/158526 on behalf of https://github.com/facebook-github-bot due to Diff reverted internally ([comment](https://github.com/pytorch/pytorch/pull/158526#issuecomment-3122186057 ))
2025-07-26 17:58:00 +00:00
anwang
cd68559d04
[Inductor] Support native Inductor as backend for MTIA ( #158526 )
...
This diff/PR includes the changes to support native Inductor integration for MTIA. The goal is to support `torch.compile(backend="inductor")` for MTIA. Inductor should generate code(triton kernel + python wrapper code) similar to CUDA. And the triton kernels can be launched eagerly.
The changes include:
- Add MTIA device interfaces used by Dynamo and Inductor, including APIs on device, stream, event, etc.
- Add required torch.mtia APIs, like is_bf16_supported, memory_allocated, set_stream_by_id, etc.
- MTIA specific codegen logic, for example, loading MTIA dynamic_library.
- Other necessary changes to integrate with Inductor codegn, following other devices like CUDA, XPU.
- Integrate with the [empty_strided_mtia](https://www.internalfb.com/code/fbsource/[0d017d3a4a1bdff7253f9c66a9f38e77bd62166b]/fbcode/caffe2/aten/src/ATen/native/mtia/EmptyTensor.cpp?lines=49%2C63%2C71%2C74%2C78 ) API that we’ve added for the new MTIA ATen backend.
- A change in Inductor runtime to avoid re-initialize MTIADriver.
- BUCK changes to include ATen-mtia in Inductor, and to use -USE_MTIA preprocessor flag.
- Update `test_mnist_e2e.py` to cover native Inductor as backend, using the `--use_native_inductor` flag.
- Add a personal script(`scripts/anwang/run_native_inductor_script.py`) for testing purpose.
Note:
- This approach(option 3) aims to provide a pytorch native approach of Inductor integration for MTIA, minimizing the onboarding overhead. The downside of this approach is that it doesn't leverage MTIA specific graph optimization, and is limited to eagerly launch overhead.
- MTIA will support another approach(option 2) to provide best performance, based on WrapperFxCodegen. We should be able to reuse the fundamental changes of this diff for option 2, like the device interfaces, steam/event APIs, etc, especially as WrapperFxCodegen inherits PythonWrapperCodegen.
Internal:
References:
- [post for context](https://fb.workplace.com/groups/mtiasw/permalink/1718377262384606/ )
- [Inductor integration discussion(option 1/2/3)](https://docs.google.com/document/d/1p6363OXtVIRv1hPoaKlRSK3j-iir3QIbDd5bjyqCNig/edit?tab=t.0#heading=h.7s4ns6wcnhmb )
- [Project design doc(option 3)](https://docs.google.com/document/d/1jXUmhgoV9WvkMf-bcY3Od_kK9K_RDOdgHdt1LoQ5Tc4/edit?tab=t.0#heading=h.y43gwdqlv46w )
- [early prototying diff](https://www.internalfb.com/diff/D75110196 )
- [MPS integration PR](https://github.com/pytorch/pytorch/pull/153959 )
- [empty_strided_xpu PR](https://github.com/pytorch/pytorch/pull/126678 )
Differential Revision: [D78458745](https://our.internmc.facebook.com/intern/diff/D78458745/ )
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158526
Approved by: https://github.com/blaine-rister , https://github.com/jansel , https://github.com/eellison
2025-07-26 08:16:34 +00:00
Mikayla Gawarecki
36cf8f1ed8
[BE] Use .md instead of .rst for nn.aliases doc ( #158666 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158666
Approved by: https://github.com/janeyx99
ghstack dependencies: #158491 , #158654
2025-07-25 22:03:55 +00:00
Mikayla Gawarecki
1e79872f2e
[BE] More torch.nn docs coverage test (except for torch.nn.parallel) ( #158654 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158654
Approved by: https://github.com/janeyx99
ghstack dependencies: #158491
2025-07-25 22:03:55 +00:00
Mikayla Gawarecki
9e8f27cc79
[BE] Make torch.nn.modules.* satisfy the docs coverage test ( #158491 )
...
Options to address the "undocumented python objects":
1. Reference the functions in the .rst via the torch.nn.modules namespace. Note that this changes the generated doc filenames / locations for most of these functions!
2. [Not an option] Monkeypatch `__module__` for these objects (broke several tests in CI due to `inspect.findsource` failing after this change)
3. Update the .rst files to also document the torch.nn.modules forms of these functions, duplicating docs.
#### [this is the docs page added](https://docs-preview.pytorch.org/pytorch/pytorch/158491/nn.aliases.html )
This PR takes option 3 by adding an rst page nn.aliases that documents the aliases in nested namespaces, removing all the torch.nn.modules.* entries from the coverage skiplist except
- NLLLoss2d (deprecated)
- Container (deprecated)
- CrossMapLRN2d (what is this?)
- NonDynamicallyQuantizableLinear
This mostly required adding docstrings to `forward`, `extra_repr` and `reset_parameters`. Since forward arguments are already part of the module docstrings I just added a very basic docstring.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158491
Approved by: https://github.com/janeyx99
2025-07-25 22:03:55 +00:00
raghavhrishi
7ef3c3357d
NUMA binding integration with elastic agent and torchrun ( #149334 )
...
Implements #148689
Pull Request resolved: https://github.com/pytorch/pytorch/pull/149334
Approved by: https://github.com/d4l3k
Co-authored-by: Paul de Supinski <pdesupinski@gmail.com>
2025-07-25 21:19:49 +00:00
Joel Schlosser
316c188a5e
Remove torch.functional entries from the doc ignore list ( #158581 )
...
Options to address the "undocumented python objects":
1. Reference the functions in the .rst via the `torch.functional` namespace. Note that this changes the generated doc filenames / locations for most of these functions!
2. Document these functions by referencing them from the `torch.` namespace instead, in line with common usage. This would also require setting the `__module__` for these functions and moving entries from `torch.functional`'s `__all__` -> `torch`'s `__all__`, which is BC-breaking.
3. Update the .rst files to also document the `torch.functional` forms of these functions, duplicating docs.
This PR takes option (3) above and:
* Removes all 20 `torch.functional` entries from the doc ignore list
* Removes `torch.functional.align_tensors()` entirely, since we don't want to document it.
* This is technically BC-breaking, although the previous impl simply errored out. This change could be moved to a separate isolated PR for safety.
* Introduces `torch.aliases.md` as a hidden page for the `torch.functional` aliases to the `torch` analogue functions
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158581
Approved by: https://github.com/janeyx99
2025-07-25 17:19:01 +00:00
PyTorch MergeBot
c8316d0e79
Revert "[BE] Remove torch deploy | remove torch deploy specific files ( #158290 )"
...
This reverts commit 6ed2cb6ccd .
Reverted https://github.com/pytorch/pytorch/pull/158290 on behalf of https://github.com/ZainRizvi due to Reverting as per offline discussion to fix internal breaks. @PaliC will reland this as a codev diff. Instructions here: https://fburl.com/fixing-ghfirst-reverts ([comment](https://github.com/pytorch/pytorch/pull/158288#issuecomment-3119037960 ))
2025-07-25 16:09:39 +00:00
PyTorch MergeBot
a9f6770edd
Revert "[BE] Remove __reduce_deploy__ ( #158291 )"
...
This reverts commit 9c68c4d08f .
Reverted https://github.com/pytorch/pytorch/pull/158291 on behalf of https://github.com/ZainRizvi due to Reverting as per offline discussion to fix internal breaks. @PaliC will reland this as a codev diff. Instructions here: https://fburl.com/fixing-ghfirst-reverts ([comment](https://github.com/pytorch/pytorch/pull/158288#issuecomment-3119037960 ))
2025-07-25 16:09:39 +00:00
Jeff Daily
9b29166f57
[ROCm] add flag torch.backends.miopen.immediate ( #158951 )
...
The MIOpen integration has changed over the years. In the past, the MIOpen default for benchmark was True and if it were set to False it would use MIOpen Immediate Mode. But with #145294 the MIOpen benchmark default changed to False and to activate immediate mode you would set the deterministic flag to True. This has proved too restrictive because benchmark and deterministic flags are independent from immediate mode. Thus, immediate mode needs its own flag. Though MIOpen still masquerades behind torch.backends.cudnn and its flags, it seemed inappropriate to add an miopen-exclusive flag to the set of cudnn flags. This PR adds the first miopen-only flag to control its immediate mode.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158951
Approved by: https://github.com/jeffdaily
Co-authored-by: Jeff Daily <jeff.daily@amd.com>
2025-07-25 04:01:51 +00:00
Xuehai Pan
f5e2de928b
[BE] fix remaining flake8 v7 warnings ( #159044 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/159044
Approved by: https://github.com/Skylion007
ghstack dependencies: #159043
2025-07-25 02:56:34 +00:00
Ti-Tai Wang
da35562bba
[ONNX] Filter out torchscript sentences ( #158850 )
...
Fixes #157300
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158850
Approved by: https://github.com/justinchuby , https://github.com/svekars
2025-07-24 20:59:06 +00:00
Wei (Will) Feng
693197eed6
[doc] remove FSDP1 developer note ( #158991 )
...
this resolve pytorch doc audit - we remove fsdp1 doc and promote fsdp2
https://docs.pytorch.org/tutorials/intermediate/FSDP_tutorial.html
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158991
Approved by: https://github.com/svekars , https://github.com/mori360
ghstack dependencies: #158989
2025-07-24 08:21:54 +00:00
Wei (Will) Feng
68349118b5
[doc] add weifengpy to torch distributed pocs ( #158989 )
...
<img width="415" height="355" alt="Screenshot 2025-07-23 at 16 02 12" src="https://github.com/user-attachments/assets/35b6bb45-d5ed-4d74-8369-e8e66aaa2618 " />
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158989
Approved by: https://github.com/mori360
2025-07-24 04:42:33 +00:00
Mikayla Gawarecki
7f649ed4f8
Add basic torch.hash_tensor op ( #154149 )
...
Added `torch.hash_tensor` reduction function with a `mode` argument that defaults to reduction with xor.
- The hash is always uint64.
- Integers will be casted to uint64 before performing the xor_sum reduction
- Floats will be upcasted to double and then bitcasted to uint64 before performing the xor_sum reduction
Pull Request resolved: https://github.com/pytorch/pytorch/pull/154149
Approved by: https://github.com/albanD
2025-07-23 22:28:03 +00:00
fduwjj
82f8e04f27
Update distributed maintainers ( #158900 )
...
I maintain couple components of distributed like devicemesh, c10d and PGNCCL, gloo, etc. Can I be marked not as emeritus? Thanks!
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158900
Approved by: https://github.com/albanD
2025-07-23 21:53:27 +00:00
PaliC
9c68c4d08f
[BE] Remove __reduce_deploy__ ( #158291 )
...
This PR removes the integration point torch.fx had with torch::deploy (and another minor change).
Note: This PR has some broken mypy errors, but I believe those should have been in the code base beforehand, and should be fixed in a separate PR
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158291
Approved by: https://github.com/albanD
ghstack dependencies: #158288 , #158290
2025-07-23 20:27:28 +00:00
PaliC
6ed2cb6ccd
[BE] Remove torch deploy | remove torch deploy specific files ( #158290 )
...
This PR removes specific files found in pytorch which are only used for torch::deploy. This is mostly testing code and a debugger.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158290
Approved by: https://github.com/albanD
ghstack dependencies: #158288
2025-07-23 20:27:28 +00:00
drisspg
691736ae07
Add kernel options to flex docs ( #158875 )
...
Fixes https://github.com/pytorch/pytorch/issues/158741
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158875
Approved by: https://github.com/BoyuanFeng , https://github.com/albanD
2025-07-23 19:05:19 +00:00
Panagiotis Kourdis
fd47401536
[doc] Updates to distributed.md for XCCL backend ( #155834 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/155834
Approved by: https://github.com/guangyey , https://github.com/AlannaBurke , https://github.com/d4l3k
Co-authored-by: Yu, Guangye <106960996+guangyey@users.noreply.github.com>
2025-07-22 21:01:43 +00:00
PyTorch MergeBot
6341311333
Revert "Add unified memory APIs for torch.accelerator ( #152932 )"
...
This reverts commit 2ad5c25cfc .
Reverted https://github.com/pytorch/pytorch/pull/152932 on behalf of https://github.com/ZainRizvi due to Very sorry but this is still breaking internally. @albanD would you be able to help get this past the finish line? D78496124 has more details on the failure and the workaround might be to do something like what's in D78684669. To validate the fixes internally, you can follow the instructions here to ghimport the changes: https://fburl.com/fixing-ghfirst-reverts ([comment](https://github.com/pytorch/pytorch/pull/138222#issuecomment-3100195370 ))
2025-07-22 01:01:41 +00:00
PyTorch MergeBot
4c18e85300
Revert "[BE] Remove torch deploy | remove torch deploy specific files ( #158290 )"
...
This reverts commit a6de309ca1 .
Reverted https://github.com/pytorch/pytorch/pull/158290 on behalf of https://github.com/ZainRizvi due to Sorry but this is breaking internally, see D78496147 for details. To validate your fixes internally, you can follow the instructions here: https://fburl.com/fixing-ghfirst-reverts ([comment](https://github.com/pytorch/pytorch/pull/158288#issuecomment-3099826158 ))
2025-07-21 23:17:39 +00:00
PyTorch MergeBot
920f26c761
Revert "[BE] Remove __reduce_deploy__ ( #158291 )"
...
This reverts commit 0b9fb91f17 .
Reverted https://github.com/pytorch/pytorch/pull/158291 on behalf of https://github.com/ZainRizvi due to Sorry but this is breaking internally, see D78496147 for details. To validate your fixes internally, you can follow the instructions here: https://fburl.com/fixing-ghfirst-reverts ([comment](https://github.com/pytorch/pytorch/pull/158288#issuecomment-3099826158 ))
2025-07-21 23:17:38 +00:00
Jane Xu
7cc5d03dfc
Document the rest of the specific optimizer module APIs ( #158669 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158669
Approved by: https://github.com/albanD
ghstack dependencies: #158483
2025-07-19 07:27:15 +00:00
Jane Xu
f73594164a
[BE] document Adadelta and Adagrad APIs properly ( #158483 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158483
Approved by: https://github.com/albanD
2025-07-19 07:27:15 +00:00
Svetlana Karslioglu
79e49efadd
Pull latest Sphinx theme ( #158595 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158595
Approved by: https://github.com/albanD
2025-07-18 18:46:47 +00:00
PyTorch MergeBot
9a7c2f1f64
Revert "Add torch compile force disable caches alias ( #158072 )"
...
This reverts commit 2ecf083b72 .
Reverted https://github.com/pytorch/pytorch/pull/158072 on behalf of https://github.com/jeffdaily due to fails on rocm, signal ignored while rocm was unstable ([comment](https://github.com/pytorch/pytorch/pull/158072#issuecomment-3086740829 ))
2025-07-18 04:58:24 +00:00
angelayi
66c9bc5062
[export] Add runnable code to export docs ( #158506 )
...
Preview: https://docs-preview.pytorch.org/pytorch/pytorch/158506/export.html
Yay I can add runnable code to export docs now
Also moved export API reference to a different file.
With these changes, we can start to consolidate the [export tutorial](https://docs.pytorch.org/tutorials/intermediate/torch_export_tutorial.html ) with the docs on pytorch docs. We just need to move the section on DDE and 0/1 specialization, and then I think we can delete the export tutorial.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158506
Approved by: https://github.com/pianpwk , https://github.com/svekars
2025-07-17 20:15:22 +00:00
Oguz Ulgen
2ecf083b72
Add torch compile force disable caches alias ( #158072 )
...
Bunch of people keep thinking current alias only disables inductor cache because it has the name inductor in it. lets globalize the name
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158072
Approved by: https://github.com/ezyang
2025-07-17 15:40:36 +00:00
Jiang, Yanbing
f4d8bc46c7
Enable TF32 as fp32 internal precision for matmul/linear/conv ( #157520 )
...
### Description
This PR is to enable TF32 as fp32 internal precision for matmul/linear/conv in `mkldnn backend`. Since we have refined fp32 precision API in https://github.com/pytorch/pytorch/pull/125888 , we can easily extend the API to support TF32 for `mkldnn backend`.
```
torch.backends.mkldnn.matmul.fp32_precision = 'tf32'
torch.backends.mkldnn.conv.fp32_precision = "tf32"
```
Related kernel update and UTs update are done. And the wrapper `bf32_on_and _off` is updated to `reduced_f32_on_and_off`, and it can run tests 3 times, one is reduced_f32 OFF, the other two are reduced_f32 ON (including `bf32 ON` and `tf32 ON`).
Pull Request resolved: https://github.com/pytorch/pytorch/pull/157520
Approved by: https://github.com/mingfeima , https://github.com/jansel
2025-07-17 08:57:34 +00:00
PaliC
0b9fb91f17
[BE] Remove __reduce_deploy__ ( #158291 )
...
This PR removes the integration point torch.fx had with torch::deploy (and another minor change).
Note: This PR has some broken mypy errors, but I believe those should have been in the code base beforehand, and should be fixed in a separate PR
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158291
Approved by: https://github.com/albanD
ghstack dependencies: #158288 , #158290
2025-07-17 05:56:26 +00:00
PaliC
a6de309ca1
[BE] Remove torch deploy | remove torch deploy specific files ( #158290 )
...
This PR removes specific files found in pytorch which are only used for torch::deploy. This is mostly testing code and a debugger.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158290
Approved by: https://github.com/albanD
ghstack dependencies: #158288
2025-07-17 05:56:18 +00:00
Yu, Guangye
2ad5c25cfc
Add unified memory APIs for torch.accelerator ( #152932 )
...
# Motivation
The following API will be put under torch.accelerator
- empty_cache
- max_memory_allocated
- max_memory_reserved
- memory_allocated
- memory_reserved
- memory_stats
- reset_accumulated_memory_stats
- reset_peak_memory_stats
Pull Request resolved: https://github.com/pytorch/pytorch/pull/152932
Approved by: https://github.com/albanD
ghstack dependencies: #138222
2025-07-17 01:56:01 +00:00
Yiming Zhou
a9ee4250d5
[4/n] Remove references to TorchScript in PyTorch docs ( #158317 )
...
Summary: jit.rst
Test Plan:
CI
Rollback Plan:
Differential Revision: D78309840
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158317
Approved by: https://github.com/svekars , https://github.com/zhxchen17
2025-07-16 20:01:34 +00:00
angelayi
1cc62c2cb9
[export] Update docs ( #157750 )
...
Preview: https://docs-preview.pytorch.org/pytorch/pytorch/157750/export.html
Changes:
* Rename draft_export.md -> export.draft_export.md for consistency.
* Removed non-strict section in export, instead pointed to programming model doc.
* Extended "Expressing Dynamism" section to include Dim hints, ShapeCollection, and AdditionalInputs.
* Removed Specialization section in favor of programming model doc
* Added pt2 archive doc
* Cleaned up sidebar
Pull Request resolved: https://github.com/pytorch/pytorch/pull/157750
Approved by: https://github.com/pianpwk
2025-07-16 19:53:12 +00:00
Jiang, Yanbing
900fba4c07
Update warning of TF32 ( #158209 )
...
Fixes #ISSUE_NUMBER
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158209
Approved by: https://github.com/jansel
2025-07-16 01:28:50 +00:00
Yiming Zhou
05dfd312cf
[3/n] Remove references to TorchScript in PyTorch docs ( #158315 )
...
Summary:
- cpp_index.rst
- fx.md
- jit_builtin_functions.rst
- jit_python_reference.md
- jit_unsupported.md
cpu_threading
large_scale_deployment
Test Plan:
CI
Rollback Plan:
Differential Revision: D78309320
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158315
Approved by: https://github.com/svekars , https://github.com/zhxchen17
2025-07-15 21:14:18 +00:00
Yiming Zhou
0640cfa38c
[2/n] Remove references to TorchScript in PyTorch docs ( #158306 )
...
Summary: Removed jit_language_reference.md
Test Plan:
CI
Rollback Plan:
Differential Revision: D78308133
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158306
Approved by: https://github.com/svekars , https://github.com/zhxchen17
2025-07-15 20:57:23 +00:00
Yiming Zhou
19625daf88
[1/n] Remove references to TorchScript in PyTorch docs ( #158305 )
...
Summary: Removed jit_language_reference_v2.md
Test Plan:
CI
Rollback Plan:
Differential Revision: D78308009
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158305
Approved by: https://github.com/jingsh , https://github.com/svekars
2025-07-15 20:16:53 +00:00
Ti-Tai Wang
5606c516fd
[ONNX] Remove legacy Dort ( #158258 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158258
Approved by: https://github.com/justinchuby , https://github.com/malfet
2025-07-15 19:14:06 +00:00
Jason Ansel
31326a9ad7
Fix typo in torch.set_float32_matmul_precision docs ( #158191 )
...
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158191
Approved by: https://github.com/Skylion007 , https://github.com/malfet
2025-07-12 18:23:11 +00:00
Ti-Tai Wang
2eff14c445
[ONNX] Delete torch.onnx.dynamo_export ( #158130 )
...
It's deprecated since torch==2.7.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158130
Approved by: https://github.com/justinchuby
2025-07-12 02:30:47 +00:00
Tristan Rice
0d77364ee3
dist2: cleanup non-option methods on PG (missing, timeouts) ( #158123 )
...
This updates the ProcessGroup.* API to include timeouts on all non-option based overloaded methods. This also adds 2 missing ones `alltoall_base` and `barrier`.
Following design in: https://docs.google.com/document/d/13R-1t_yESTvmAjcCN-wQjQQadIEu0JNIdS65uZawZzY/edit?tab=t.0#heading=h.3ctbqqopzc89
Test plan:
```
pytest test/distributed/test_dist2.py
```
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158123
Approved by: https://github.com/Skylion007 , https://github.com/fduwjj
2025-07-12 00:06:37 +00:00
Shivam Raikundalia
11d6ad8b2e
[Docs] Update PT2 Profiler Torch-Compiled Region Image ( #158066 )
...
Summary: In Pytorch 2.5 we added source code attribution to PT2 traces. Each Torch-Compiled Region will now have its frame id and frame compile id associated with it. Update the image in the doc and add a description of this in the doc itself
Test Plan:
{F1980179183}
Rollback Plan:
Differential Revision: D78118228
Pull Request resolved: https://github.com/pytorch/pytorch/pull/158066
Approved by: https://github.com/aaronenyeshi
2025-07-11 07:56:45 +00:00
zeshengzong
b4fc42ca80
Add torch.segment_reduce docs ( #154352 )
...
Fixes #153138
## Test Result

Pull Request resolved: https://github.com/pytorch/pytorch/pull/154352
Approved by: https://github.com/albanD
2025-07-11 06:16:38 +00:00
Jerry Zhang
11a86ad2fa
Remove pytorch quant docs since we are moving to torchao ( #157766 )
...
Summary:
att
Test Plan:
doc page generated from CI
Reviewers:
Subscribers:
Tasks:
Tags:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/157766
Approved by: https://github.com/Skylion007
2025-07-11 03:21:47 +00:00
Howard Huang
8532033679
RPC tutorial audit ( #157938 )
...
Fix [T228333894](https://www.internalfb.com/intern/tasks/?t=228333894 )
Pull Request resolved: https://github.com/pytorch/pytorch/pull/157938
Approved by: https://github.com/AlannaBurke
2025-07-10 14:15:37 +00:00
Dmitry Rogozhkin
b146ca74f0
docs: add get_default_backend_for_device to distributed documentation ( #156783 )
...
`torch.distributed.get_default_backend_for_device()` API was added to torch 2.6, but is still missing in distributed documentation. This commit addresses the gap.
CC: @guangyey, @EikanWang
Pull Request resolved: https://github.com/pytorch/pytorch/pull/156783
Approved by: https://github.com/guangyey , https://github.com/malfet
2025-07-10 05:11:30 +00:00
Tristan Rice
ed051c3084
torch.distributed: add initial _dist2 prototype API ( #157841 )
...
This adds the initial dist2 API as proposed in https://docs.google.com/document/d/13R-1t_yESTvmAjcCN-wQjQQadIEu0JNIdS65uZawZzY/edit?tab=t.0#heading=h.3ctbqqopzc89
This is a WIP experimental API and is a sandbox for a number of new features and quality of life improvements/changes to c10d.
Test plan:
```
pytest test/distributed/test_dist2.py
```
Docs
```
cd docs
make html
```

Pull Request resolved: https://github.com/pytorch/pytorch/pull/157841
Approved by: https://github.com/fduwjj
2025-07-09 23:40:43 +00:00