Yuanyuan Chen
|
315ffdc1e4
|
[4/N] Apply ruff UP035 rule to python code (#164206)
Follows #164104
Pull Request resolved: https://github.com/pytorch/pytorch/pull/164206
Approved by: https://github.com/albanD
|
2025-10-01 19:05:53 +00:00 |
|
Xuehai Pan
|
5c4545f857
|
[BE][Easy] enable PYFMT for torch/[a-s]*/ (#138447)
Reproduce command:
```bash
ghstack checkout https://github.com/pytorch/pytorch/pull/138447
git checkout HEAD~1 torch/
lintrunner -a --take "PYFMT" --all-files
```
Pull Request resolved: https://github.com/pytorch/pytorch/pull/138447
Approved by: https://github.com/ezyang
|
2024-12-23 14:04:00 +00:00 |
|
Yu, Guangye
|
eb7adc3ae0
|
Refactor gpu trace to be device-agnostic (#121794)
# Motivation
Refactor gpu trace to be device-agnostic. gpu trace is usually used in runtime components, including Device, Stream, Event, Guard, and Allocator. It should be device-agnostic and can be shared among each device backend.
# Solution
move `_cuda_trace.py` to `_gpu_trace.py`, which makes each device backend owns their callback, respectively.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/121794
Approved by: https://github.com/jgong5, https://github.com/albanD, https://github.com/EikanWang, https://github.com/gujinghui
|
2024-03-30 13:04:38 +00:00 |
|
PyTorch MergeBot
|
968c4c4154
|
Revert "Refactor gpu trace to be device-agnostic (#121794)"
This reverts commit 74deacbf31.
Reverted https://github.com/pytorch/pytorch/pull/121794 on behalf of https://github.com/huydhn due to Sorry for reverting your change but it breaks ROCm jobs in trunk 74deacbf31, please help take a look and reland the change ([comment](https://github.com/pytorch/pytorch/pull/121794#issuecomment-2013674083))
|
2024-03-21 20:33:17 +00:00 |
|
Yu, Guangye
|
74deacbf31
|
Refactor gpu trace to be device-agnostic (#121794)
# Motivation
Refactor gpu trace to be device-agnostic. gpu trace is usually used in runtime components, including Device, Stream, Event, Guard, and Allocator. It should be device-agnostic and can be shared among each device backend.
# Solution
move `_cuda_trace.py` to `_gpu_trace.py`, which makes each device backend owns their callback, respectively.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/121794
Approved by: https://github.com/jgong5, https://github.com/albanD, https://github.com/EikanWang, https://github.com/gujinghui
|
2024-03-21 01:52:58 +00:00 |
|
PyTorch MergeBot
|
f9ed1c432d
|
Revert "Refactor gpu trace to be device-agnostic (#121794)"
This reverts commit 0ff1109e26.
Reverted https://github.com/pytorch/pytorch/pull/121794 on behalf of https://github.com/jeanschmidt due to Reverting to see if rocm trunk errors are related ([comment](https://github.com/pytorch/pytorch/pull/121794#issuecomment-2007519408))
|
2024-03-19 15:40:26 +00:00 |
|
Yu, Guangye
|
0ff1109e26
|
Refactor gpu trace to be device-agnostic (#121794)
# Motivation
Refactor gpu trace to be device-agnostic. gpu trace is usually used in runtime components, including Device, Stream, Event, Guard, and Allocator. It should be device-agnostic and can be shared among each device backend.
# Solution
move `_cuda_trace.py` to `_gpu_trace.py`, which makes each device backend owns their callback, respectively.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/121794
Approved by: https://github.com/jgong5, https://github.com/albanD, https://github.com/EikanWang, https://github.com/gujinghui
|
2024-03-19 06:02:28 +00:00 |
|