Junjie Bai
f54ab540af
Rename cuda_gpu_id to device_id in DeviceOption ( #12456 )
...
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/12456
codemod with 'Yes to all'
codemod -d . --extensions h,cc,cpp,cu,py,proto,pbtxt,pb.txt,config cuda_gpu_id device_id
Overload TextFormat::ParseFromString to do string replace when parsing from protobuf format
Reviewed By: Yangqing
Differential Revision: D10240535
fbshipit-source-id: 5e6992bec961214be8dbe26f16f5794154a22b25
2018-10-09 15:54:04 -07:00
Junjie Bai
ff608a9ff3
Back out "Revert D10123245: Back out "codemod cuda_gpu_id to device_id"" ( #12232 )
...
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/12232
Original commit changeset: fca91fea58b7
This adds proper modifications to the DeviceType <->DeviceOption conversion code added in D10033396
Reviewed By: jerryzh168
Differential Revision: D10132473
fbshipit-source-id: 801ef777e2950982cb47b48051b1471a0a91e64b
2018-10-01 21:54:52 -07:00
Rick Ratmansky
3010dc4208
Revert D10123245: Back out "codemod cuda_gpu_id to device_id"
...
Differential Revision:
D10123245
Original commit changeset: d83da8e00a12
fbshipit-source-id: fca91fea58b7df208edc2e218a1d514f9821ec7b
2018-10-01 12:22:36 -07:00
Yang Liu
7d7d336c45
Back out "codemod cuda_gpu_id to device_id"
...
Summary:
Original commit changeset: f5614a5d2607
D9986213 is causing Multifeed Aggregator a [huge performance different](https://our.intern.facebook.com/intern/ads/analyze_canary/412951953278781781/ ) and is blocking aggregator push since last Friday night: https://fburl.com/feedtools/b6izvwjz
We need to land this revert ASAP to unblock aggregator push.
Reviewed By: orionr
Differential Revision: D10123245
fbshipit-source-id: d83da8e00a1250f5d09811a0a587c127e377aab2
2018-10-01 11:31:14 -07:00
Junjie Bai
3eb5940cf5
codemod cuda_gpu_id to device_id ( #12022 )
...
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/12022
codemod -d . --extensions h,cc,cpp,cu,py,proto,pbtxt,pb.txt,config cuda_gpu_id device_id
codemod with 'Yes to all'
Reviewed By: orionr
Differential Revision: D9986213
fbshipit-source-id: f5614a5d26078817aee8caf79a494abfd6a95ff1
2018-09-27 20:24:53 -07:00
Dmytro Dzhulgakov
496c999f7d
[core] NUMA-aware pinned allocator
...
Using cudaHostRegister/Unregister instead of cudaMallocHost to move memory to a
specific NUMA node
2018-03-06 00:33:11 -08:00
Dmytro Dzhulgakov
9e71de398b
[core] Graph-level NUMA awareness in Caffe2
...
Adding NUMA awareness through numa_node_id in DeviceOption. Blobs of operators
with numa_node_id are allocated on corr. memory banks, using CPU pools with
NUMA affinity set to run operators.
2018-03-06 00:33:11 -08:00