pytorch/torch/cuda
Nikita Shulga 43d746305c
Preserve CUDA gencode flags (#41212)
Summary:
Add `torch._C._cuda_getArchFlags()` that returns list of architecture `torch_cuda` were compiled with
Add `torch.cuda.get_arch_list()` and `torch.cuda.get_gencode_flags()` methods that returns architecture list and gencode flags PyTorch were compiled with
Print warning if some of GPUs is not compatible with any of the CUBINs

Pull Request resolved: https://github.com/pytorch/pytorch/pull/41173

Differential Revision: D22459998

Pulled By: malfet

fbshipit-source-id: 65d40ae29e54a0ba0f3f2da11b821fdb4d452d95
2020-07-09 17:34:50 -07:00
..
amp Make custom_fwd a no-op when not executed under autocast (#36171) 2020-06-23 10:23:02 -07:00
__init__.py Preserve CUDA gencode flags (#41212) 2020-07-09 17:34:50 -07:00
_utils.py Device name should be a string, not bytes (#40322) 2020-06-22 19:27:25 -07:00
comm.py Avoid initializing unnecessary tensors in nccl.reduce (#39688) 2020-06-14 10:11:32 -07:00
error.py
memory.py Move torch.cuda annotations inline (#40075) 2020-06-18 15:52:29 -07:00
nccl.py Avoid initializing unnecessary tensors in nccl.reduce (#39688) 2020-06-14 10:11:32 -07:00
nvtx.py Stop using ctypes to interface with CUDA libraries. (#33678) 2020-03-11 07:22:46 -07:00
profiler.py Move torch.cuda annotations inline (#40075) 2020-06-18 15:52:29 -07:00
random.py Move torch.cuda annotations inline (#40075) 2020-06-18 15:52:29 -07:00
sparse.py
streams.py Move _dummy_type to _utils.py (#40177) 2020-06-17 22:50:02 -07:00