pytorch/test/cpp_extensions/cuda_extension.cu
Peter Goldsborough 008ba18c5b Improve CUDA extension support (#5324)
* Also pass torch includes to nvcc build

* Export ATen/cuda headers with install

* Refactor flags common to C++ and CUDA

* Improve tests for C++/CUDA extensions

* Export .cuh files under THC

* Refactor and clean cpp_extension.py slightly

* Include ATen in cuda extension test

* Clarifying comment in cuda_extension.cu

* Replace cuda_extension.cu with cuda_extension_kernel.cu in setup.py

* Copy compile args in C++ extension and add second kernel

* Conditionally add -std=c++11 to cuda_flags

* Also export cuDNN headers

* Add comment about deepcopy
2018-02-23 10:15:30 -05:00

30 lines
1.1 KiB
Plaintext

// NOTE: This is a copy of cuda_extension_kernel.cu. It's kept here to test
// collision handling when a C++ file and CUDA file share the same filename.
// Setuptools can't deal with this at all, so the setup.py-based test uses
// cuda_extension_kernel.cu and the JIT test uses this file. Symlinks don't
// work well on Windows, so this is the most thorough solution right now.
#include <cuda.h>
#include <cuda_runtime.h>
#include <ATen/ATen.h>
__global__ void sigmoid_add_kernel(
const float* __restrict__ x,
const float* __restrict__ y,
float* __restrict__ output,
const int size) {
const int index = blockIdx.x * blockDim.x + threadIdx.x;
if (index < size) {
const float sigmoid_x = 1.0f / (1.0f + __expf(-x[index]));
const float sigmoid_y = 1.0f / (1.0f + __expf(-y[index]));
output[index] = sigmoid_x + sigmoid_y;
}
}
void sigmoid_add_cuda(const float* x, const float* y, float* output, int size) {
const int threads = 1024;
const int blocks = (size + threads - 1) / threads;
sigmoid_add_kernel<<<blocks, threads>>>(x, y, output, size);
}