mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-06 12:20:52 +01:00
Test Plan: revert-hammer Differential Revision: D20630760 Original commit changeset: 7d2f27aca6b1 fbshipit-source-id: 28ac92b3390651a4a67061d6ebf208515b9b9463
1183 lines
36 KiB
C++
1183 lines
36 KiB
C++
#ifdef TORCH_ENABLE_LLVM
|
|
#include "test/cpp/tensorexpr/test_base.h"
|
|
|
|
#include "test/cpp/tensorexpr/padded_buffer.h"
|
|
#include "test/cpp/tensorexpr/test_utils.h"
|
|
#include "torch/csrc/jit/tensorexpr/buffer.h"
|
|
#include "torch/csrc/jit/tensorexpr/eval.h"
|
|
#include "torch/csrc/jit/tensorexpr/function.h"
|
|
#include "torch/csrc/jit/tensorexpr/ir.h"
|
|
#include "torch/csrc/jit/tensorexpr/ir_printer.h"
|
|
#include "torch/csrc/jit/tensorexpr/llvm_codegen.h"
|
|
#include "torch/csrc/jit/tensorexpr/loopnest.h"
|
|
#include "torch/csrc/jit/tensorexpr/tensor.h"
|
|
|
|
#include <numeric>
|
|
|
|
namespace torch {
|
|
namespace jit {
|
|
using namespace torch::jit::tensorexpr;
|
|
using namespace torch::jit::tensorexpr;
|
|
|
|
using LLVMExprEval = ExprEval<LLVMCodeGen>;
|
|
|
|
// Typed tests, can't use gtest params here due to the way we instantiate tests.
|
|
#define TEST_LLVM_SCALAR_TYPES(_) \
|
|
_(uint8_t, Byte, 24) \
|
|
_(int8_t, Char, -20) \
|
|
_(int16_t, Short, 3332) \
|
|
_(int, Int, 123456) \
|
|
_(int64_t, Long, 2631563121321) \
|
|
_(float, Float, 0.122) \
|
|
_(double, Double, 0.21312) \
|
|
_(at::Half, Half, 0.128f)
|
|
|
|
#define IMM_TEST(Type, Name, Val) \
|
|
void testLLVM##Name##ImmTest() { \
|
|
KernelScope kernel_scope; \
|
|
auto a = Name##Imm::make(Val); \
|
|
LLVMExprEval cg(a); \
|
|
if (std::is_floating_point<decltype(Val)>()) { \
|
|
EXPECT_NEAR(cg.value<Type>(), Val, 0.1); \
|
|
} else { \
|
|
EXPECT_EQ(cg.value<Type>(), Val); \
|
|
} \
|
|
}
|
|
TEST_LLVM_SCALAR_TYPES(IMM_TEST)
|
|
#undef IMM_TEST
|
|
|
|
#define ADD_TEST(Type, Name, Val) \
|
|
void testLLVM##Name##AddTest() { \
|
|
KernelScope kernel_scope; \
|
|
auto a = Name##Imm::make(Val); \
|
|
auto b = Name##Imm::make(Val * 2); \
|
|
auto c = Add::make(a, b); \
|
|
LLVMExprEval cg(c); \
|
|
if (std::is_floating_point<decltype(Val)>()) { \
|
|
EXPECT_NEAR(cg.value<Type>(), Val * 3, 0.1); \
|
|
} else { \
|
|
EXPECT_EQ(cg.value<Type>(), Val * 3); \
|
|
} \
|
|
}
|
|
TEST_LLVM_SCALAR_TYPES(ADD_TEST)
|
|
#undef ADD_TEST
|
|
|
|
#define SUB_TEST(Type, Name, Val) \
|
|
void testLLVM##Name##SubTest() { \
|
|
KernelScope kernel_scope; \
|
|
auto a = Name##Imm::make(Val * 2); \
|
|
auto b = Name##Imm::make(Val); \
|
|
auto c = Sub::make(a, b); \
|
|
LLVMExprEval cg(c); \
|
|
if (std::is_floating_point<decltype(Val)>()) { \
|
|
EXPECT_NEAR(cg.value<Type>(), Val, 0.1); \
|
|
} else { \
|
|
EXPECT_EQ(cg.value<Type>(), Val); \
|
|
} \
|
|
}
|
|
TEST_LLVM_SCALAR_TYPES(SUB_TEST)
|
|
#undef SUB_TEST
|
|
|
|
#define MUL_TEST(Type, Name, Val) \
|
|
void testLLVM##Name##MulTest() { \
|
|
KernelScope kernel_scope; \
|
|
auto a = Name##Imm::make(Val); \
|
|
auto b = Name##Imm::make((Type)4); \
|
|
auto c = Mul::make(a, b); \
|
|
LLVMExprEval cg(c); \
|
|
if (std::is_floating_point<decltype(Val)>()) { \
|
|
EXPECT_NEAR(cg.value<Type>(), Val * 4, 0.1); \
|
|
} else { \
|
|
EXPECT_EQ(cg.value<Type>(), Val * 4); \
|
|
} \
|
|
}
|
|
TEST_LLVM_SCALAR_TYPES(MUL_TEST)
|
|
#undef MUL_TEST
|
|
|
|
#define DIV_TEST(Type, Name, Val) \
|
|
void testLLVM##Name##DivTest() { \
|
|
KernelScope kernel_scope; \
|
|
auto a = Name##Imm::make((Type)6); \
|
|
auto b = Name##Imm::make((Type)3); \
|
|
auto c = Div::make(a, b); \
|
|
LLVMExprEval cg(c); \
|
|
if (std::is_floating_point<decltype(Val)>()) { \
|
|
EXPECT_NEAR(cg.value<Type>(), 2, 0.1); \
|
|
} else { \
|
|
EXPECT_EQ(cg.value<Type>(), 2); \
|
|
} \
|
|
}
|
|
TEST_LLVM_SCALAR_TYPES(DIV_TEST)
|
|
#undef DIV_TEST
|
|
|
|
void testLLVMIntToFloatCastTest() {
|
|
KernelScope kernel_scope;
|
|
auto a = IntImm::make(2);
|
|
auto b = Cast::make(kFloat, a);
|
|
LLVMExprEval cg(b, {});
|
|
EXPECT_EQ(cg.value<float>(), 2.0);
|
|
}
|
|
|
|
void testLLVMFloatToIntCastTest() {
|
|
KernelScope kernel_scope;
|
|
auto a = FloatImm::make(2.0);
|
|
auto b = Cast::make(kInt, a);
|
|
LLVMExprEval cg(b);
|
|
EXPECT_EQ(cg.value<int>(), 2);
|
|
}
|
|
|
|
void testLLVMIntToLongCastTest() {
|
|
KernelScope kernel_scope;
|
|
auto a = IntImm::make(12345);
|
|
auto b = Cast::make(kLong, a);
|
|
LLVMExprEval cg(b);
|
|
EXPECT_EQ(cg.value<int64_t>(), 12345);
|
|
}
|
|
|
|
void testLLVMByteToCharCastTest() {
|
|
KernelScope kernel_scope;
|
|
auto a = ByteImm::make(250);
|
|
auto b = Cast::make(kChar, a);
|
|
LLVMExprEval cg(b);
|
|
EXPECT_EQ(cg.value<int8_t>(), (int8_t)250);
|
|
}
|
|
|
|
void testLLVMHalfToLongCastTest() {
|
|
KernelScope kernel_scope;
|
|
auto a = HalfImm::make(2.0);
|
|
auto b = Cast::make(kLong, a);
|
|
LLVMExprEval cg(b);
|
|
EXPECT_EQ(cg.value<int64_t>(), 2);
|
|
}
|
|
|
|
void testLLVMByteToDoubleCastTest() {
|
|
KernelScope kernel_scope;
|
|
auto a = ByteImm::make(2);
|
|
auto b = Cast::make(kDouble, a);
|
|
LLVMExprEval cg(b);
|
|
EXPECT_EQ(cg.value<double>(), 2);
|
|
}
|
|
|
|
void testLLVMLetTest01() {
|
|
KernelScope kernel_scope;
|
|
VarHandle x("x", kFloat);
|
|
ExprHandle value = ExprHandle(3.f);
|
|
ExprHandle body = ExprHandle(2.f) + (x * ExprHandle(3.f) + ExprHandle(4.f));
|
|
ExprHandle result = Let::make(x, ExprHandle(3.f), body);
|
|
LLVMExprEval cg(result, {});
|
|
EXPECT_EQ(cg.value<float>(), 2.f + (3.f * 3.f + 4.f));
|
|
}
|
|
|
|
void testLLVMLetTest02() {
|
|
KernelScope kernel_scope;
|
|
VarHandle x("x", kFloat);
|
|
VarHandle y("y", kFloat);
|
|
ExprHandle value = ExprHandle(3.f);
|
|
ExprHandle body =
|
|
ExprHandle(2.f) + (x * ExprHandle(3.f) + ExprHandle(4.f) * y);
|
|
ExprHandle e1 = Let::make(x, ExprHandle(3.f), body);
|
|
ExprHandle e2 = Let::make(y, ExprHandle(6.f), e1);
|
|
LLVMExprEval cg(e2, {});
|
|
EXPECT_EQ(cg.value<float>(), 2.f + (3.f * 3.f + 4.f * 6.f));
|
|
}
|
|
|
|
void testLLVMLetTestMultitype() {
|
|
KernelScope kernel_scope;
|
|
VarHandle x("x", kByte);
|
|
VarHandle y("y", kHalf);
|
|
ExprHandle value = ExprHandle((short)3);
|
|
ExprHandle body = ExprHandle((double)2.f) +
|
|
(x * ExprHandle(3) + ExprHandle((int64_t)4) * y);
|
|
ExprHandle e1 = Let::make(x, ExprHandle((uint8_t)3), body);
|
|
ExprHandle e2 = Let::make(y, ExprHandle((at::Half)6.f), e1);
|
|
LLVMExprEval cg(e2, {});
|
|
EXPECT_EQ(cg.value<double>(), 2.f + (3 * 3 + 4 * 6.f));
|
|
}
|
|
|
|
void testLLVMBufferTest() {
|
|
KernelScope kernel_scope;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {32});
|
|
std::vector<int32_t> v(5);
|
|
std::vector<void*> args({v.data()});
|
|
auto rv = IntImm::make(0);
|
|
LLVMExprEval cg(rv, {a});
|
|
EXPECT_EQ(cg.value<int>(args), 0);
|
|
}
|
|
|
|
void testLLVMBlockTest() {
|
|
KernelScope kernel_scope;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {32});
|
|
std::vector<int32_t> v = {1, 2};
|
|
std::vector<void*> args({v.data()});
|
|
|
|
auto block = Block::make({
|
|
Store::make(a, IntImm::make(0), IntImm::make(3), IntImm::make(1)),
|
|
Store::make(a, IntImm::make(1), IntImm::make(4), IntImm::make(1)),
|
|
Store::make(a, IntImm::make(0), IntImm::make(4), IntImm::make(1)),
|
|
});
|
|
|
|
LLVMCodeGen cg(block, {a});
|
|
EXPECT_EQ(cg.value<int>(args), 0);
|
|
EXPECT_EQ(v[0], 4);
|
|
EXPECT_EQ(v[1], 4);
|
|
}
|
|
|
|
void testLLVMLoadStoreTest() {
|
|
KernelScope kernel_scope;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {1});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {1});
|
|
std::vector<int32_t> a_buffer = {42};
|
|
std::vector<int32_t> b_buffer = {-11};
|
|
|
|
auto store = Store::make(
|
|
b,
|
|
IntImm::make(0),
|
|
Load::make(a, IntImm::make(0), IntImm::make(1)),
|
|
IntImm::make(1));
|
|
LLVMCodeGen cg(store, {a, b});
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data()});
|
|
EXPECT_EQ(cg.value<int>(args), 0);
|
|
EXPECT_EQ(a_buffer[0], 42);
|
|
EXPECT_EQ(b_buffer[0], 42);
|
|
}
|
|
|
|
void testLLVMIfThenElseTest() {
|
|
KernelScope kernel_scope;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {1});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {1});
|
|
Buffer c(VarHandle("C", kHandle), kInt, {1});
|
|
std::vector<int32_t> a_buffer = {42};
|
|
std::vector<int32_t> b_buffer = {-11};
|
|
std::vector<int32_t> c_buffer = {1};
|
|
|
|
auto store = Store::make(
|
|
b,
|
|
IntImm::make(0),
|
|
IfThenElse::make(
|
|
Load::make(c, IntImm::make(0), IntImm::make(1)), // cond
|
|
Load::make(a, IntImm::make(0), IntImm::make(1)), // then
|
|
IntImm::make(0)), // else
|
|
IntImm::make(1));
|
|
LLVMCodeGen cg(store, {a, b, c});
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
EXPECT_EQ(cg.value<int>(args), 0);
|
|
EXPECT_EQ(a_buffer[0], 42);
|
|
EXPECT_EQ(b_buffer[0], 42);
|
|
}
|
|
|
|
void testLLVMVecLoadStoreTest() {
|
|
KernelScope kernel_scope;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {1});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {1});
|
|
std::vector<int32_t> a_buffer = {1, 1, 1, 1};
|
|
std::vector<int32_t> b_buffer = {2, 2, 2, 2};
|
|
|
|
auto store = Store::make(
|
|
b,
|
|
Ramp::make(0, 1, 4),
|
|
Load::make(a, Ramp::make(0, 1, 4), Broadcast::make(IntImm::make(1), 4)),
|
|
Broadcast::make(IntImm::make(1), 4));
|
|
LLVMCodeGen cg(store, {a, b});
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data()});
|
|
EXPECT_EQ(cg.value<int>(args), 0);
|
|
EXPECT_EQ(a_buffer[0], 1);
|
|
EXPECT_EQ(a_buffer[1], 1);
|
|
EXPECT_EQ(a_buffer[2], 1);
|
|
EXPECT_EQ(a_buffer[3], 1);
|
|
EXPECT_EQ(b_buffer[0], 1);
|
|
EXPECT_EQ(b_buffer[1], 1);
|
|
EXPECT_EQ(b_buffer[2], 1);
|
|
EXPECT_EQ(b_buffer[3], 1);
|
|
}
|
|
|
|
#define FLOAT_INTRINSICS_TEST(Name, Lanes) \
|
|
void testLLVMVecFloat_##Name##Lane##Lanes##Test() { \
|
|
KernelScope kernel_scope; \
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {1}); \
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {1}); \
|
|
float val = 0.5f; \
|
|
std::vector<float> a_buffer(Lanes, val); \
|
|
std::vector<float> b_buffer(Lanes, val); \
|
|
auto store = Store::make( \
|
|
b, \
|
|
Ramp::make(0, 1, Lanes), \
|
|
Name(Load::make( \
|
|
a, \
|
|
Ramp::make(0, 1, Lanes), \
|
|
Broadcast::make(IntImm::make(1), Lanes))), \
|
|
Broadcast::make(IntImm::make(1), Lanes)); \
|
|
LLVMCodeGen cg(store, {a, b}); \
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data()}); \
|
|
float ref = std::Name(0.5f); \
|
|
EXPECT_EQ(cg.value<int>(args), 0); \
|
|
for (int i = 0; i < Lanes; i++) { \
|
|
EXPECT_FLOAT_EQ(a_buffer[i], val); \
|
|
} \
|
|
} // namespace jit
|
|
FLOAT_INTRINSICS_TEST(erf, 4)
|
|
FLOAT_INTRINSICS_TEST(erfc, 4)
|
|
FLOAT_INTRINSICS_TEST(acos, 4)
|
|
FLOAT_INTRINSICS_TEST(asin, 4)
|
|
FLOAT_INTRINSICS_TEST(atan, 4)
|
|
FLOAT_INTRINSICS_TEST(cosh, 4)
|
|
FLOAT_INTRINSICS_TEST(sinh, 4)
|
|
FLOAT_INTRINSICS_TEST(tanh, 4)
|
|
FLOAT_INTRINSICS_TEST(expm1, 4)
|
|
FLOAT_INTRINSICS_TEST(lgamma, 4)
|
|
FLOAT_INTRINSICS_TEST(erf, 8)
|
|
FLOAT_INTRINSICS_TEST(erfc, 8)
|
|
FLOAT_INTRINSICS_TEST(acos, 8)
|
|
FLOAT_INTRINSICS_TEST(asin, 8)
|
|
FLOAT_INTRINSICS_TEST(atan, 8)
|
|
FLOAT_INTRINSICS_TEST(cosh, 8)
|
|
FLOAT_INTRINSICS_TEST(sinh, 8)
|
|
FLOAT_INTRINSICS_TEST(tanh, 8)
|
|
FLOAT_INTRINSICS_TEST(expm1, 8)
|
|
FLOAT_INTRINSICS_TEST(lgamma, 8)
|
|
#undef FLOAT_INTRINSICS_TEST
|
|
|
|
#define DOUBLE_INTRINSICS_TEST(Name, Lanes) \
|
|
void testLLVMVecDouble_##Name##Lane##Lanes##Test() { \
|
|
KernelScope kernel_scope; \
|
|
Buffer a(VarHandle("A", kHandle), kDouble, {1}); \
|
|
Buffer b(VarHandle("B", kHandle), kDouble, {1}); \
|
|
float val = 0.5f; \
|
|
std::vector<double> a_buffer(Lanes, val); \
|
|
std::vector<double> b_buffer(Lanes, val); \
|
|
auto store = Store::make( \
|
|
b, \
|
|
Ramp::make(0, 1, Lanes), \
|
|
Name(Load::make( \
|
|
a, \
|
|
Ramp::make(0, 1, Lanes), \
|
|
Broadcast::make(IntImm::make(1), Lanes))), \
|
|
Broadcast::make(IntImm::make(1), Lanes)); \
|
|
LLVMCodeGen cg(store, {a, b}); \
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data()}); \
|
|
float ref = std::Name(0.5f); \
|
|
EXPECT_EQ(cg.value<int>(args), 0); \
|
|
for (int i = 0; i < Lanes; i++) { \
|
|
EXPECT_FLOAT_EQ(a_buffer[i], val); \
|
|
} \
|
|
} // namespace jit
|
|
DOUBLE_INTRINSICS_TEST(erf, 2)
|
|
DOUBLE_INTRINSICS_TEST(erfc, 2)
|
|
DOUBLE_INTRINSICS_TEST(acos, 2)
|
|
DOUBLE_INTRINSICS_TEST(asin, 2)
|
|
DOUBLE_INTRINSICS_TEST(atan, 2)
|
|
DOUBLE_INTRINSICS_TEST(cosh, 2)
|
|
DOUBLE_INTRINSICS_TEST(sinh, 2)
|
|
DOUBLE_INTRINSICS_TEST(tanh, 2)
|
|
DOUBLE_INTRINSICS_TEST(expm1, 2)
|
|
DOUBLE_INTRINSICS_TEST(lgamma, 2)
|
|
DOUBLE_INTRINSICS_TEST(erf, 4)
|
|
DOUBLE_INTRINSICS_TEST(erfc, 4)
|
|
DOUBLE_INTRINSICS_TEST(acos, 4)
|
|
DOUBLE_INTRINSICS_TEST(asin, 4)
|
|
DOUBLE_INTRINSICS_TEST(atan, 4)
|
|
DOUBLE_INTRINSICS_TEST(cosh, 4)
|
|
DOUBLE_INTRINSICS_TEST(sinh, 4)
|
|
DOUBLE_INTRINSICS_TEST(tanh, 4)
|
|
DOUBLE_INTRINSICS_TEST(expm1, 4)
|
|
DOUBLE_INTRINSICS_TEST(lgamma, 4)
|
|
#undef DOUBLE_INTRINSICS_TEST
|
|
|
|
void testLLVMVectorizerLoadStoreTest() {
|
|
KernelScope kernel_scope;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {1});
|
|
|
|
Tensor* c = Compute("c", {{4, "i"}}, [&](const VarHandle& i) {
|
|
return Load::make(a, i, 1);
|
|
});
|
|
|
|
Buffer c_buf(VarHandle(c->func_var()), kInt, {4});
|
|
LoopNest l({c});
|
|
Stmt* s = l.root_stmt();
|
|
l.vectorize(*dynamic_cast<Block*>(s)->stmts().begin());
|
|
|
|
EXPECT_TRUE(dynamic_cast<For*>(*dynamic_cast<Block*>(s)->stmts().begin()) == nullptr);
|
|
|
|
LLVMCodeGen cg(s, {a, c_buf});
|
|
|
|
std::vector<int> a_vec(4, 21);
|
|
std::vector<int> c_vec(4, 0);
|
|
std::vector<void*> args({a_vec.data(), c_vec.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
assertAllEqual(c_vec, 21);
|
|
}
|
|
|
|
void testLLVMMemcpyTest() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 32;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {N});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {N});
|
|
std::vector<int32_t> a_buffer(N, 42);
|
|
std::vector<int32_t> b_buffer(N, 0);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr =
|
|
For::make(i, 0, N, Store::make(b, i, Load::make(a, i, mask), mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 42);
|
|
assertAllEqual(b_buffer, 42);
|
|
}
|
|
|
|
void testLLVMBzeroTest() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 32;
|
|
Buffer b(VarHandle("B", kHandle), kInt, {N});
|
|
std::vector<int32_t> b_buffer(N, 11);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(i, 0, N, Store::make(b, i, IntImm::make(0), mask));
|
|
|
|
LLVMCodeGen cg(expr, {b});
|
|
|
|
std::vector<void*> args({b_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
assertAllEqual(b_buffer, 0);
|
|
}
|
|
|
|
void testLLVMElemwiseAdd() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {N});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {N});
|
|
Buffer c(VarHandle("C", kHandle), kInt, {N});
|
|
std::vector<int32_t> a_buffer(N, 41);
|
|
std::vector<int32_t> b_buffer(N, 1);
|
|
std::vector<int32_t> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Add::make(Load::make(a, i, mask), Load::make(b, i, mask)),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41);
|
|
assertAllEqual(b_buffer, 1);
|
|
assertAllEqual(c_buffer, 42);
|
|
}
|
|
|
|
void testLLVMElemwiseAddFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, 41);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(c, i, Load::make(a, i, mask) + Load::make(b, i, mask), mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41.0f);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 42.0f);
|
|
}
|
|
|
|
void testLLVMElemwiseLog10Float() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, 10.0f);
|
|
std::vector<float> b_buffer(N, 2.0f);
|
|
|
|
auto mask = Broadcast::make(IntImm::make(1), 4);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N / 4,
|
|
Store::make(
|
|
b,
|
|
Ramp::make(i * 4, 1, 4),
|
|
log10(Load::make(a, Ramp::make(i * 4, 1, 4), mask)),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 10.0f);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
}
|
|
|
|
void testLLVMElemwiseMaxInt() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {N});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {N});
|
|
Buffer c(VarHandle("C", kHandle), kInt, {N});
|
|
std::vector<int> a_buffer(N, 41);
|
|
std::vector<int> b_buffer(N, 1);
|
|
std::vector<int> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Max::make(Load::make(a, i, mask), Load::make(b, i, mask), false),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41);
|
|
assertAllEqual(b_buffer, 1);
|
|
assertAllEqual(c_buffer, 41);
|
|
}
|
|
|
|
void testLLVMElemwiseMinInt() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {N});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {N});
|
|
Buffer c(VarHandle("C", kHandle), kInt, {N});
|
|
std::vector<int> a_buffer(N, 41);
|
|
std::vector<int> b_buffer(N, 1);
|
|
std::vector<int> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Min::make(Load::make(a, i, mask), Load::make(b, i, mask), false),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41);
|
|
assertAllEqual(b_buffer, 1);
|
|
assertAllEqual(c_buffer, 1);
|
|
}
|
|
|
|
void testLLVMElemwiseMaxNumFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, 41);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Max::make(Load::make(a, i, mask), Load::make(b, i, mask), false),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41.0f);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 41.0f);
|
|
}
|
|
|
|
void testLLVMElemwiseMaxNumNaNFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, NAN);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Max::make(Load::make(a, i, mask), Load::make(b, i, mask), false),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 1.0f);
|
|
}
|
|
|
|
void testLLVMElemwiseMinNumFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, 41);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Min::make(Load::make(a, i, mask), Load::make(b, i, mask), false),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41.0f);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 1.0f);
|
|
}
|
|
|
|
void testLLVMElemwiseMinNumNaNFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, NAN);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Min::make(Load::make(a, i, mask), Load::make(b, i, mask), false),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 1.0f);
|
|
}
|
|
|
|
#if 1 // LLVM doesn't currently have implementations for maximum/minimum on x86
|
|
void testLLVMElemwiseMaximumFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, 41);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Max::make(Load::make(a, i, mask), Load::make(b, i, mask), true),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41.0f);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 41.0f);
|
|
}
|
|
|
|
void testLLVMElemwiseMaximumNaNFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, NAN);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Max::make(Load::make(a, i, mask), Load::make(b, i, mask), true),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
for (int i = 0; i < N; ++i) {
|
|
ASSERT_TRUE(std::isnan(a_buffer[i]));
|
|
ASSERT_TRUE(std::isnan(c_buffer[i]));
|
|
}
|
|
}
|
|
|
|
void testLLVMElemwiseMinimumFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, 41);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Min::make(Load::make(a, i, mask), Load::make(b, i, mask), true),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
assertAllEqual(a_buffer, 41.0f);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 1.0f);
|
|
}
|
|
|
|
void testLLVMElemwiseMinimumNaNFloat() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kFloat, {N});
|
|
std::vector<float> a_buffer(N, NAN);
|
|
std::vector<float> b_buffer(N, 1);
|
|
std::vector<float> c_buffer(N, 1);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
Min::make(Load::make(a, i, mask), Load::make(b, i, mask), true),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
for (int i = 0; i < N; ++i) {
|
|
ASSERT_TRUE(std::isnan(a_buffer[i]));
|
|
ASSERT_TRUE(std::isnan(c_buffer[i]));
|
|
}
|
|
}
|
|
#endif
|
|
|
|
void testLLVMCompareSelectIntEQ() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kInt, {N});
|
|
Buffer b(VarHandle("B", kHandle), kInt, {N});
|
|
Buffer c(VarHandle("C", kHandle), kInt, {N});
|
|
std::vector<int> a_buffer(N, 1);
|
|
std::vector<int> b_buffer(N, 1);
|
|
std::vector<int> c_buffer(N, 0);
|
|
std::vector<int> c_ref(N, 1);
|
|
|
|
for (int i = 0; i < N / 2; i++) {
|
|
b_buffer[i] = 0;
|
|
c_ref[i] = 0;
|
|
}
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
CompareSelect::make(
|
|
Load::make(a, i, mask),
|
|
Load::make(b, i, mask),
|
|
CompareSelectOperation::kEQ),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
|
|
assertAllEqual(a_buffer, 1);
|
|
for (int i = 0; i < N; i++) {
|
|
ASSERT_EQ(c_ref[i], c_buffer[i]);
|
|
}
|
|
}
|
|
|
|
void testLLVMCompareSelectFloatEQ() {
|
|
KernelScope kernel_scope;
|
|
constexpr int N = 1024;
|
|
Buffer a(VarHandle("A", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("B", kHandle), kFloat, {N});
|
|
Buffer c(VarHandle("C", kHandle), kInt, {N});
|
|
std::vector<float> a_buffer(N, 1.0f);
|
|
std::vector<float> b_buffer(N, 1.0f);
|
|
std::vector<int> c_buffer(N, 0);
|
|
|
|
auto mask = IntImm::make(1);
|
|
VarHandle i("i", kInt);
|
|
auto expr = For::make(
|
|
i,
|
|
0,
|
|
N,
|
|
Store::make(
|
|
c,
|
|
i,
|
|
CompareSelect::make(
|
|
Load::make(a, i, mask),
|
|
Load::make(b, i, mask),
|
|
CompareSelectOperation::kEQ),
|
|
mask));
|
|
|
|
LLVMCodeGen cg(expr, {a, b, c});
|
|
|
|
std::vector<void*> args({a_buffer.data(), b_buffer.data(), c_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
ASSERT_EQ(a_buffer.size(), N);
|
|
ASSERT_EQ(b_buffer.size(), N);
|
|
ASSERT_EQ(c_buffer.size(), N);
|
|
|
|
assertAllEqual(a_buffer, 1.0f);
|
|
assertAllEqual(b_buffer, 1.0f);
|
|
assertAllEqual(c_buffer, 1);
|
|
}
|
|
|
|
void testLLVMStoreFloat() {
|
|
KernelScope kernel_scope;
|
|
Buffer result(VarHandle("result", kHandle), kFloat, {1});
|
|
std::vector<float> result_buffer = {0.0f};
|
|
auto expr = Store::make(
|
|
result, IntImm::make(0), FloatImm::make(3.14f), IntImm::make(1));
|
|
LLVMCodeGen cg(expr, {result});
|
|
std::vector<void*> args({result_buffer.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
EXPECT_EQ(result_buffer[0], 3.14f);
|
|
}
|
|
|
|
void testLLVMSimpleMath01() {
|
|
KernelScope kernel_scope;
|
|
const int N = 1024;
|
|
Tensor* tensor = Compute("f", {{N, "i"}}, [](const VarHandle& i) {
|
|
return cast<float>(i * i + 1);
|
|
});
|
|
LoopNest l({tensor});
|
|
Stmt* stmt = l.root_stmt();
|
|
Buffer f_buf(VarHandle(tensor->func_var()), kFloat, {N});
|
|
LLVMCodeGen cg(stmt, {f_buf});
|
|
|
|
PaddedBuffer<float> f_v(N, "f_v");
|
|
std::vector<void*> args({f_v.data()});
|
|
int value = cg.value<int>(args);
|
|
ASSERT_EQ(value, 0);
|
|
PaddedBuffer<float> f_ref(N, "f_ref");
|
|
for (int i = 0; i < N; i++) {
|
|
f_ref(i) = i * i + 1;
|
|
}
|
|
ExpectAllNear(f_v, f_ref, 1e-5);
|
|
}
|
|
|
|
void testLLVMComputeMul() {
|
|
KernelScope kernel_scope;
|
|
const int N = 1024;
|
|
Buffer a(VarHandle("a", kHandle), kFloat, {N});
|
|
Buffer b(VarHandle("b", kHandle), kFloat, {N});
|
|
Tensor* c = Compute("c", {{N, "i"}}, [&](const VarHandle& i) {
|
|
return Load::make(a, i, 1) * Load::make(b, i, 1);
|
|
});
|
|
|
|
Buffer c_buf(VarHandle(c->func_var()), kFloat, {N});
|
|
LoopNest l({c});
|
|
Stmt* s = l.root_stmt();
|
|
|
|
LLVMCodeGen cg(s, {a, b, c_buf});
|
|
|
|
std::vector<float> a_vec(N, 21.0f);
|
|
std::vector<float> b_vec(N, 2.0f);
|
|
std::vector<float> c_vec(N, 0.0f);
|
|
std::vector<void*> args({a_vec.data(), b_vec.data(), c_vec.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
assertAllEqual(c_vec, 42.0f);
|
|
}
|
|
|
|
void testLLVMBroadcastAdd() {
|
|
KernelScope kernel_scope;
|
|
const int M = 32;
|
|
const int N = 1024;
|
|
Buffer a(VarHandle("a", kHandle), kFloat, {M, N});
|
|
Buffer b(VarHandle("b", kHandle), kFloat, {N});
|
|
Tensor* c = Compute(
|
|
"c", {{M, "i"}, {N, "j"}}, [&](const VarHandle& i, const VarHandle& j) {
|
|
ExprHandle mask(1);
|
|
return Load::make(a, i * N + j, mask) + Load::make(b, j, mask);
|
|
});
|
|
|
|
Buffer c_buf(VarHandle(c->func_var()), kFloat, {M, N});
|
|
LoopNest l({c});
|
|
Stmt* s = l.root_stmt();
|
|
|
|
LLVMCodeGen cg(s, {a, b, c_buf});
|
|
|
|
std::vector<float> av(M * N);
|
|
std::iota(av.begin(), av.end(), 0);
|
|
std::vector<float> bv(N);
|
|
std::iota(bv.begin(), bv.end(), 0);
|
|
std::vector<float> cv(M * N, 0);
|
|
std::vector<void*> args({av.data(), bv.data(), cv.data()});
|
|
ASSERT_EQ(cg.value<int>(args), 0);
|
|
|
|
for (int i = 0; i < M; i++) {
|
|
for (int j = 0; j < N; j++) {
|
|
ASSERT_EQ(cv[i * N + j], av[i * N + j] + bv[j]);
|
|
}
|
|
}
|
|
}
|
|
|
|
void testLLVMBitwiseOps() {
|
|
KernelScope kernel_scope;
|
|
auto a = IntImm::make(59);
|
|
auto b = IntImm::make(11);
|
|
auto c = IntImm::make(101);
|
|
auto d = IntImm::make(2);
|
|
|
|
ExprHandle f = (((a ^ (b << 1)) & c) >> 2) | d;
|
|
LLVMExprEval cg(f);
|
|
|
|
EXPECT_EQ(cg.value<int>(), 11);
|
|
}
|
|
|
|
void testLLVMDynamicShapeAdd() {
|
|
KernelScope kernel_scope;
|
|
auto testWithSize = [](int32_t size) {
|
|
VarHandle n("n", kInt);
|
|
Buffer a(VarHandle("a", kHandle), kFloat, {n});
|
|
Buffer b(VarHandle("b", kHandle), kFloat, {n});
|
|
Buffer c(VarHandle("c", kHandle), kFloat, {n});
|
|
VarHandle i("i", kInt);
|
|
Stmt* s = For::make(i, 0, n, Store::make(c, i, a(i) + b(i), 1));
|
|
std::vector<float> aData(size, 1.0f);
|
|
std::vector<float> bData(size, 2.0f);
|
|
std::vector<float> cData(size, 0.0f);
|
|
LLVMCodeGen cg(s, {a, b, c, n});
|
|
std::vector<void*> args({aData.data(), bData.data(), cData.data(), &size});
|
|
cg.value<float>(args);
|
|
ExpectAllNear(cData, std::vector<float>(size, 3.0f), 1e-7);
|
|
};
|
|
testWithSize(1);
|
|
testWithSize(16);
|
|
testWithSize(37);
|
|
}
|
|
|
|
void testLLVMBindDynamicShapeAdd() {
|
|
KernelScope kernel_scope;
|
|
auto testWithSize = [](int32_t size) {
|
|
VarHandle n("n", kInt);
|
|
Buffer a(VarHandle("a", kHandle), kFloat, {n});
|
|
Buffer b(VarHandle("b", kHandle), kFloat, {n});
|
|
Buffer c(VarHandle("c", kHandle), kFloat, {n});
|
|
VarHandle i("i", kInt);
|
|
Stmt* s = For::make(i, 0, n, Store::make(c, i, a(i) + b(i), 1));
|
|
std::vector<float> aData(size, 1.0f);
|
|
std::vector<float> bData(size, 2.0f);
|
|
std::vector<float> cData(size, 0.0f);
|
|
LLVMCodeGen cg(s, {a, b, c, n});
|
|
cg.call({aData, bData, cData, size});
|
|
ExpectAllNear(cData, std::vector<float>(size, 3.0f), 1e-7);
|
|
};
|
|
testWithSize(1);
|
|
testWithSize(16);
|
|
testWithSize(37);
|
|
}
|
|
|
|
void testLLVMTensorDynamicShapeAdd() {
|
|
KernelScope kernel_scope;
|
|
auto testWithSize = [](int32_t size) {
|
|
VarHandle n("n", kInt);
|
|
Buffer a(VarHandle("a", kHandle), kFloat, {n});
|
|
Buffer b(VarHandle("b", kHandle), kFloat, {n});
|
|
Tensor* c = Compute(
|
|
"c", {{n, "n"}}, [&](const VarHandle& i) { return a(i) + b(i); });
|
|
LoopNest l({c});
|
|
Stmt* s = l.root_stmt();
|
|
LLVMCodeGen cg(s, {a, b, c, n});
|
|
std::vector<float> aData(size, 1.0f);
|
|
std::vector<float> bData(size, 2.0f);
|
|
std::vector<float> cData(size, 0.0f);
|
|
cg.call({aData, bData, cData, size});
|
|
ExpectAllNear(cData, std::vector<float>(size, 3.0f), 1e-7);
|
|
};
|
|
testWithSize(1);
|
|
testWithSize(16);
|
|
testWithSize(37);
|
|
}
|
|
|
|
void testLLVMDynamicShape2D() {
|
|
KernelScope kernel_scope;
|
|
auto testWithSize = [](int32_t M, int32_t N) {
|
|
VarHandle m("m", kInt);
|
|
VarHandle n("n", kInt);
|
|
Buffer a(VarHandle("a", kHandle), kFloat, {m, n});
|
|
Buffer b(VarHandle("b", kHandle), kFloat, {m, n});
|
|
Tensor* c = Compute(
|
|
"c", {{m, "m"}, {n, "n"}}, [&](const VarHandle& i, const VarHandle& j) {
|
|
return a(i, j) + b(i, j);
|
|
});
|
|
LoopNest l({c});
|
|
Stmt* s = l.root_stmt();
|
|
LLVMCodeGen cg(s, {a, b, c, m, n});
|
|
std::vector<float> aData(M * N, 1.0f);
|
|
std::vector<float> bData(M * N, 2.0f);
|
|
std::vector<float> cData(M * N, 0.0f);
|
|
cg.call({aData, bData, cData, M, N});
|
|
ExpectAllNear(cData, std::vector<float>(M * N, 3.0f), 1e-7);
|
|
};
|
|
testWithSize(1, 8);
|
|
testWithSize(16, 32);
|
|
testWithSize(37, 11);
|
|
}
|
|
|
|
} // namespace jit
|
|
} // namespace torch
|
|
|
|
#endif // TORCH_ENABLE_LLVM
|