pytorch/test/cpp/api/module.cpp
Peter Goldsborough c5de3314cf Add name() to C++ modules (#7409)
* Add name() to C++ modules

* Use RTTI to get module name by default

* Add functional.cpp to CMakeLists.txt

* Call typeid() inside name() instead of constructor

* Add tests and use default constructor
2018-05-10 08:52:38 -07:00

140 lines
3.7 KiB
C++

#include <catch.hpp>
#include <torch/torch.h>
using namespace torch;
using namespace torch::nn;
using Catch::StartsWith;
struct AGIUnit : CloneableModule<AGIUnit> {
variable_list forward(variable_list) {
return {};
}
};
namespace test {
struct AGIUnit : CloneableModule<AGIUnit> {
variable_list forward(variable_list) {
return {};
}
};
struct AGIUnit2 : CloneableModule<AGIUnit2> {
AGIUnit2() : CloneableModule<AGIUnit2>("Foo") {}
variable_list forward(variable_list) {
return {};
}
};
} // namespace test
TEST_CASE("module/training-mode") {
auto model = make(Linear(3, 4));
REQUIRE(model->is_training());
SECTION("Enable eval mode") {
model->eval();
REQUIRE(!model->is_training());
}
SECTION("Enable train mode") {
model->train();
REQUIRE(model->is_training());
}
}
TEST_CASE("module/zero-grad") {
auto model = make(Linear(3, 4));
auto weights = Var(at::ones(at::CPU(at::kFloat), {8, 3}));
auto loss = model->forward({weights}).front().sum();
backward(loss);
for (auto& parameter : model->parameters()) {
Variable grad = parameter.second.grad();
REQUIRE(grad.defined());
REQUIRE(grad.sum().toCFloat() != 0);
}
model->zero_grad();
for (auto& parameter : model->parameters()) {
Variable grad = parameter.second.grad();
REQUIRE(grad.defined());
REQUIRE(grad.sum().toCFloat() == 0);
}
}
TEST_CASE("module/name") {
AGIUnit agi;
// Call it twice just to make sure there are no bugs in the lazy
// initialization semantics.
REQUIRE(agi.name() == "AGIUnit");
REQUIRE(agi.name() == "AGIUnit");
SECTION("correctly demangled") {
REQUIRE(test::AGIUnit().name() == "test::AGIUnit");
REQUIRE(test::AGIUnit2().name() == "Foo");
}
}
TEST_CASE("module/conversions", "[cuda]") {
auto model = make(LSTM(128, 64).nlayers(3).dropout(0.2));
SECTION("starts as float on CPU") {
for (auto& parameter : model->parameters()) {
REQUIRE(parameter.second.type().backend() == at::kCPU);
REQUIRE(parameter.second.type().scalarType() == at::kFloat);
}
}
SECTION("to(CUDA)") {
model->cuda();
for (auto& parameter : model->parameters()) {
REQUIRE(parameter.second.type().backend() == at::kCUDA);
}
}
SECTION("to(CPU)") {
model->to(at::kCPU);
for (auto& parameter : model->parameters()) {
REQUIRE(parameter.second.type().backend() == at::kCPU);
}
}
SECTION("to(Int)") {
model->to(at::kInt);
for (auto& parameter : model->parameters()) {
REQUIRE(parameter.second.type().scalarType() == at::kInt);
}
}
SECTION("to(Double)") {
model->to(at::kDouble);
for (auto& parameter : model->parameters()) {
REQUIRE(parameter.second.type().scalarType() == at::kDouble);
}
}
SECTION("to(CUDA(Float))") {
model->to(at::CUDA(at::kFloat));
for (auto& parameter : model->parameters()) {
REQUIRE(parameter.second.type().backend() == at::kCUDA);
REQUIRE(parameter.second.type().scalarType() == at::kFloat);
}
}
}
TEST_CASE("module/clone") {
SECTION(
"a module that does not override clone() throws when clone() is called") {
struct UnCloneable : Module {
variable_list forward(variable_list) override {
return {};
}
};
UnCloneable module;
REQUIRE_THROWS_WITH(
module.clone(), StartsWith("clone() has not been implemented"));
}
SECTION(
"a module that overrides clone() does not throw when clone() is called ") {
struct Cloneable : Module {
variable_list forward(variable_list) override {
return {};
}
std::unique_ptr<Module> clone() const override {
return nullptr;
}
};
Cloneable module;
REQUIRE_NOTHROW(module.clone());
}
}