#include #include using namespace torch; using namespace torch::nn; using Catch::StartsWith; struct AGIUnit : CloneableModule { variable_list forward(variable_list) { return {}; } }; namespace test { struct AGIUnit : CloneableModule { variable_list forward(variable_list) { return {}; } }; struct AGIUnit2 : CloneableModule { AGIUnit2() : CloneableModule("Foo") {} variable_list forward(variable_list) { return {}; } }; } // namespace test TEST_CASE("module/training-mode") { auto model = make(Linear(3, 4)); REQUIRE(model->is_training()); SECTION("Enable eval mode") { model->eval(); REQUIRE(!model->is_training()); } SECTION("Enable train mode") { model->train(); REQUIRE(model->is_training()); } } TEST_CASE("module/zero-grad") { auto model = make(Linear(3, 4)); auto weights = Var(at::ones(at::CPU(at::kFloat), {8, 3})); auto loss = model->forward({weights}).front().sum(); backward(loss); for (auto& parameter : model->parameters()) { Variable grad = parameter.second.grad(); REQUIRE(grad.defined()); REQUIRE(grad.sum().toCFloat() != 0); } model->zero_grad(); for (auto& parameter : model->parameters()) { Variable grad = parameter.second.grad(); REQUIRE(grad.defined()); REQUIRE(grad.sum().toCFloat() == 0); } } TEST_CASE("module/name") { // CHECK instead of REQUIRE because demangling may fail. AGIUnit agi; // Call it twice just to make sure there are no bugs in the lazy // initialization semantics. CHECK(agi.name() == "AGIUnit"); CHECK(agi.name() == "AGIUnit"); SECTION("correctly demangled") { CHECK(test::AGIUnit().name() == "test::AGIUnit"); CHECK(test::AGIUnit2().name() == "Foo"); } } TEST_CASE("module/conversions", "[cuda]") { auto model = make(LSTM(/*in=*/128, /*out=*/64, /*layers=*/3, /*dropout=*/0.2)); SECTION("starts as float on CPU") { for (auto& parameter : model->parameters()) { REQUIRE(parameter.second.type().backend() == at::kCPU); REQUIRE(parameter.second.type().scalarType() == at::kFloat); } } SECTION("to(CUDA)") { model->cuda(); for (auto& parameter : model->parameters()) { REQUIRE(parameter.second.type().backend() == at::kCUDA); } } SECTION("to(CPU)") { model->to(at::kCPU); for (auto& parameter : model->parameters()) { REQUIRE(parameter.second.type().backend() == at::kCPU); } } SECTION("to(Int)") { model->to(at::kInt); for (auto& parameter : model->parameters()) { REQUIRE(parameter.second.type().scalarType() == at::kInt); } } SECTION("to(Double)") { model->to(at::kDouble); for (auto& parameter : model->parameters()) { REQUIRE(parameter.second.type().scalarType() == at::kDouble); } } SECTION("to(CUDA(Float))") { model->to(at::CUDA(at::kFloat)); for (auto& parameter : model->parameters()) { REQUIRE(parameter.second.type().backend() == at::kCUDA); REQUIRE(parameter.second.type().scalarType() == at::kFloat); } } } TEST_CASE("module/clone") { SECTION( "a module that does not override clone() throws when clone() is called") { struct UnCloneable : Module { variable_list forward(variable_list) override { return {}; } }; UnCloneable module; REQUIRE_THROWS_WITH( module.clone(), StartsWith("clone() has not been implemented")); } SECTION( "a module that overrides clone() does not throw when clone() is called ") { struct Cloneable : Module { variable_list forward(variable_list) override { return {}; } std::unique_ptr clone() const override { return nullptr; } }; Cloneable module; REQUIRE_NOTHROW(module.clone()); } SECTION("Cloning creates distinct parameters") { struct TestModel : public CloneableModule { TestModel() { add(make(Linear(10, 3)), "l1"); add(make(Linear(3, 5)), "l2"); add(make(Linear(5, 100)), "l3"); } variable_list forward(variable_list input) override { return input; } }; auto model = make(TestModel()); auto model2 = model->clone(); auto m1param = model->parameters(); auto m2param = model2->parameters(); for (auto& param : m1param) { REQUIRE(param.second.allclose(m2param[param.first])); param.second.data().mul_(2); } for (auto& param : m1param) { REQUIRE(!param.second.allclose(m2param[param.first])); } } }