mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-06 12:20:52 +01:00
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/57441 debug info Previous diffs did not save operator name in debug info. For delegated backends that only idenfity op for profiling with debug handle, operator name should be stores as well. Furthermore to complete debug informaton also serialize function name. Test Plan: Existing lite interpreter and backend tests Existing lite interpreter and backend tests Imported from OSS Differential Revision: D28144581 D28144581 Reviewed By: raziel Pulled By: kimishpatel fbshipit-source-id: 415210f147530a53b444b07f1d6ee699a3570d99
580 lines
20 KiB
C++
580 lines
20 KiB
C++
#include <gtest/gtest.h>
|
|
#include <test/cpp/jit/test_utils.h>
|
|
#include <torch/csrc/jit/api/module.h>
|
|
#include <torch/csrc/jit/backends/backend_detail.h>
|
|
#include <torch/csrc/jit/mobile/import.h>
|
|
#include <torch/csrc/jit/serialization/import.h>
|
|
#include <torch/torch.h>
|
|
|
|
// Tests go in torch::jit
|
|
namespace torch {
|
|
namespace jit {
|
|
// NOLINTNEXTLINE(cppcoreguidelines-avoid-non-const-global-variables)
|
|
TEST(BackendTest, ToBackend) {
|
|
Module m("m");
|
|
m.define(R"(
|
|
def forward(self, x, h):
|
|
return self.accum(x, h), self.sub_accum(x, h)
|
|
|
|
def accum(self, x, h):
|
|
return x + h
|
|
|
|
def sub_accum(self, x, h):
|
|
return x - h
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(2.0 * torch::ones({}));
|
|
inputs.emplace_back(1.0 * torch::ones({}));
|
|
auto ref = m.forward(inputs).toTuple()->elements();
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
auto lm = torch::jit::detail::codegen_backend_module(
|
|
"test_backend", m, compile_spec, any_dict_ty);
|
|
// lowered module code:
|
|
/*
|
|
class test_backendLoweredModule(Module):
|
|
__parameters__ = []
|
|
__buffers__ = []
|
|
__processed_module : Any
|
|
__method_compile_spec : Dict[str, Any]
|
|
__backend : __torch__.torch.classes.__backends__.test_backend
|
|
__handles : Dict[str, Any]
|
|
def __create_backend(self: torch.jit.test_backendLoweredModule) -> None:
|
|
_0 =
|
|
__torch__.torch.classes.__backends__.test_backend.__new__(__torch__.torch.classes.__backends__.test_backend)
|
|
_1 = (_0).__init__()
|
|
self.__backend = _0
|
|
return None
|
|
def __getstate__(self: torch.jit.test_backendLoweredModule) ->
|
|
Tuple[Dict[str, Any], Any]: _2 = (self.__method_compile_spec,
|
|
self.__processed_module) return _2 def __setstate__(self:
|
|
torch.jit.test_backendLoweredModule, state: Tuple[Dict[str, Any], Any]) ->
|
|
None: self.__method_compile_spec = (state)[0] self.__processed_module =
|
|
(state)[1] _3 = (self).__create_backend() _4 =
|
|
(self.__backend).compile(self.__processed_module,
|
|
self.__method_compile_spec, ) self.__handles = _4 return None def
|
|
forward(self: torch.jit.test_backendLoweredModule, x: Tensor, h: Tensor) ->
|
|
Tuple[Tensor, Tensor]: _5 = uninitialized(Tensor) typed_inputs =
|
|
annotate(List[Any], [x, h]) _6 =
|
|
(self.__backend).execute((self.__handles)["forward"], typed_inputs, ) _7,
|
|
_8, = _6 _9 = isinstance(_7, Tensor) if _9: _10 = unchecked_cast(Tensor, _7)
|
|
else:
|
|
ops.prim.RaiseException("AssertionError: ")
|
|
_10 = _5
|
|
_11 = isinstance(_8, Tensor)
|
|
if _11:
|
|
_12 = unchecked_cast(Tensor, _8)
|
|
else:
|
|
ops.prim.RaiseException("AssertionError: ")
|
|
_12 = _5
|
|
return (_10, _12)
|
|
|
|
*/
|
|
auto res = lm.forward(inputs).toTuple()->elements();
|
|
AT_ASSERT(res[0].toTensor().equal(ref[0].toTensor()));
|
|
AT_ASSERT(res[1].toTensor().equal(ref[1].toTensor()));
|
|
}
|
|
|
|
// NOLINTNEXTLINE(cppcoreguidelines-avoid-non-const-global-variables)
|
|
TEST(BackendTest, ToBackendNotAvailable) {
|
|
Module m("m");
|
|
m.define(R"(
|
|
def forward(self, x, h):
|
|
return self.accum(x, h), self.sub_accum(x, h)
|
|
|
|
def accum(self, x, h):
|
|
return x + h
|
|
|
|
def sub_accum(self, x, h):
|
|
return x - h
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(2.0 * torch::ones({}));
|
|
inputs.emplace_back(1.0 * torch::ones({}));
|
|
auto ref = m.forward(inputs).toTuple()->elements();
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// Produce lowered module (backend not available).
|
|
// Exception is not thrown at this point.
|
|
auto lm = torch::jit::detail::codegen_backend_module(
|
|
"test_backend_unavailable", m, compile_spec, any_dict_ty);
|
|
// Validate exception is thrown when trying to execute and
|
|
// the backend is not available.
|
|
ASSERT_THROWS_WITH_MESSAGE(
|
|
lm.forward(inputs).toTuple()->elements(), "Backend is not available.");
|
|
}
|
|
|
|
// NOLINTNEXTLINE(cppcoreguidelines-avoid-non-const-global-variables)
|
|
TEST(BackendTest, TestCompiler) {
|
|
Module m("m");
|
|
m.define(R"(
|
|
def forward(self, x, h):
|
|
return x + h
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(2.0 * torch::ones({}));
|
|
inputs.emplace_back(1.0 * torch::ones({}));
|
|
auto ref = m.forward(inputs);
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
auto lm = torch::jit::detail::codegen_backend_module(
|
|
"backend_with_compiler_demo", m, compile_spec, any_dict_ty);
|
|
auto res = lm.forward(inputs);
|
|
AT_ASSERT(res.toTensor().equal(ref.toTensor()));
|
|
|
|
std::stringstream ss;
|
|
lm._save_for_mobile(ss);
|
|
auto mlm = _load_for_mobile(ss);
|
|
auto mres = mlm.forward(inputs);
|
|
AT_ASSERT(mres.toTensor().equal(ref.toTensor()));
|
|
}
|
|
|
|
// NOLINTNEXTLINE(cppcoreguidelines-avoid-non-const-global-variables)
|
|
TEST(BackendTest, TestCompilerNotSupport) {
|
|
Module m("m");
|
|
m.define(R"(
|
|
def forward(self, x, h):
|
|
return x * h
|
|
)");
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
ASSERT_THROWS_WITH_MESSAGE(
|
|
torch::jit::detail::codegen_backend_module(
|
|
"backend_with_compiler_demo", m, compile_spec, any_dict_ty),
|
|
"The node of aten::mul is not supported in this compiler. Source code:");
|
|
}
|
|
|
|
TEST(BackendTestDebugInfo, TestCompiler) {
|
|
Module m("m");
|
|
m.define(R"(
|
|
def forward(self, x, h):
|
|
return x + h
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(torch::rand({2, 4}));
|
|
inputs.emplace_back(torch::rand({13, 9}));
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
auto lm = torch::jit::detail::codegen_backend_module(
|
|
"backend_with_compiler_demo", m, compile_spec, any_dict_ty);
|
|
|
|
std::stringstream ss;
|
|
lm._save_for_mobile(ss, ExtraFilesMap(), true);
|
|
auto mlm = _load_for_mobile(ss);
|
|
std::string error_pattern = R"(
|
|
Module hierarchy:top(backend_with_compiler_demoLoweredModule).aten::add
|
|
Traceback of TorchScript (most recent call last):
|
|
File "<string>", line 5, in FunctionName_UNKNOWN
|
|
typed_inputs: List[Any] = [x, h, ]
|
|
if self.__backend.is_available() :
|
|
_0, = self.__backend.execute(self.__handles["forward"], typed_inputs)
|
|
~~~~~~~~~~~~~~~~~~~~~~ <--- HERE
|
|
assert isinstance(_0, Tensor)
|
|
return _0
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, h):
|
|
return x + h
|
|
~~~~~ <--- HERE
|
|
)";
|
|
ASSERT_THROWS_WITH_MESSAGE(mlm.forward(inputs), error_pattern);
|
|
}
|
|
|
|
TEST(BackendTestDebugInfo, TestExceptionStackForCompilerWithModuleHierarchy) {
|
|
Module a("A");
|
|
a.define(R"(
|
|
def forward(self, x, y):
|
|
return x + y
|
|
)");
|
|
Module b("B");
|
|
b.define(R"(
|
|
def forward(self, x):
|
|
return x + 2
|
|
)");
|
|
Module c("C");
|
|
c.register_module("A0", a);
|
|
c.register_module("B0", b);
|
|
c.define(R"(
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + self.B0.forward(x)
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(torch::rand({2, 4}));
|
|
inputs.emplace_back(torch::rand({13, 9}));
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
auto lm = torch::jit::detail::codegen_backend_module(
|
|
"backend_with_compiler_demo", c, compile_spec, any_dict_ty);
|
|
|
|
std::stringstream ss;
|
|
lm._save_for_mobile(ss, ExtraFilesMap(), true);
|
|
auto mlm = _load_for_mobile(ss);
|
|
std::string error_pattern = R"(
|
|
Module hierarchy:top(backend_with_compiler_demoLoweredModule).A0(A).aten::add
|
|
Traceback of TorchScript (most recent call last):
|
|
File "<string>", line 5, in FunctionName_UNKNOWN
|
|
typed_inputs: List[Any] = [x, y, ]
|
|
if self.__backend.is_available() :
|
|
_0, = self.__backend.execute(self.__handles["forward"], typed_inputs)
|
|
~~~~~~~~~~~~~~~~~~~~~~ <--- HERE
|
|
assert isinstance(_0, Tensor)
|
|
return _0
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + self.B0.forward(x)
|
|
~~~~~~~~~~~~~~~ <--- HERE
|
|
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return x + y
|
|
~~~~~ <--- HERE
|
|
)";
|
|
ASSERT_THROWS_WITH_MESSAGE(mlm.forward(inputs), error_pattern);
|
|
}
|
|
|
|
TEST(
|
|
BackendTestDebugInfo,
|
|
TestExceptionStackForCompilerWithTwoLevelModuleHierarchy) {
|
|
Module a("A");
|
|
a.define(R"(
|
|
def forward(self, x, y):
|
|
return x + y
|
|
)");
|
|
Module b("B");
|
|
b.register_module("A0", a);
|
|
b.define(R"(
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + 2
|
|
)");
|
|
Module c("C");
|
|
c.register_module("B0", b);
|
|
c.define(R"(
|
|
def forward(self, x, y):
|
|
return self.B0.forward(x, y) + 3
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(torch::rand({2, 4}));
|
|
inputs.emplace_back(torch::rand({13, 9}));
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
auto lm = torch::jit::detail::codegen_backend_module(
|
|
"backend_with_compiler_demo", c, compile_spec, any_dict_ty);
|
|
|
|
std::stringstream ss;
|
|
lm._save_for_mobile(ss, ExtraFilesMap(), true);
|
|
auto mlm = _load_for_mobile(ss);
|
|
/*
|
|
* Error stack throw will look like this:
|
|
* Module hierarchy:top(backend_with_compiler_demoLoweredModule).B0(B).A0(A)
|
|
* Traceback of TorchScript (most recent call last):
|
|
* File "<string>", line 5, in FunctionName_UNKNOWN
|
|
* typed_inputs: List[Any] = [x, y, ]
|
|
* if self.__backend.is_available() :
|
|
* _0, = self.__backend.execute(self.__handles["forward"],
|
|
* typed_inputs)
|
|
* ~~~~~~~~~~~~~~~~~~~~~~ <--- HERE
|
|
* assert isinstance(_0, Tensor)
|
|
* return _0
|
|
* File "<string>", line 3, in FunctionName_UNKNOWN
|
|
*
|
|
* def forward(self, x, y):
|
|
* return self.B0.forward(x, y) + 3
|
|
* ~~~~~~~~~~~~~~~ <--- HERE
|
|
*
|
|
* File "<string>", line 3, in FunctionName_UNKNOWN
|
|
*
|
|
* def forward(self, x, y):
|
|
* return self.A0.forward(x, y) + 2
|
|
* ~~~~~~~~~~~~~~~ <--- HERE
|
|
*
|
|
* File "<string>", line 3, in FunctionName_UNKNOWN
|
|
*
|
|
* def forward(self, x, y):
|
|
* return x + y
|
|
* ~~~~~ <--- HERE
|
|
*
|
|
*/
|
|
std::string error_pattern = R"(
|
|
Module hierarchy:top(backend_with_compiler_demoLoweredModule).B0(B).A0(A).aten::add
|
|
Traceback of TorchScript (most recent call last):
|
|
File "<string>", line 5, in FunctionName_UNKNOWN
|
|
typed_inputs: List[Any] = [x, y, ]
|
|
if self.__backend.is_available() :
|
|
_0, = self.__backend.execute(self.__handles["forward"], typed_inputs)
|
|
~~~~~~~~~~~~~~~~~~~~~~ <--- HERE
|
|
assert isinstance(_0, Tensor)
|
|
return _0
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return self.B0.forward(x, y) + 3
|
|
~~~~~~~~~~~~~~~ <--- HERE
|
|
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + 2
|
|
~~~~~~~~~~~~~~~ <--- HERE
|
|
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return x + y
|
|
~~~~~ <--- HERE
|
|
)";
|
|
ASSERT_THROWS_WITH_MESSAGE(mlm.forward(inputs), error_pattern);
|
|
}
|
|
|
|
TEST(BackendTestDebugInfo, TestExceptionStackForCompilerWithLoweredSubModule) {
|
|
std::shared_ptr<CompilationUnit> cu = std::make_shared<CompilationUnit>();
|
|
Module a("A");
|
|
a.define(R"(
|
|
def forward(self, x, y):
|
|
return x + y
|
|
)");
|
|
Module b("B");
|
|
b.define(R"(
|
|
def forward(self, x):
|
|
return x + 2
|
|
)");
|
|
Module c("C");
|
|
c.register_module("A0", a);
|
|
c.register_module("B0", b);
|
|
c.define(R"(
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + self.B0.forward(x)
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(torch::rand({2, 4}));
|
|
inputs.emplace_back(torch::rand({13, 9}));
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
IValue submodule = c.attr("A0");
|
|
Module current_sm = submodule.toModule();
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
auto lowered_submodule = torch::jit::detail::codegen_backend_module(
|
|
"backend_with_compiler_demo", current_sm, compile_spec, any_dict_ty);
|
|
|
|
c.type()->unsafeChangeAttributeType("A0", lowered_submodule.type());
|
|
c.setattr("A0", lowered_submodule._ivalue());
|
|
std::unordered_map<TypePtr, TypePtr> type_remap;
|
|
type_remap[a.type()] = lowered_submodule.type();
|
|
auto type_remap_fn = [&type_remap](TypePtr in) {
|
|
auto it = type_remap.find(in);
|
|
if (it == type_remap.end())
|
|
return in;
|
|
return it->second;
|
|
};
|
|
for (auto& fn : c.type()->methods()) {
|
|
auto method = c.get_method(fn->name());
|
|
auto graph = method.graph();
|
|
graph->remapTypes(type_remap_fn);
|
|
auto new_schema = fn->getSchema().cloneWithRemappedTypes(type_remap_fn);
|
|
fn->setSchema(new_schema);
|
|
}
|
|
|
|
std::stringstream ss;
|
|
c._save_for_mobile(ss, ExtraFilesMap(), true);
|
|
auto c_loaded = _load_for_mobile(ss);
|
|
std::string error_pattern = R"(
|
|
Module hierarchy:top(C).A0(backend_with_compiler_demoLoweredModule).aten::add
|
|
Traceback of TorchScript (most recent call last):
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + self.B0.forward(x)
|
|
~~~~~~~~~~~~~~~ <--- HERE
|
|
|
|
File "<string>", line 5, in FunctionName_UNKNOWN
|
|
typed_inputs: List[Any] = [x, y, ]
|
|
if self.__backend.is_available() :
|
|
_0, = self.__backend.execute(self.__handles["forward"], typed_inputs)
|
|
~~~~~~~~~~~~~~~~~~~~~~ <--- HERE
|
|
assert isinstance(_0, Tensor)
|
|
return _0
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return x + y
|
|
~~~~~ <--- HERE
|
|
)";
|
|
ASSERT_THROWS_WITH_MESSAGE(c_loaded.forward(inputs), error_pattern);
|
|
}
|
|
|
|
TEST(
|
|
BackendTestDebugInfo,
|
|
TestExceptionStackForCompilerWithSelectiveLoweredSubModule) {
|
|
std::shared_ptr<CompilationUnit> cu = std::make_shared<CompilationUnit>();
|
|
Module aa("AA");
|
|
aa.define(R"(
|
|
def forward(self, x, y):
|
|
return x + y
|
|
)");
|
|
Module a("A");
|
|
a.register_module("AA0", aa);
|
|
a.define(R"(
|
|
def forward(self, x, y):
|
|
return self.AA0.forward(x, y) + 3
|
|
)");
|
|
Module b("B");
|
|
b.define(R"(
|
|
def forward(self, x):
|
|
return x + 2
|
|
)");
|
|
Module c("C");
|
|
c.register_module("A0", a);
|
|
c.register_module("B0", b);
|
|
c.define(R"(
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + self.B0.forward(x)
|
|
)");
|
|
|
|
std::vector<IValue> inputs;
|
|
inputs.emplace_back(torch::rand({2, 4}));
|
|
inputs.emplace_back(torch::rand({13, 9}));
|
|
|
|
c10::Dict<IValue, IValue> compile_spec(StringType::get(), AnyType::get());
|
|
c10::Dict<IValue, IValue> fake_dict(StringType::get(), AnyType::get());
|
|
fake_dict.insert("", "");
|
|
compile_spec.insert("forward", fake_dict);
|
|
IValue submodule = c.attr("A0");
|
|
Module current_sm = submodule.toModule();
|
|
auto any_dict_ty = DictType::create(StringType::get(), AnyType::get());
|
|
// lowered module
|
|
auto lowered_submodule = torch::jit::detail::codegen_backend_module(
|
|
"backend_with_compiler_demo", current_sm, compile_spec, any_dict_ty);
|
|
|
|
c.type()->unsafeChangeAttributeType("A0", lowered_submodule.type());
|
|
c.setattr("A0", lowered_submodule._ivalue());
|
|
std::unordered_map<TypePtr, TypePtr> type_remap;
|
|
type_remap[a.type()] = lowered_submodule.type();
|
|
auto type_remap_fn = [&type_remap](TypePtr in) {
|
|
auto it = type_remap.find(in);
|
|
if (it == type_remap.end())
|
|
return in;
|
|
return it->second;
|
|
};
|
|
for (auto& fn : c.type()->methods()) {
|
|
auto method = c.get_method(fn->name());
|
|
auto graph = method.graph();
|
|
graph->remapTypes(type_remap_fn);
|
|
auto new_schema = fn->getSchema().cloneWithRemappedTypes(type_remap_fn);
|
|
fn->setSchema(new_schema);
|
|
}
|
|
|
|
std::stringstream ss;
|
|
c._save_for_mobile(ss, ExtraFilesMap(), true);
|
|
auto c_loaded = _load_for_mobile(ss);
|
|
/*
|
|
* Erro stack trace will look like this:
|
|
* Module hierarchy:top(C).A0(backend_with_compiler_demoLoweredModule).AA0(AA)
|
|
* Traceback of TorchScript (most recent call last):
|
|
* File "<string>", line 3, in FunctionName_UNKNOWN
|
|
*
|
|
* def forward(self, x, y):
|
|
* return self.A0.forward(x, y) + self.B0.forward(x)
|
|
* ~~~~~~~~~~~~~~~ <--- HERE
|
|
*
|
|
* File "<string>", line 5, in FunctionName_UNKNOWN
|
|
* typed_inputs: List[Any] = [x, y, ]
|
|
* if self.__backend.is_available() :
|
|
* _0, = self.__backend.execute(self.__handles["forward"],
|
|
* typed_inputs)
|
|
* ~~~~~~~~~~~~~~~~~~~~~~ <--- HERE
|
|
* assert isinstance(_0, Tensor)
|
|
* return _0
|
|
* File "<string>", line 3, in FunctionName_UNKNOWN
|
|
*
|
|
* def forward(self, x, y):
|
|
* return self.AA0.forward(x, y) + 3
|
|
* ~~~~~~~~~~~~~~~~ <--- HERE
|
|
*
|
|
* File "<string>", line 3, in FunctionName_UNKNOWN
|
|
*
|
|
* def forward(self, x, y):
|
|
* return x + y
|
|
* ~~~~~ <--- HERE
|
|
*
|
|
*
|
|
* */
|
|
std::string error_pattern = R"(
|
|
Module hierarchy:top(C).A0(backend_with_compiler_demoLoweredModule).AA0(AA).aten::add
|
|
Traceback of TorchScript (most recent call last):
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return self.A0.forward(x, y) + self.B0.forward(x)
|
|
~~~~~~~~~~~~~~~ <--- HERE
|
|
|
|
File "<string>", line 5, in FunctionName_UNKNOWN
|
|
typed_inputs: List[Any] = [x, y, ]
|
|
if self.__backend.is_available() :
|
|
_0, = self.__backend.execute(self.__handles["forward"], typed_inputs)
|
|
~~~~~~~~~~~~~~~~~~~~~~ <--- HERE
|
|
assert isinstance(_0, Tensor)
|
|
return _0
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return self.AA0.forward(x, y) + 3
|
|
~~~~~~~~~~~~~~~~ <--- HERE
|
|
|
|
File "<string>", line 3, in FunctionName_UNKNOWN
|
|
|
|
def forward(self, x, y):
|
|
return x + y
|
|
~~~~~ <--- HERE
|
|
)";
|
|
ASSERT_THROWS_WITH_MESSAGE(c_loaded.forward(inputs), error_pattern);
|
|
}
|
|
|
|
} // namespace jit
|
|
} // namespace torch
|