#pragma once #include #include #include namespace torch { namespace jit { using Stack = std::vector; enum OpCode : uint8_t; namespace mobile { struct Code; class Function { public: TORCH_API Function(c10::QualifiedName name); TORCH_API bool run(Stack& stack) const; c10::IValue operator()(Stack& stack) const; const std::string& name() const; TORCH_API const c10::QualifiedName& qualname() const; void append_instruction(OpCode op, int X, int N, int64_t dbg_handle = -1); bool append_operator( const std::string& name, const std::string& overload_name, const c10::optional& num_specified_args, int64_t model_version); /* TODO: T90339189 deprecate all v3 when v3 models are removed */ void append_constant(const c10::IValue& constant); void append_type(const c10::TypePtr& type); void set_register_size(size_t size); int64_t get_debug_handle(size_t pc) const; const std::shared_ptr get_code() const; void setSchema(c10::FunctionSchema schema); const at::optional& getSchema() const; // Returns the debug handle corresponding to where the execution // is halted due to exception. // If no corresponding debug handle is found then -1 is returned. int64_t getExceptionDebugHandle() const; private: c10::QualifiedName name_; std::shared_ptr code_; at::optional schema_; // (byte-code version 4+) }; } // namespace mobile } // namespace jit } // namespace torch