pytorch/benchmarks/operator_benchmark/benchmark_pytorch.py
Mingzhe Li b93f29ded3 add JIT path to the benchmark (#22309)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/22309

This diff enables PT operators to run with JIT mode. Users can control eager and JIT mode using the `use_jit` flag.

In this diff, we are putting operators in a loop and passed it to JIT. One extra step which wraps the operator with the `_consume` op is introduced to avoid dead code elimination optimization in JIT.  With that, the reported time includes the real operator execution time plus the `_consume` (directly return input, nothing else if happening inside) op.

Reviewed By: zheng-xq

Differential Revision: D16033082

fbshipit-source-id: e03be89fd5a505e44e81015dfc63db9cd76fb8a1
2019-07-03 17:18:03 -07:00

128 lines
4.5 KiB
Python

from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import benchmark_core
import torch
import cpp_extension # noqa
"""PyTorch performance microbenchmarks.
This module contains PyTorch-specific functionalities for performance
microbenchmarks.
"""
class TorchBenchmarkBase(object):
""" This is a base class used to create Pytorch operator benchmark.
module_name is the name of the operator being benchmarked.
test_name is the name (it's created by concatenating all the
inputs) of a specific test
"""
def __init__(self):
self.user_given_name = None
self._jit_forward = None
def forward(self):
pass
def _wrap_forward(self, foo):
""" The function passed to JIT trace must have at least one argument,
this function is to wrap the forward method to meet that requirement.
_consume op is used to avoid the dead-code-elimination optimization
in JIT.
"""
return torch.ops.operator_benchmark._consume(self.forward())
def _generate_jit_forward_graph(self):
""" generate a graph for the forward function via tracing
"""
func = torch.jit.trace(self._wrap_forward, torch.rand(1))
place_holder = torch.rand(1) # noqa
@torch.jit.script
def _jit_forward_graph(iters, place_holder):
# type: (int, Tensor)
result = torch.jit.annotate(torch.Tensor, None)
for _ in range(iters):
result = func(place_holder)
return result
return _jit_forward_graph
def module_name(self):
""" this is used to label the operator being benchmarked
"""
if self.user_given_name:
return self.user_given_name
return self.__class__.__name__
def set_module_name(self, name):
self.user_given_name = name
def test_name(self, **kargs):
""" this is a globally unique name which can be used to
label a specific test
"""
test_name_str = []
for key in kargs:
value = kargs[key]
test_name_str.append(
key + str(value if type(value) != bool else int(value)))
name = (self.module_name() + '_' +
'_'.join(test_name_str)).replace(" ", "")
return name
class PyTorchOperatorTestCase(object):
""" This class includes all the information needed to benchmark an operator.
op_bench: it's a user-defined class (child of TorchBenchmarkBase)
which includes input and operator, .etc
test_config: a namedtuple includes test_name, input_shape, tag, run_backward.
When run_backward is false, the run_forward method will be executed,
When run_backward is true, run_forward_eager and _output_mean will be
executed to generate output. Then, run_backward will be executed.
"""
def __init__(self, op_bench, test_config):
self.test_config = test_config
self.op_bench = op_bench
self.place_holder_tensor = torch.ones(1)
self.framework = "PyTorch"
def run_jit_forward(self, num_runs):
""" Run the forward path of an op with JIT mode
"""
if self.op_bench._jit_forward is None:
self.op_bench._jit_forward = self.op_bench._generate_jit_forward_graph()
self.op_bench._jit_forward(num_runs, self.place_holder_tensor)
def run_forward(self, num_runs):
""" Run the forward path of an op with eager mode
"""
for _ in range(num_runs):
self.output = self.op_bench.forward()
def _output_mean(self):
""" TODO (mingzhe): it is not necessary to sum up everything by myself,
torch.autograd.backward do take a gradient tensor. By default, it
is the same shape as your output tensor, with all 1s.
Mathematically, it is the same as if the output is summed together.
So we should be able to get ride of this method.
dummy function for gradient calculation
"""
self.mean = self.output.mean()
def run_backward(self, num_runs):
""" Run the backward path of an op in many iterations
"""
# TODO: can we use JIT here to reduce python overhead?
for _ in range(num_runs):
self.mean.backward(retain_graph=True)
def register_pytorch_op_test_case(op_bench, test_config):
test_case = PyTorchOperatorTestCase(op_bench, test_config)
benchmark_core._register_test(test_case)