pytorch/test/quantization/core
Xia, Weiwen a6d3da1835 [Quant] Add int8 linear op impl for quantization PT2E with Inductor. input is an int8 CPU tensor; weight is an int8 MdkldnnCPU tensor. (#105818)
**Summary**
Add a new onednn qlinear op for quantization PT2E with Inductor. input is an int8 CPU tensor and weight is an int8 MkldnnCPU tensor.

**Test plan**
python test/test_quantization.py -k test_qlinear_pt2e

Pull Request resolved: https://github.com/pytorch/pytorch/pull/105818
Approved by: https://github.com/jgong5, https://github.com/leslie-fang-intel, https://github.com/jerryzh168
2023-08-27 08:13:12 +00:00
..
experimental pt2: make aot_eager backend handle basic float8 operations (#107783) 2023-08-23 18:10:53 +00:00
__init__.py
test_backend_config.py
test_docs.py [BE]: Update ruff to 0.285 (#107519) 2023-08-22 23:16:38 +00:00
test_quantized_functional.py
test_quantized_module.py [BE] f-stringify torch/ and scripts (#105538) 2023-07-21 19:35:24 +00:00
test_quantized_op.py [Quant] Add int8 linear op impl for quantization PT2E with Inductor. input is an int8 CPU tensor; weight is an int8 MdkldnnCPU tensor. (#105818) 2023-08-27 08:13:12 +00:00
test_quantized_tensor.py Back out "Reland "Make adding buffers more like adding parameters (#104069)" (#106224)" (#106743) 2023-08-08 15:27:34 +00:00
test_top_level_apis.py
test_utils.py
test_workflow_module.py [BE] Enable ruff's UP rules and autoformat test/ (#105434) 2023-07-19 20:36:06 +00:00
test_workflow_ops.py