pytorch/torch/quantization
Supriya Rao 11c32611d7 [quant] Support 4-bit embedding_bag operators using the dtype quint4x2 (#45752)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/45752

Use the torch.quint4x2 dtype to create 4-bit packed tensors in the previous PR.
These packed tensors can be directly consumed by the operator.
Serialization of the packed tensors is supported using torchbind custom class.
Module support will follow in a later PR.

Test Plan:
python test/test_quantization.py TestEmbeddingBagOps

Imported from OSS

Reviewed By: jerryzh168

Differential Revision: D24120996

fbshipit-source-id: 2639353b3343ebc69e058b5ba237d3fc56728e1c
2020-10-06 21:11:49 -07:00
..
fx [quant][graphmode][fx] Add warning for unsupported case (#45714) 2020-10-06 15:33:34 -07:00
__init__.py [quant][eagermode] Custom module support (#44835) 2020-09-23 15:39:40 -07:00
_correct_bias.py Remove py2 compatible future imports (#44735) 2020-09-16 12:55:57 -07:00
_equalize.py Fix type annotation errors in torch.functional (#43446) 2020-08-26 08:27:59 -07:00
_learnable_fake_quantize.py Remove py2 compatible future imports (#44735) 2020-09-16 12:55:57 -07:00
_numeric_suite.py Remove py2 compatible future imports (#44735) 2020-09-16 12:55:57 -07:00
custom_module_class_mappings.py [quant][graphmode][fx] Custom module support (#44766) 2020-09-22 17:11:46 -07:00
default_mappings.py [quant] ConvTranspose1d / ConvTranspose2d (#40371) 2020-09-14 14:25:06 -07:00
fake_quantize.py Enable type check for torch.quantization.fake_quantize (#45701) 2020-10-02 09:27:34 -07:00
fuse_modules.py Remove py2 compatible future imports (#44735) 2020-09-16 12:55:57 -07:00
fuser_method_mappings.py [quant][eagermode][refactor] Add set/get method for quantization and fusion mappings (#43990) 2020-09-10 21:29:39 -07:00
observer.py [quant] Support 4-bit embedding_bag operators using the dtype quint4x2 (#45752) 2020-10-06 21:11:49 -07:00
qconfig.py Remove py2 compatible future imports (#44735) 2020-09-16 12:55:57 -07:00
quant_type.py Remove py2 compatible future imports (#44735) 2020-09-16 12:55:57 -07:00
quantization_mappings.py [quant][graphmode][fx][eagermode] Add leaky relu support in quantization workflows (#45712) 2020-10-06 12:16:04 -07:00
quantize_fx.py [reland][quant][graphmode][fx] Merge all quantization mode (#45292) (#45672) 2020-10-01 15:47:11 -07:00
quantize_jit.py [quant][graphmode][jit][api] Expose preserved_attrs from finalize to convert_jit (#44490) 2020-09-22 19:37:25 -07:00
quantize.py [quant][eagermode] Custom module support (#44835) 2020-09-23 15:39:40 -07:00
stubs.py Factored out the default mappings 2019-10-03 11:52:21 -07:00