pytorch/torch/jit
Mingzhe Li 29881c7f02 Fix LSTM int8 quantization model size issue (#23577)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/23577

This diff is fixing a model size issue introduced in #23291. After that PR, the model size after in8 quantization is the same as that of the original unquantized model. The reason is that we save original weight for int8 quantization even when that's not needed anymore. This diff fixes that by only saving original weight for fp16 quantization path.

Reviewed By: llyfacebook

Differential Revision: D16557619

fbshipit-source-id: f924ae8d155a0d525b86a7440b3c7147d5bead0a
2019-08-02 13:38:30 -07:00
..
__init__.py Remove more uses of DimensionedTensorType 2019-08-01 21:19:28 -07:00
_logging.py Experimental logging/counters API (#18235) 2019-03-29 17:14:03 -07:00
_pickle.py Add all list specializations to pickler (#20191) 2019-05-10 17:14:42 -07:00
annotations.py Include recursive class compilations in error call stack (#23454) 2019-07-30 17:29:54 -07:00
frontend.py Fix frontend error message 2019-08-02 11:37:21 -07:00
quantized.py Fix LSTM int8 quantization model size issue (#23577) 2019-08-02 13:38:30 -07:00
supported_ops.py Make it possible for self.forward to return a ScriptMethod (#19217) 2019-04-24 11:14:34 -07:00