pytorch/test/quantization
kausik 4f60a2e39c Set correct output dtype for dequantize op during convert_pt2e in decomposed mode (#128953)
Earlier the signature of dequantize ops for decomposed quantized Tensor was changed for wider use-cases where the output dtype can be different from torch.float and needs to be passed during dequantization.
Please refer: https://github.com/pytorch/pytorch/pull/121450

However, setting of correct output dtype for dequantize ops was still missing in convert_pt2e flow.

This change enables the users to use PT2E quantization flow with non torch.float unquantized dtype, such as torch.bfloat16.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/128953
Approved by: https://github.com/jgong5, https://github.com/jerryzh168
2024-07-19 04:58:02 +00:00
..
ao_migration Enable UFMT on all of test/quantization/ao_migration &bc (#123994) 2024-04-13 06:36:10 +00:00
bc Enable UFMT on all of test/quantization/ao_migration &bc (#123994) 2024-04-13 06:36:10 +00:00
core Fix for HistogramObserver (#129387) 2024-07-02 15:41:44 +00:00
eager Support min/max carry over for eager mode from_float method (#127309) 2024-05-29 19:33:26 +00:00
fx [BE]: Update ruff to v0.4.4 (#125031) 2024-05-12 20:02:37 +00:00
jit Enable UFMT on all of test/quantization/jit &pt2e (#124010) 2024-04-14 06:07:23 +00:00
pt2e Set correct output dtype for dequantize op during convert_pt2e in decomposed mode (#128953) 2024-07-19 04:58:02 +00:00
serialized
__init__.py