mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-07 12:21:27 +01:00
Summary: This PR added constant folding for quantize ops so that instead of storing fp32 weight in the quantized model, we'll get int8/int16 etc. weight Test Plan: python test/test_quantization.py TestQuantizePT2E.test_fold_quantize also will verify in executorch later Reviewers: Subscribers: Tasks: Tags: Differential Revision: [D49399210](https://our.internmc.facebook.com/intern/diff/D49399210) Pull Request resolved: https://github.com/pytorch/pytorch/pull/109343 Approved by: https://github.com/kimishpatel, https://github.com/jgong5 |
||
|---|---|---|
| .. | ||
| test_duplicate_dq.py | ||
| test_graph_utils.py | ||
| test_metadata_porting.py | ||
| test_quantize_pt2e_qat.py | ||
| test_quantize_pt2e.py | ||
| test_representation.py | ||
| test_x86inductor_quantizer.py | ||