pytorch/test/quantization
Angel Li fe5ccb1a74 bf16 support for per tensor backward (#165362)
Adding bf16 for the backward pass of `torch._fake_quantize_learnable_per_tensor_affine()`.

Note that for testing, we modified the seed to avoid increasing tolerance due to cases where difference in Python vs CPP downcasting causes tensor mismatches. (e.g. 27.87704 vs  27.8408 before downcasting, 27.7500 vs 27.8750 after downcasting for Python vs CPP op)

Pull Request resolved: https://github.com/pytorch/pytorch/pull/165362
Approved by: https://github.com/andrewor14
2025-10-16 17:47:01 +00:00
..
ao_migration [BE][PYFMT] migrate PYFMT for test/[i-z]*/ to ruff format (#144556) 2025-07-29 03:26:09 +00:00
bc Add __main__ guards to quantization tests (#154728) 2025-06-10 19:46:07 +00:00
core bf16 support for per tensor backward (#165362) 2025-10-16 17:47:01 +00:00
eager Enable ruff rule E721 (#165162) 2025-10-13 01:48:55 +00:00
fx Enable ruff rule E721 (#165162) 2025-10-13 01:48:55 +00:00
jit [BE][PYFMT] migrate PYFMT for test/[i-z]*/ to ruff format (#144556) 2025-07-29 03:26:09 +00:00
pt2e Enable ruff rule E721 (#165162) 2025-10-13 01:48:55 +00:00
serialized [ao] fix incorrect integer cast on histogram observer bounds (#90355) 2022-12-12 20:30:44 +00:00
__init__.py