pytorch/test/quantization
andrewor14 7da3c938cf [quant][be] Move QAT tests to its own file (#108061)
Test Plan:
python test/test_quantization.py TestQuantizePT2EQAT
python test/test_quantization.py TestQuantizePT2EQATModels

Reviewers: jerryzh168, kimishpatel

Subscribers: jerryzh168, kimishpatel, supriyar
Pull Request resolved: https://github.com/pytorch/pytorch/pull/108061
Approved by: https://github.com/jerryzh168
2023-09-15 18:34:44 +00:00
..
ao_migration
bc [BE] Enable ruff's UP rules and autoformat test/ (#105434) 2023-07-19 20:36:06 +00:00
core [Quant] Add int8 linear op impl for quantization PT2E with Inductor. input is an int8 CPU tensor; weight is an int8 MdkldnnCPU tensor. (#105818) 2023-08-27 08:13:12 +00:00
eager [pytorch][ao] Add torch.matmul in FloatFunctional/QFunctional (#106831) 2023-08-10 22:43:36 +00:00
fx Revert "[quant][pt2e] store scale/zero_point as tensor attributes to support serialization (#105894)" 2023-07-28 01:16:02 +00:00
jit Reland: Remove remaining global set_default_dtype calls from tests (#108088) 2023-09-07 03:04:34 +00:00
pt2e [quant][be] Move QAT tests to its own file (#108061) 2023-09-15 18:34:44 +00:00
serialized
__init__.py