pytorch/test/quantization/core
leslie-fang-intel 86e2d16ba0 [Inductor][Quant] Change the schema of QLinear Binary (#129049)
**Summary**
We change the schema of QLinear Binary, so it will be easier to enable the corresponding gemm template.

- Extra input of binary post-op is a tensor which needs to be an input node of autotuning, we need to move it at front of `output_scale` which is a scalar.
- We also move it at front of `bias`, since `bias` is optional tensor for this fusion, but `other` is a must to have for linear binary fusion.

**Test Plan**
```
python -u -m pytest -s -v test/quantization/core/test_quantized_op.py -k qlinear
python -u -m pytest -s -v test/inductor/test_mkldnn_pattern_matcher.py -k qlinear
```

Pull Request resolved: https://github.com/pytorch/pytorch/pull/129049
Approved by: https://github.com/jgong5, https://github.com/jansel
ghstack dependencies: #128825, #129048
2024-07-02 12:36:38 +00:00
..
experimental Enable torch.empty for float8 dtypes + deterministic mode + cpu (#128744) 2024-06-15 02:05:30 +00:00
__init__.py
test_backend_config.py
test_docs.py Revert "[BE][Easy] use pathlib.Path instead of dirname / ".." / pardir (#129374)" 2024-06-29 00:47:15 +00:00
test_quantized_functional.py
test_quantized_module.py
test_quantized_op.py [Inductor][Quant] Change the schema of QLinear Binary (#129049) 2024-07-02 12:36:38 +00:00
test_quantized_tensor.py [quant] Make per_group and per_token quant match torch.fake_quantize (#125781) 2024-05-14 18:18:54 +00:00
test_top_level_apis.py
test_utils.py
test_workflow_module.py
test_workflow_ops.py