onnxruntime/include
Adrian Lizarraga b47e1e64d7
[QNN EP] Make offloading graph input/output quantization (to CPU) the default (#23368)
### Description
Makes the QNN provider option `offload_graph_io_quantization` enabled by
default. It was previously disabled by default.



### Motivation and Context
Enabling this option significantly decreases inference latency for many
models.
2025-02-04 11:42:46 -08:00
..
onnxruntime/core [QNN EP] Make offloading graph input/output quantization (to CPU) the default (#23368) 2025-02-04 11:42:46 -08:00