mirror of
https://github.com/saymrwulf/transformers.git
synced 2026-05-14 20:58:08 +00:00
* add tests for linear shape behavior * fix linear shape behavior ended up adding the reshape at the end, after f8f8bf16_rowwise, because adding it directly after quantize_fp8_per_row caused f8f8bf16_rowwise to drop the seq_len dimension. (i.e., (17, 23, 1014) -> (17, 1024)) * save shape up front + comment |
||
|---|---|---|
| .. | ||
| aqlm_integration | ||
| autoawq | ||
| bnb | ||
| eetq_integration | ||
| fbgemm_fp8 | ||
| ggml | ||
| gptq | ||
| hqq | ||
| quanto_integration | ||
| torchao_integration | ||