transformers/tests/quantization
Penut Chen 1c122a46dc
Support dequantizing GGUF FP16 format (#31783)
* support gguf fp16

* support gguf bf16 with pytorch

* add gguf f16 test

* remove bf16
2024-07-24 17:59:59 +02:00
..
aqlm_integration Cache: Static cache as a standalone object (#30476) 2024-04-30 16:37:19 +01:00
autoawq Skip tests properly (#31308) 2024-06-26 21:59:08 +01:00
bnb Skip tests properly (#31308) 2024-06-26 21:59:08 +01:00
eetq_integration [FEAT]: EETQ quantizer support (#30262) 2024-04-22 20:38:58 +01:00
fbgemm_fp8 Add new quant method (#32047) 2024-07-22 20:21:59 +02:00
ggml Support dequantizing GGUF FP16 format (#31783) 2024-07-24 17:59:59 +02:00
gptq 🚨 Remove dataset with restrictive license (#31452) 2024-06-17 17:56:51 +01:00
hqq Quantization / HQQ: Fix HQQ tests on our runner (#30668) 2024-05-06 11:33:52 +02:00
quanto_integration Skip tests properly (#31308) 2024-06-26 21:59:08 +01:00