transformers/tests/quantization
湛露先生 ebd2029483
Change GPUS to GPUs (#36945)
Signed-off-by: zhanluxianshen <zhanluxianshen@163.com>
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
2025-03-25 17:25:39 +01:00
..
aqlm_integration Skipping aqlm non working inference tests till fix merged (#34865) 2024-11-26 11:09:30 +01:00
autoawq fix xpu tests (#36656) 2025-03-17 15:57:49 +01:00
bitnet_integration Fix : BitNet tests (#34895) 2024-11-25 16:47:14 +01:00
bnb Fix Device map for bitsandbytes tests (#36800) 2025-03-19 11:57:13 +01:00
compressed_tensors Fix Expected output for compressed-tensors tests (#36425) 2025-02-26 21:17:24 +01:00
eetq_integration Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
fbgemm_fp8 Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
finegrained_fp8 Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
ggml Guard against unset resolved_archive_file (#35628) 2025-02-14 14:44:31 +01:00
gptq Fix typos (#36910) 2025-03-24 14:08:29 +00:00
higgs Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
hqq Fix hqq skipped modules and dynamic quant (#36821) 2025-03-20 15:31:49 +01:00
quanto_integration Changing the test model in Quanto kv cache (#36670) 2025-03-13 12:23:34 +01:00
quark_integration Support loading Quark quantized models in Transformers (#36372) 2025-03-20 15:40:51 +01:00
spqr_integration Efficient Inference Kernel for SpQR (#34976) 2025-02-13 16:22:58 +01:00
torchao_integration Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
vptq_integration Fix : VPTQ test (#35394) 2024-12-23 16:27:46 +01:00