transformers/tests/quantization
jiqing-feng 3a6ab46a0b
add gpt2 test on XPU (#37028)
* add gpt2 test on XPU

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* auto dtype has been fixed

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

* convert model to train mode

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>

---------

Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
2025-04-01 11:09:29 +02:00
..
aqlm_integration Skipping aqlm non working inference tests till fix merged (#34865) 2024-11-26 11:09:30 +01:00
autoawq [tests] remove cuda-only test marker in AwqConfigTest (#37032) 2025-03-31 11:53:02 +02:00
bitnet_integration Fix : BitNet tests (#34895) 2024-11-25 16:47:14 +01:00
bnb add gpt2 test on XPU (#37028) 2025-04-01 11:09:29 +02:00
compressed_tensors Fix Expected output for compressed-tensors tests (#36425) 2025-02-26 21:17:24 +01:00
eetq_integration Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
fbgemm_fp8 Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
finegrained_fp8 Skip FP8 linear tests For device capability < 9.0(#37008) 2025-03-27 12:38:37 +01:00
ggml Guard against unset resolved_archive_file (#35628) 2025-02-14 14:44:31 +01:00
gptq Fix typos (#36910) 2025-03-24 14:08:29 +00:00
higgs Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
hqq Fix hqq skipped modules and dynamic quant (#36821) 2025-03-20 15:31:49 +01:00
quanto_integration Changing the test model in Quanto kv cache (#36670) 2025-03-13 12:23:34 +01:00
quark_integration Support loading Quark quantized models in Transformers (#36372) 2025-03-20 15:40:51 +01:00
spqr_integration Efficient Inference Kernel for SpQR (#34976) 2025-02-13 16:22:58 +01:00
torchao_integration Change GPUS to GPUs (#36945) 2025-03-25 17:25:39 +01:00
vptq_integration Fix : VPTQ test (#35394) 2024-12-23 16:27:46 +01:00