transformers/tests/quantization
Younes Belkada 6d4306160a
GGUF: Fix llama 3 GGUF (#31358)
* Create push-important-models.yml

* llama3 support for GGUF

* fixup

* Update src/transformers/integrations/ggml.py

* fix pre-tokenizer

* fix

* fix

* fix

* fix

* fix

* fix

* address final comment

* handle special tokens + add tests
2024-06-20 14:29:58 +02:00
..
aqlm_integration Cache: Static cache as a standalone object (#30476) 2024-04-30 16:37:19 +01:00
autoawq FIX / TST: Fix expected results on Mistral AWQ test (#30971) 2024-05-24 14:06:31 +02:00
bnb FIX / Quantization: Add extra validation for bnb config (#31135) 2024-05-30 11:45:03 +02:00
eetq_integration [FEAT]: EETQ quantizer support (#30262) 2024-04-22 20:38:58 +01:00
ggml GGUF: Fix llama 3 GGUF (#31358) 2024-06-20 14:29:58 +02:00
gptq 🚨 Remove dataset with restrictive license (#31452) 2024-06-17 17:56:51 +01:00
hqq Quantization / HQQ: Fix HQQ tests on our runner (#30668) 2024-05-06 11:33:52 +02:00
quanto_integration Fix quantized cache output (#31143) 2024-05-31 12:08:55 +02:00