transformers/tests/quantization/ggml
g-prz fe484726aa
Add falcon gguf (#33437)
* feat(gguf): add falcon q2 k

* fix(gguf): remove useless renaming

* feat(gguf): seperate falcon 7b and 40b

* feat(gguf): apply fixup

* fix(test): error rebase

* feat(gguf): add fp16 weight comparison for falcon

* feat(gguf): test weight of all layers

* test(gguf): add falcon 40b under skip decorator

* feat(gguf): quick example for extracting model size
2024-10-02 14:10:39 +02:00
..
__init__.py Loading GGUF files support (#30391) 2024-05-15 14:28:20 +02:00
test_ggml.py Add falcon gguf (#33437) 2024-10-02 14:10:39 +02:00