transformers/docs/source
Vivek Khandelwal 2963e196ee
Add support for loading GPTQ models on CPU (#26719)
* Add support for loading GPTQ models on CPU

Right now, we can only load the GPTQ Quantized model on the CUDA
device. The attribute `gptq_supports_cpu` checks if the current
auto_gptq version is the one which has the cpu support for the
model or not.
The larger variants of the model are hard to load/run/trace on
the GPU and that's the rationale behind adding this attribute.

Signed-Off By: Vivek Khandelwal <vivek@nod-labs.com>

* Update quantization.md

* Update quantization.md

* Update quantization.md
2023-10-31 13:45:23 +00:00
..
de [i18n-DE] contribute chapter (#26481) 2023-10-02 09:56:40 -07:00
en Add support for loading GPTQ models on CPU (#26719) 2023-10-31 13:45:23 +00:00
es docs: update link huggingface map (#26077) 2023-09-11 12:57:04 +01:00
fr Fix typos (#25936) 2023-09-04 11:15:12 +01:00
hi Hindi translation of pipeline_tutorial.md (#26837) 2023-10-25 11:21:49 -07:00
it [Docs] Fix un-rendered images (#25561) 2023-08-17 12:08:11 +02:00
ja Translating en/main_classes folder docs to Japanese 🇯🇵 (#26894) 2023-10-30 09:39:14 -07:00
ko 🌐 [i18n-KO] Translated big_models.md to Korean (#26245) 2023-10-12 15:00:12 -07:00
ms Remove broken links to s-JoL/Open-Llama (#27164) 2023-10-31 10:17:54 +00:00
pt docs: update link huggingface map (#26077) 2023-09-11 12:57:04 +01:00
te Added Telugu [te] translations (#26828) 2023-10-20 15:27:55 -07:00
zh 🌐 [i18n-ZH] Translate serialization.md into Chinese (#27076) 2023-10-30 08:50:29 -07:00
_config.py Adding evaluate to the list of libraries required in generated notebooks (#20850) 2022-12-21 14:04:08 +01:00