mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-25 23:38:59 +06:00
![]() * archive_file may not be specified When loading a pre-trained model from a gguf file, resolved_archive_file may not be set. Guard against that case in the safetensors availability check. * Remap partial disk offload to cpu for GGUF files GGUF files don't support disk offload so attempt to remap them to the CPU when device_map is auto. If device_map is anything else but None, raise a NotImplementedError. * Don't remap auto device_map and raise RuntimeError If device_map=auto and modules are selected for disk offload, don't attempt to map them to any other device. Raise a runtime error when a GGUF model is configured to map any modules to disk. --------- Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com> |
||
---|---|---|
.. | ||
aqlm_integration | ||
autoawq | ||
bitnet_integration | ||
bnb | ||
compressed_tensor | ||
eetq_integration | ||
fbgemm_fp8 | ||
finegrained_fp8 | ||
ggml | ||
gptq | ||
higgs | ||
hqq | ||
quanto_integration | ||
spqr_integration | ||
torchao_integration | ||
vptq_integration |