mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
Disallow Offload to disk for gguf files (#36933)
update Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com>
This commit is contained in:
parent
91455c1825
commit
2b8a15cc3f
@ -4327,6 +4327,12 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin, GenerationMixin, PushToHubMix
|
||||
"You cannot combine Quantization and loading a model from a GGUF file, try again by making sure you did not passed a `quantization_config` or that you did not load a quantized model from the Hub."
|
||||
)
|
||||
|
||||
if gguf_file and device_map is not None and "disk" in device_map.values():
|
||||
raise RuntimeError(
|
||||
"One or more modules is configured to be mapped to disk. Disk offload is not supported for models "
|
||||
"loaded from GGUF files."
|
||||
)
|
||||
|
||||
checkpoint_files, sharded_metadata = _get_resolved_checkpoint_files(
|
||||
pretrained_model_name_or_path=pretrained_model_name_or_path,
|
||||
subfolder=subfolder,
|
||||
|
Loading…
Reference in New Issue
Block a user