mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-02 11:11:05 +06:00
[from_pretrained
] Simpler code for peft (#25726)
* refactor complicated from pretrained for peft * nits * more nits * Update src/transformers/modeling_utils.py Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * make tests happy * fixup after merge --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
This commit is contained in:
parent
0a365c3e6a
commit
fecf08560c
@ -2391,33 +2391,23 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin, GenerationMixin, PushToHubMix
|
|||||||
else:
|
else:
|
||||||
commit_hash = getattr(config, "_commit_hash", None)
|
commit_hash = getattr(config, "_commit_hash", None)
|
||||||
|
|
||||||
if is_peft_available() and _adapter_model_path is None:
|
if is_peft_available():
|
||||||
maybe_adapter_model_path = find_adapter_config_file(
|
if _adapter_model_path is None:
|
||||||
pretrained_model_name_or_path,
|
_adapter_model_path = find_adapter_config_file(
|
||||||
cache_dir=cache_dir,
|
pretrained_model_name_or_path,
|
||||||
force_download=force_download,
|
cache_dir=cache_dir,
|
||||||
resume_download=resume_download,
|
force_download=force_download,
|
||||||
proxies=proxies,
|
resume_download=resume_download,
|
||||||
local_files_only=local_files_only,
|
proxies=proxies,
|
||||||
token=token,
|
local_files_only=local_files_only,
|
||||||
revision=revision,
|
token=token,
|
||||||
subfolder=subfolder,
|
revision=revision,
|
||||||
_commit_hash=commit_hash,
|
subfolder=subfolder,
|
||||||
)
|
_commit_hash=commit_hash,
|
||||||
elif is_peft_available() and _adapter_model_path is not None:
|
)
|
||||||
maybe_adapter_model_path = _adapter_model_path
|
if _adapter_model_path is not None and os.path.isfile(_adapter_model_path):
|
||||||
else:
|
with open(_adapter_model_path, "r", encoding="utf-8"):
|
||||||
maybe_adapter_model_path = None
|
_adapter_model_path = pretrained_model_name_or_path
|
||||||
|
|
||||||
has_adapter_config = maybe_adapter_model_path is not None
|
|
||||||
|
|
||||||
if has_adapter_config:
|
|
||||||
if _adapter_model_path is not None:
|
|
||||||
adapter_model_id = _adapter_model_path
|
|
||||||
else:
|
|
||||||
with open(maybe_adapter_model_path, "r", encoding="utf-8") as f:
|
|
||||||
adapter_model_id = pretrained_model_name_or_path
|
|
||||||
pretrained_model_name_or_path = json.load(f)["base_model_name_or_path"]
|
|
||||||
|
|
||||||
# change device_map into a map if we passed an int, a str or a torch.device
|
# change device_map into a map if we passed an int, a str or a torch.device
|
||||||
if isinstance(device_map, torch.device):
|
if isinstance(device_map, torch.device):
|
||||||
@ -3246,9 +3236,9 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin, GenerationMixin, PushToHubMix
|
|||||||
if quantization_method_from_config == QuantizationMethod.GPTQ:
|
if quantization_method_from_config == QuantizationMethod.GPTQ:
|
||||||
model = quantizer.post_init_model(model)
|
model = quantizer.post_init_model(model)
|
||||||
|
|
||||||
if has_adapter_config:
|
if _adapter_model_path is not None:
|
||||||
model.load_adapter(
|
model.load_adapter(
|
||||||
adapter_model_id,
|
_adapter_model_path,
|
||||||
adapter_name=adapter_name,
|
adapter_name=adapter_name,
|
||||||
revision=revision,
|
revision=revision,
|
||||||
token=token,
|
token=token,
|
||||||
|
Loading…
Reference in New Issue
Block a user