diff --git a/src/transformers/modeling_utils.py b/src/transformers/modeling_utils.py index ba02bd6c910..8fdcf662fc1 100644 --- a/src/transformers/modeling_utils.py +++ b/src/transformers/modeling_utils.py @@ -1005,6 +1005,7 @@ def _get_resolved_checkpoint_files( user_agent: dict, revision: str, commit_hash: Optional[str], + is_remote_code: bool, # Because we can't determine this inside this function, we need it to be passed in transformers_explicit_filename: Optional[str] = None, ) -> Tuple[Optional[List[str]], Optional[Dict]]: """Get all the checkpoint filenames based on `pretrained_model_name_or_path`, and optional metadata if the @@ -1201,7 +1202,10 @@ def _get_resolved_checkpoint_files( "_commit_hash": commit_hash, **has_file_kwargs, } - if not has_file(pretrained_model_name_or_path, safe_weights_name, **has_file_kwargs): + if ( + not has_file(pretrained_model_name_or_path, safe_weights_name, **has_file_kwargs) + and not is_remote_code + ): Thread( target=auto_conversion, args=(pretrained_model_name_or_path,), @@ -4550,6 +4554,7 @@ class PreTrainedModel(nn.Module, ModuleUtilsMixin, PushToHubMixin, PeftAdapterMi user_agent=user_agent, revision=revision, commit_hash=commit_hash, + is_remote_code=cls._auto_class is not None, transformers_explicit_filename=transformers_explicit_filename, )