diff --git a/src/transformers/models/paligemma/modeling_paligemma.py b/src/transformers/models/paligemma/modeling_paligemma.py index f6b8f2ac46e..0d7eca8aa1d 100644 --- a/src/transformers/models/paligemma/modeling_paligemma.py +++ b/src/transformers/models/paligemma/modeling_paligemma.py @@ -196,7 +196,6 @@ class PaliGemmaPreTrainedModel(PreTrainedModel): _supports_cache_class = True _supports_quantized_cache = True _supports_static_cache = True - _supports_cache_class = True _supports_flash_attn_2 = True _supports_sdpa = True diff --git a/src/transformers/tokenization_utils_base.py b/src/transformers/tokenization_utils_base.py index 86e07a382f8..b9b24d681b2 100644 --- a/src/transformers/tokenization_utils_base.py +++ b/src/transformers/tokenization_utils_base.py @@ -2292,13 +2292,6 @@ class PreTrainedTokenizerBase(SpecialTokensMixin, PushToHubMixin): "Unable to load vocabulary from file. " "Please check that the provided vocabulary is accessible and not corrupted." ) - except RuntimeError as e: - if "sentencepiece_processor.cc" in str(e): - logger.info( - "Unable to load tokenizer model from SPM, loading from TikToken will be attempted instead." - "(SentencePiece RuntimeError: Tried to load SPM model with non-SPM vocab file).", - ) - return False if added_tokens_decoder != {} and max(list(added_tokens_decoder.keys())[-1], 0) > tokenizer.vocab_size: logger.info(