diff --git a/src/transformers/models/t5/tokenization_t5.py b/src/transformers/models/t5/tokenization_t5.py index 922d9b67105..4ff1cf29c03 100644 --- a/src/transformers/models/t5/tokenization_t5.py +++ b/src/transformers/models/t5/tokenization_t5.py @@ -183,7 +183,7 @@ class T5Tokenizer(PreTrainedTokenizer): self._added_tokens_decoder = {} for i in range(len(extra_tokens)): self._added_tokens_decoder[len(self.sp_model) - 1 + extra_ids - i] = AddedToken( - f"", single_word=True, lstrip=True, rstrip=True, special=True + f"", single_word=False, lstrip=True, rstrip=True, special=True, normalized=False ) if legacy is None: