mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
parent
71963a6633
commit
7777db159f
@ -3179,7 +3179,7 @@ class PreTrainedTokenizerBase(SpecialTokensMixin):
|
||||
assert already_has_special_tokens and token_ids_1 is None, (
|
||||
"You cannot use ``already_has_special_tokens=False`` with this tokenizer. "
|
||||
"Please use a slow (full python) tokenizer to activate this argument."
|
||||
"Or set `return_special_token_mask=True` when calling the encoding method "
|
||||
"Or set `return_special_tokens_mask=True` when calling the encoding method "
|
||||
"to get the special tokens mask in any tokenizer. "
|
||||
)
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user