mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
[clean_up_tokenization_spaces
] Pl bart was failing, updating (#33735)
`clean_up_tokenization_spaces=True` for pl bart
This commit is contained in:
parent
294477aafb
commit
5f4420587a
@ -130,6 +130,7 @@ class PLBartTokenizer(PreTrainedTokenizer):
|
||||
tgt_lang=None,
|
||||
sp_model_kwargs: Optional[Dict[str, Any]] = None,
|
||||
additional_special_tokens=None,
|
||||
clean_up_tokenization_spaces=True,
|
||||
**kwargs,
|
||||
):
|
||||
# Mask token behave like a normal word, i.e. include the space before it
|
||||
@ -200,6 +201,7 @@ class PLBartTokenizer(PreTrainedTokenizer):
|
||||
tgt_lang=tgt_lang,
|
||||
additional_special_tokens=_additional_special_tokens,
|
||||
sp_model_kwargs=self.sp_model_kwargs,
|
||||
clean_up_tokenization_spaces=clean_up_tokenization_spaces,
|
||||
**kwargs,
|
||||
)
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user