mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-02 19:21:31 +06:00
Update tokenization_openai.py
This commit is contained in:
parent
bb04446285
commit
562b998366
@ -92,6 +92,7 @@ class OpenAIGPTTokenizer(PreTrainedTokenizer):
|
||||
from spacy.lang.en import English
|
||||
_nlp = English()
|
||||
self.nlp = nlp.Defaults.create_tokenizer(_nlp)
|
||||
self.fix_text = ftfy.fix_text
|
||||
except ImportError:
|
||||
logger.warning("ftfy or spacy is not installed using BERT BasicTokenizer instead of SpaCy & ftfy.")
|
||||
self.nlp = BasicTokenizer(do_lower_case=True)
|
||||
|
Loading…
Reference in New Issue
Block a user