mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-03 03:31:05 +06:00
Fix issue avoid-missing-comma found at https://codereview.doctor (#16768)
This commit is contained in:
parent
de8b06f9bf
commit
1b7de41a07
@ -173,7 +173,7 @@ class BertJapaneseTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def test_wordpiece_tokenizer(self):
|
def test_wordpiece_tokenizer(self):
|
||||||
vocab_tokens = ["[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは" "ばんは", "##こん", "##にちは", "##ばんは"]
|
vocab_tokens = ["[UNK]", "[CLS]", "[SEP]", "こんにちは", "こん", "にちは", "ばんは", "##こん", "##にちは", "##ばんは"]
|
||||||
|
|
||||||
vocab = {}
|
vocab = {}
|
||||||
for (i, token) in enumerate(vocab_tokens):
|
for (i, token) in enumerate(vocab_tokens):
|
||||||
@ -246,7 +246,7 @@ class BertJapaneseCharacterTokenizationTest(TokenizerTesterMixin, unittest.TestC
|
|||||||
)
|
)
|
||||||
|
|
||||||
def test_character_tokenizer(self):
|
def test_character_tokenizer(self):
|
||||||
vocab_tokens = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界" "、", "。"]
|
vocab_tokens = ["[UNK]", "[CLS]", "[SEP]", "こ", "ん", "に", "ち", "は", "ば", "世", "界", "、", "。"]
|
||||||
|
|
||||||
vocab = {}
|
vocab = {}
|
||||||
for (i, token) in enumerate(vocab_tokens):
|
for (i, token) in enumerate(vocab_tokens):
|
||||||
|
Loading…
Reference in New Issue
Block a user