mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-02 03:01:07 +06:00
[ci] Close #4481
This commit is contained in:
parent
a3af8e86cb
commit
865d4d595e
@ -26,7 +26,7 @@ from transformers.tokenization_bert_japanese import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
from .test_tokenization_common import TokenizerTesterMixin
|
from .test_tokenization_common import TokenizerTesterMixin
|
||||||
from .utils import custom_tokenizers, slow
|
from .utils import custom_tokenizers
|
||||||
|
|
||||||
|
|
||||||
@custom_tokenizers
|
@custom_tokenizers
|
||||||
@ -126,7 +126,6 @@ class BertJapaneseTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
|
|||||||
|
|
||||||
self.assertListEqual(tokenizer.tokenize("こんばんは こんばんにちは こんにちは"), ["こん", "##ばんは", "[UNK]", "こんにちは"])
|
self.assertListEqual(tokenizer.tokenize("こんばんは こんばんにちは こんにちは"), ["こん", "##ばんは", "[UNK]", "こんにちは"])
|
||||||
|
|
||||||
@slow
|
|
||||||
def test_sequence_builders(self):
|
def test_sequence_builders(self):
|
||||||
tokenizer = self.tokenizer_class.from_pretrained("bert-base-japanese")
|
tokenizer = self.tokenizer_class.from_pretrained("bert-base-japanese")
|
||||||
|
|
||||||
@ -141,6 +140,7 @@ class BertJapaneseTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
|
|||||||
assert encoded_pair == [2] + text + [3] + text_2 + [3]
|
assert encoded_pair == [2] + text + [3] + text_2 + [3]
|
||||||
|
|
||||||
|
|
||||||
|
@custom_tokenizers
|
||||||
class BertJapaneseCharacterTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
|
class BertJapaneseCharacterTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
|
||||||
|
|
||||||
tokenizer_class = BertJapaneseTokenizer
|
tokenizer_class = BertJapaneseTokenizer
|
||||||
@ -187,7 +187,6 @@ class BertJapaneseCharacterTokenizationTest(TokenizerTesterMixin, unittest.TestC
|
|||||||
|
|
||||||
self.assertListEqual(tokenizer.tokenize("こんにちほ"), ["こ", "ん", "に", "ち", "[UNK]"])
|
self.assertListEqual(tokenizer.tokenize("こんにちほ"), ["こ", "ん", "に", "ち", "[UNK]"])
|
||||||
|
|
||||||
@slow
|
|
||||||
def test_sequence_builders(self):
|
def test_sequence_builders(self):
|
||||||
tokenizer = self.tokenizer_class.from_pretrained("bert-base-japanese-char")
|
tokenizer = self.tokenizer_class.from_pretrained("bert-base-japanese-char")
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user