mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-13 09:40:06 +06:00
Skip RoFormer ONNX test if rjieba not installed (#16981)
* Skip RoFormer ONNX test if rjieba not installed * Update deps table * Skip RoFormer serialization test * Fix RoFormer vocab * Add rjieba to CircleCI
This commit is contained in:
parent
db034660fb
commit
4bb1d0ec84
@ -549,7 +549,7 @@ jobs:
|
|||||||
- v0.4-custom_tokenizers-{{ checksum "setup.py" }}
|
- v0.4-custom_tokenizers-{{ checksum "setup.py" }}
|
||||||
- v0.4-{{ checksum "setup.py" }}
|
- v0.4-{{ checksum "setup.py" }}
|
||||||
- run: pip install --upgrade pip
|
- run: pip install --upgrade pip
|
||||||
- run: pip install .[ja,testing,sentencepiece,jieba,spacy,ftfy]
|
- run: pip install .[ja,testing,sentencepiece,jieba,spacy,ftfy,rjieba]
|
||||||
- run: python -m unidic download
|
- run: python -m unidic download
|
||||||
- save_cache:
|
- save_cache:
|
||||||
key: v0.4-custom_tokenizers-{{ checksum "setup.py" }}
|
key: v0.4-custom_tokenizers-{{ checksum "setup.py" }}
|
||||||
@ -785,7 +785,7 @@ jobs:
|
|||||||
- v0.4-torch-{{ checksum "setup.py" }}
|
- v0.4-torch-{{ checksum "setup.py" }}
|
||||||
- v0.4-{{ checksum "setup.py" }}
|
- v0.4-{{ checksum "setup.py" }}
|
||||||
- run: pip install --upgrade pip
|
- run: pip install --upgrade pip
|
||||||
- run: pip install .[torch,testing,sentencepiece,onnxruntime,vision]
|
- run: pip install .[torch,testing,sentencepiece,onnxruntime,vision,rjieba]
|
||||||
- save_cache:
|
- save_cache:
|
||||||
key: v0.4-onnx-{{ checksum "setup.py" }}
|
key: v0.4-onnx-{{ checksum "setup.py" }}
|
||||||
paths:
|
paths:
|
||||||
|
4
setup.py
4
setup.py
@ -140,6 +140,7 @@ _deps = [
|
|||||||
"ray[tune]",
|
"ray[tune]",
|
||||||
"regex!=2019.12.17",
|
"regex!=2019.12.17",
|
||||||
"requests",
|
"requests",
|
||||||
|
"rjieba",
|
||||||
"rouge-score",
|
"rouge-score",
|
||||||
"sacrebleu>=1.4.12,<2.0.0",
|
"sacrebleu>=1.4.12,<2.0.0",
|
||||||
"sacremoses",
|
"sacremoses",
|
||||||
@ -288,7 +289,8 @@ extras["testing"] = (
|
|||||||
"nltk",
|
"nltk",
|
||||||
"GitPython",
|
"GitPython",
|
||||||
"hf-doc-builder",
|
"hf-doc-builder",
|
||||||
'sacremoses'
|
"sacremoses",
|
||||||
|
"rjieba"
|
||||||
)
|
)
|
||||||
+ extras["retrieval"]
|
+ extras["retrieval"]
|
||||||
+ extras["modelcreation"]
|
+ extras["modelcreation"]
|
||||||
|
@ -47,6 +47,7 @@ deps = {
|
|||||||
"ray[tune]": "ray[tune]",
|
"ray[tune]": "ray[tune]",
|
||||||
"regex": "regex!=2019.12.17",
|
"regex": "regex!=2019.12.17",
|
||||||
"requests": "requests",
|
"requests": "requests",
|
||||||
|
"rjieba": "rjieba",
|
||||||
"rouge-score": "rouge-score",
|
"rouge-score": "rouge-score",
|
||||||
"sacrebleu": "sacrebleu>=1.4.12,<2.0.0",
|
"sacrebleu": "sacrebleu>=1.4.12,<2.0.0",
|
||||||
"sacremoses": "sacremoses",
|
"sacremoses": "sacremoses",
|
||||||
|
@ -27,7 +27,7 @@ from .tokenization_utils import JiebaPreTokenizer
|
|||||||
|
|
||||||
logger = logging.get_logger(__name__)
|
logger = logging.get_logger(__name__)
|
||||||
|
|
||||||
VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"}
|
VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"}
|
||||||
|
|
||||||
PRETRAINED_VOCAB_FILES_MAP = {
|
PRETRAINED_VOCAB_FILES_MAP = {
|
||||||
"vocab_file": {
|
"vocab_file": {
|
||||||
|
@ -71,3 +71,7 @@ class RoFormerTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
|
|||||||
# can't train new_tokenizer via Tokenizers lib
|
# can't train new_tokenizer via Tokenizers lib
|
||||||
def test_training_new_tokenizer_with_special_tokens_change(self):
|
def test_training_new_tokenizer_with_special_tokens_change(self):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
# can't serialise custom PreTokenizer
|
||||||
|
def test_save_slow_from_fast_and_reload_fast(self):
|
||||||
|
pass
|
||||||
|
@ -16,7 +16,7 @@ from transformers.onnx import (
|
|||||||
validate_model_outputs,
|
validate_model_outputs,
|
||||||
)
|
)
|
||||||
from transformers.onnx.utils import compute_effective_axis_dimension, compute_serialized_parameters_size
|
from transformers.onnx.utils import compute_effective_axis_dimension, compute_serialized_parameters_size
|
||||||
from transformers.testing_utils import require_onnx, require_tf, require_torch, require_vision, slow
|
from transformers.testing_utils import require_onnx, require_rjieba, require_tf, require_torch, require_vision, slow
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available() or is_tf_available():
|
if is_torch_available() or is_tf_available():
|
||||||
@ -287,6 +287,7 @@ class OnnxExportTestCaseV2(TestCase):
|
|||||||
@slow
|
@slow
|
||||||
@require_torch
|
@require_torch
|
||||||
@require_vision
|
@require_vision
|
||||||
|
@require_rjieba
|
||||||
def test_pytorch_export(self, test_name, name, model_name, feature, onnx_config_class_constructor):
|
def test_pytorch_export(self, test_name, name, model_name, feature, onnx_config_class_constructor):
|
||||||
self._onnx_export(test_name, name, model_name, feature, onnx_config_class_constructor)
|
self._onnx_export(test_name, name, model_name, feature, onnx_config_class_constructor)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user