mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-03 21:00:08 +06:00
Skip RoFormer ONNX test if rjieba not installed (#16981)
* Skip RoFormer ONNX test if rjieba not installed * Update deps table * Skip RoFormer serialization test * Fix RoFormer vocab * Add rjieba to CircleCI
This commit is contained in:
parent
db034660fb
commit
4bb1d0ec84
@ -549,7 +549,7 @@ jobs:
|
||||
- v0.4-custom_tokenizers-{{ checksum "setup.py" }}
|
||||
- v0.4-{{ checksum "setup.py" }}
|
||||
- run: pip install --upgrade pip
|
||||
- run: pip install .[ja,testing,sentencepiece,jieba,spacy,ftfy]
|
||||
- run: pip install .[ja,testing,sentencepiece,jieba,spacy,ftfy,rjieba]
|
||||
- run: python -m unidic download
|
||||
- save_cache:
|
||||
key: v0.4-custom_tokenizers-{{ checksum "setup.py" }}
|
||||
@ -785,7 +785,7 @@ jobs:
|
||||
- v0.4-torch-{{ checksum "setup.py" }}
|
||||
- v0.4-{{ checksum "setup.py" }}
|
||||
- run: pip install --upgrade pip
|
||||
- run: pip install .[torch,testing,sentencepiece,onnxruntime,vision]
|
||||
- run: pip install .[torch,testing,sentencepiece,onnxruntime,vision,rjieba]
|
||||
- save_cache:
|
||||
key: v0.4-onnx-{{ checksum "setup.py" }}
|
||||
paths:
|
||||
|
4
setup.py
4
setup.py
@ -140,6 +140,7 @@ _deps = [
|
||||
"ray[tune]",
|
||||
"regex!=2019.12.17",
|
||||
"requests",
|
||||
"rjieba",
|
||||
"rouge-score",
|
||||
"sacrebleu>=1.4.12,<2.0.0",
|
||||
"sacremoses",
|
||||
@ -288,7 +289,8 @@ extras["testing"] = (
|
||||
"nltk",
|
||||
"GitPython",
|
||||
"hf-doc-builder",
|
||||
'sacremoses'
|
||||
"sacremoses",
|
||||
"rjieba"
|
||||
)
|
||||
+ extras["retrieval"]
|
||||
+ extras["modelcreation"]
|
||||
|
@ -47,6 +47,7 @@ deps = {
|
||||
"ray[tune]": "ray[tune]",
|
||||
"regex": "regex!=2019.12.17",
|
||||
"requests": "requests",
|
||||
"rjieba": "rjieba",
|
||||
"rouge-score": "rouge-score",
|
||||
"sacrebleu": "sacrebleu>=1.4.12,<2.0.0",
|
||||
"sacremoses": "sacremoses",
|
||||
|
@ -27,7 +27,7 @@ from .tokenization_utils import JiebaPreTokenizer
|
||||
|
||||
logger = logging.get_logger(__name__)
|
||||
|
||||
VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt"}
|
||||
VOCAB_FILES_NAMES = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"}
|
||||
|
||||
PRETRAINED_VOCAB_FILES_MAP = {
|
||||
"vocab_file": {
|
||||
|
@ -71,3 +71,7 @@ class RoFormerTokenizationTest(TokenizerTesterMixin, unittest.TestCase):
|
||||
# can't train new_tokenizer via Tokenizers lib
|
||||
def test_training_new_tokenizer_with_special_tokens_change(self):
|
||||
pass
|
||||
|
||||
# can't serialise custom PreTokenizer
|
||||
def test_save_slow_from_fast_and_reload_fast(self):
|
||||
pass
|
||||
|
@ -16,7 +16,7 @@ from transformers.onnx import (
|
||||
validate_model_outputs,
|
||||
)
|
||||
from transformers.onnx.utils import compute_effective_axis_dimension, compute_serialized_parameters_size
|
||||
from transformers.testing_utils import require_onnx, require_tf, require_torch, require_vision, slow
|
||||
from transformers.testing_utils import require_onnx, require_rjieba, require_tf, require_torch, require_vision, slow
|
||||
|
||||
|
||||
if is_torch_available() or is_tf_available():
|
||||
@ -287,6 +287,7 @@ class OnnxExportTestCaseV2(TestCase):
|
||||
@slow
|
||||
@require_torch
|
||||
@require_vision
|
||||
@require_rjieba
|
||||
def test_pytorch_export(self, test_name, name, model_name, feature, onnx_config_class_constructor):
|
||||
self._onnx_export(test_name, name, model_name, feature, onnx_config_class_constructor)
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user