mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-04 13:20:12 +06:00

* Initial support for upload to hub * push -> upload * Fixes + examples * Fix torchhub test * Torchhub test I hate you * push_model_to_hub -> push_to_hub * Apply mixin to other pretrained models * Remove ABC inheritance * Add tests * Typo * Run tests * Install git-lfs * Change approach * Add push_to_hub to all * Staging test suite * Typo * Maybe like this? * More deps * Cache * Adapt name * Quality * MOAR tests * Put it in testing_utils * Docs + torchhub last hope * Styling * Wrong method * Typos * Update src/transformers/file_utils.py Co-authored-by: Julien Chaumond <julien@huggingface.co> * Address review comments * Apply suggestions from code review Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by: Julien Chaumond <julien@huggingface.co> Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
141 lines
5.3 KiB
Python
141 lines
5.3 KiB
Python
# coding=utf-8
|
|
# Copyright 2019 HuggingFace Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
|
|
import json
|
|
import os
|
|
import tempfile
|
|
import unittest
|
|
|
|
from huggingface_hub import HfApi
|
|
from requests.exceptions import HTTPError
|
|
from transformers import BertConfig
|
|
from transformers.testing_utils import ENDPOINT_STAGING, PASS, USER, is_staging_test
|
|
|
|
|
|
class ConfigTester(object):
|
|
def __init__(self, parent, config_class=None, has_text_modality=True, **kwargs):
|
|
self.parent = parent
|
|
self.config_class = config_class
|
|
self.has_text_modality = has_text_modality
|
|
self.inputs_dict = kwargs
|
|
|
|
def create_and_test_config_common_properties(self):
|
|
config = self.config_class(**self.inputs_dict)
|
|
if self.has_text_modality:
|
|
self.parent.assertTrue(hasattr(config, "vocab_size"))
|
|
self.parent.assertTrue(hasattr(config, "hidden_size"))
|
|
self.parent.assertTrue(hasattr(config, "num_attention_heads"))
|
|
self.parent.assertTrue(hasattr(config, "num_hidden_layers"))
|
|
|
|
def create_and_test_config_to_json_string(self):
|
|
config = self.config_class(**self.inputs_dict)
|
|
obj = json.loads(config.to_json_string())
|
|
for key, value in self.inputs_dict.items():
|
|
self.parent.assertEqual(obj[key], value)
|
|
|
|
def create_and_test_config_to_json_file(self):
|
|
config_first = self.config_class(**self.inputs_dict)
|
|
|
|
with tempfile.TemporaryDirectory() as tmpdirname:
|
|
json_file_path = os.path.join(tmpdirname, "config.json")
|
|
config_first.to_json_file(json_file_path)
|
|
config_second = self.config_class.from_json_file(json_file_path)
|
|
|
|
self.parent.assertEqual(config_second.to_dict(), config_first.to_dict())
|
|
|
|
def create_and_test_config_from_and_save_pretrained(self):
|
|
config_first = self.config_class(**self.inputs_dict)
|
|
|
|
with tempfile.TemporaryDirectory() as tmpdirname:
|
|
config_first.save_pretrained(tmpdirname)
|
|
config_second = self.config_class.from_pretrained(tmpdirname)
|
|
|
|
self.parent.assertEqual(config_second.to_dict(), config_first.to_dict())
|
|
|
|
def create_and_test_config_with_num_labels(self):
|
|
config = self.config_class(**self.inputs_dict, num_labels=5)
|
|
self.parent.assertEqual(len(config.id2label), 5)
|
|
self.parent.assertEqual(len(config.label2id), 5)
|
|
|
|
config.num_labels = 3
|
|
self.parent.assertEqual(len(config.id2label), 3)
|
|
self.parent.assertEqual(len(config.label2id), 3)
|
|
|
|
def check_config_can_be_init_without_params(self):
|
|
if self.config_class.is_composition:
|
|
return
|
|
config = self.config_class()
|
|
self.parent.assertIsNotNone(config)
|
|
|
|
def run_common_tests(self):
|
|
self.create_and_test_config_common_properties()
|
|
self.create_and_test_config_to_json_string()
|
|
self.create_and_test_config_to_json_file()
|
|
self.create_and_test_config_from_and_save_pretrained()
|
|
self.create_and_test_config_with_num_labels()
|
|
self.check_config_can_be_init_without_params()
|
|
|
|
|
|
@is_staging_test
|
|
class ConfigPushToHubTester(unittest.TestCase):
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
cls._api = HfApi(endpoint=ENDPOINT_STAGING)
|
|
cls._token = cls._api.login(username=USER, password=PASS)
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
try:
|
|
cls._api.delete_repo(token=cls._token, name="test-model")
|
|
except HTTPError:
|
|
pass
|
|
|
|
try:
|
|
cls._api.delete_repo(token=cls._token, name="test-model-org", organization="valid_org")
|
|
except HTTPError:
|
|
pass
|
|
|
|
def test_push_to_hub(self):
|
|
config = BertConfig(
|
|
vocab_size=99, hidden_size=32, num_hidden_layers=5, num_attention_heads=4, intermediate_size=37
|
|
)
|
|
with tempfile.TemporaryDirectory() as tmp_dir:
|
|
config.save_pretrained(tmp_dir, push_to_hub=True, repo_name="test-model", use_auth_token=self._token)
|
|
|
|
new_config = BertConfig.from_pretrained(f"{USER}/test-model")
|
|
for k, v in config.__dict__.items():
|
|
if k != "transformers_version":
|
|
self.assertEqual(v, getattr(new_config, k))
|
|
|
|
def test_push_to_hub_in_organization(self):
|
|
config = BertConfig(
|
|
vocab_size=99, hidden_size=32, num_hidden_layers=5, num_attention_heads=4, intermediate_size=37
|
|
)
|
|
|
|
with tempfile.TemporaryDirectory() as tmp_dir:
|
|
config.save_pretrained(
|
|
tmp_dir,
|
|
push_to_hub=True,
|
|
repo_name="test-model-org",
|
|
use_auth_token=self._token,
|
|
organization="valid_org",
|
|
)
|
|
|
|
new_config = BertConfig.from_pretrained("valid_org/test-model-org")
|
|
for k, v in config.__dict__.items():
|
|
if k != "transformers_version":
|
|
self.assertEqual(v, getattr(new_config, k))
|