mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
Fix tests skip (#32012)
* [run-slow] clip * [run-slow] clip * Fix skip -> skipTest * [run-slow] clip
This commit is contained in:
parent
24cfcc2114
commit
691586b0dc
@ -716,7 +716,7 @@ class BigBirdModelIntegrationTest(unittest.TestCase):
|
||||
"""
|
||||
|
||||
if not self.test_attention_probs:
|
||||
self.skip("test_attention_probs is set to False")
|
||||
self.skipTest("test_attention_probs is set to False")
|
||||
|
||||
model = BigBirdModel.from_pretrained(
|
||||
"google/bigbird-roberta-base", attention_type="block_sparse", num_random_blocks=3, block_size=16
|
||||
|
@ -3335,7 +3335,7 @@ class WhisperEncoderModelTest(ModelTesterMixin, GenerationTesterMixin, unittest.
|
||||
fx_model_class_name = "Flax" + model_class.__name__
|
||||
|
||||
if not hasattr(transformers, fx_model_class_name):
|
||||
self.skip("Flax model does not exist")
|
||||
self.skipTest("Flax model does not exist")
|
||||
|
||||
# Output all for aggressive testing
|
||||
config.output_hidden_states = True
|
||||
|
@ -2678,7 +2678,7 @@ class TokenizerTesterMixin:
|
||||
config = config_class()
|
||||
|
||||
if config.is_encoder_decoder or config.pad_token_id is None:
|
||||
self.skip("Model is not an encoder-decoder model or has no set pad token id")
|
||||
self.skipTest("Model is not an encoder-decoder model or has no set pad token id")
|
||||
|
||||
# Build sequence
|
||||
first_ten_tokens = list(tokenizer.get_vocab().keys())[:10]
|
||||
|
Loading…
Reference in New Issue
Block a user