mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-01 02:31:11 +06:00
Avoid many test failures for LlavaNextVideoForConditionalGeneration
(#34070)
* skip * [run-slow] llava_next_video * skip * [run-slow] video_llava, llava_next_video * skip * [run-slow] llava_next_video --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
This commit is contained in:
parent
37ac078535
commit
80bee7b114
@ -1230,6 +1230,9 @@ class GenerationTesterMixin:
|
|||||||
if any(model_name in model_class.__name__.lower() for model_name in ["marian", "mbart", "pegasus"]):
|
if any(model_name in model_class.__name__.lower() for model_name in ["marian", "mbart", "pegasus"]):
|
||||||
self.skipTest("DoLa is not supported for models that don't return layerwise hidden states")
|
self.skipTest("DoLa is not supported for models that don't return layerwise hidden states")
|
||||||
|
|
||||||
|
if any(model_name == model_class.__name__ for model_name in ["LlavaNextVideoForConditionalGeneration"]):
|
||||||
|
self.skipTest(f"DoLa is failing for {model_class.__name__}")
|
||||||
|
|
||||||
# enable cache if the model is not openai-gpt, xlnet, cpm, or xlm
|
# enable cache if the model is not openai-gpt, xlnet, cpm, or xlm
|
||||||
config, inputs_dict = self.prepare_config_and_inputs_for_generate()
|
config, inputs_dict = self.prepare_config_and_inputs_for_generate()
|
||||||
main_input = inputs_dict[model_class.main_input_name]
|
main_input = inputs_dict[model_class.main_input_name]
|
||||||
|
Loading…
Reference in New Issue
Block a user