transformers/tests/models/hubert
Matt edb704b26e
Fix inverted conditional in TF common test! (#22540)
* Fix inverted conditional in TF common test!

* Make the same change in the PT tests file

* Make sure hidden states for GPT2 have the same output shape in PT/TF

* Minor fix to PT implementation of token classification loss

* Skip loss equivalence test for TFHubert because it keeps overflowing to inf

* Compute LM loss for TF the (weird) way it's computed in PT

* Skip loss equivalence test for Wav2Vec2 for the same reason as Hubert

* Fix - don't try to access the hidden states property when output is a tuple
2023-04-04 21:59:54 +01:00
..
__init__.py Move test model folders (#17034) 2022-05-03 14:42:02 +02:00
test_modeling_hubert.py 🔥Rework pipeline testing by removing PipelineTestCaseMeta 🚀 (#21516) 2023-02-28 19:40:57 +01:00
test_modeling_tf_hubert.py Fix inverted conditional in TF common test! (#22540) 2023-04-04 21:59:54 +01:00