mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 18:22:34 +06:00
TF: add initializer_std with a small value in TFFunnelModelTester (#15684)
This commit is contained in:
parent
416dff736c
commit
f8ff3fad87
@ -63,6 +63,7 @@ class TFFunnelModelTester:
|
|||||||
activation_dropout=0.0,
|
activation_dropout=0.0,
|
||||||
max_position_embeddings=512,
|
max_position_embeddings=512,
|
||||||
type_vocab_size=3,
|
type_vocab_size=3,
|
||||||
|
initializer_std=0.02, # Set to a smaller value, so we can keep the small error threshold (1e-5) in the test
|
||||||
num_labels=3,
|
num_labels=3,
|
||||||
num_choices=4,
|
num_choices=4,
|
||||||
scope=None,
|
scope=None,
|
||||||
@ -92,6 +93,7 @@ class TFFunnelModelTester:
|
|||||||
self.num_labels = num_labels
|
self.num_labels = num_labels
|
||||||
self.num_choices = num_choices
|
self.num_choices = num_choices
|
||||||
self.scope = scope
|
self.scope = scope
|
||||||
|
self.initializer_std = initializer_std
|
||||||
|
|
||||||
# Used in the tests to check the size of the first attention layer
|
# Used in the tests to check the size of the first attention layer
|
||||||
self.num_attention_heads = n_head
|
self.num_attention_heads = n_head
|
||||||
@ -137,6 +139,7 @@ class TFFunnelModelTester:
|
|||||||
activation_dropout=self.activation_dropout,
|
activation_dropout=self.activation_dropout,
|
||||||
max_position_embeddings=self.max_position_embeddings,
|
max_position_embeddings=self.max_position_embeddings,
|
||||||
type_vocab_size=self.type_vocab_size,
|
type_vocab_size=self.type_vocab_size,
|
||||||
|
initializer_std=self.initializer_std,
|
||||||
)
|
)
|
||||||
|
|
||||||
return (
|
return (
|
||||||
|
Loading…
Reference in New Issue
Block a user