mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-24 06:48:58 +06:00
fix head_mask for albert encoder part(AlbertTransformer
) (#11596)
* fix head mask for albert encoder part * fix head_mask for albert encoder part
This commit is contained in:
parent
864c1dfe34
commit
c1780ce7a4
@ -450,6 +450,8 @@ class AlbertTransformer(nn.Module):
|
||||
all_hidden_states = (hidden_states,) if output_hidden_states else None
|
||||
all_attentions = () if output_attentions else None
|
||||
|
||||
head_mask = [None] * self.config.num_hidden_layers if head_mask is None else head_mask
|
||||
|
||||
for i in range(self.config.num_hidden_layers):
|
||||
# Number of layers in a hidden group
|
||||
layers_per_group = int(self.config.num_hidden_layers / self.config.num_hidden_groups)
|
||||
|
Loading…
Reference in New Issue
Block a user