mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
fix switch transformers
This commit is contained in:
parent
f19d166d2b
commit
ab7fac4a01
@ -979,9 +979,7 @@ class SwitchTransformersStack(SwitchTransformersPreTrainedModel):
|
||||
attention_mask,
|
||||
inputs_embeds,
|
||||
cache_position,
|
||||
past_key_values.self_attention_cache
|
||||
if isinstance(past_key_values, EncoderDecoderCache) is not None
|
||||
else None,
|
||||
past_key_values.self_attention_cache if isinstance(past_key_values, EncoderDecoderCache) else None,
|
||||
output_attentions,
|
||||
)
|
||||
else:
|
||||
|
Loading…
Reference in New Issue
Block a user