mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
![]() * Fix cross-attention head mask for Torch BART models * Fix head masking for cross-attention module for the following models: BART, Blenderbot, Blenderbot_small, M2M_100, Marian, MBart, Pegasus * Enable test_headmasking for M2M_100 model * Fix cross_head_mask for FSMT, LED and T5 * This commit fixes `head_mask` for cross-attention modules in the following models: FSMT, LED, T5 * It also contains some smaller changes in doc so that it is be perfectly clear the shape of `cross_head_mask` is the same as of `decoder_head_mask` * Update template * Fix template for BartForCausalLM * Fix cross_head_mask for Speech2Text models * Fix cross_head_mask in templates * Fix args order in BartForCausalLM template * Fix doc in BART templates * Make more explicit naming * `cross_head_mask` -> `cross_attn_head_mask` * `cross_layer_head_mask` -> `cross_attn_layer_head_mask` * Fix doc * make style quality * Fix speech2text docstring |
||
---|---|---|
.. | ||
__init__.py | ||
{{cookiecutter.lowercase_modelname}}.rst | ||
configuration_{{cookiecutter.lowercase_modelname}}.py | ||
configuration.json | ||
modeling_{{cookiecutter.lowercase_modelname}}.py | ||
modeling_tf_{{cookiecutter.lowercase_modelname}}.py | ||
test_modeling_{{cookiecutter.lowercase_modelname}}.py | ||
test_modeling_tf_{{cookiecutter.lowercase_modelname}}.py | ||
to_replace_{{cookiecutter.lowercase_modelname}}.py | ||
tokenization_{{cookiecutter.lowercase_modelname}}.py | ||
tokenization_fast_{{cookiecutter.lowercase_modelname}}.py |