mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-30 17:52:35 +06:00
![]() * Add cross_attn_head_mask to BART * Fix cross_attentions in TFBart-like models * This commit enables returning of `cross_attentions` for TFBart-like models * It also fixes attention head masking in cross-attenion module * Update TF model templates * Fix missing , in TF model templates * Fix typo: congig -> config |
||
---|---|---|
.. | ||
__init__.py | ||
{{cookiecutter.lowercase_modelname}}.rst | ||
configuration_{{cookiecutter.lowercase_modelname}}.py | ||
configuration.json | ||
modeling_{{cookiecutter.lowercase_modelname}}.py | ||
modeling_tf_{{cookiecutter.lowercase_modelname}}.py | ||
test_modeling_{{cookiecutter.lowercase_modelname}}.py | ||
test_modeling_tf_{{cookiecutter.lowercase_modelname}}.py | ||
to_replace_{{cookiecutter.lowercase_modelname}}.py | ||
tokenization_{{cookiecutter.lowercase_modelname}}.py | ||
tokenization_fast_{{cookiecutter.lowercase_modelname}}.py |