transformers/templates/adding_a_new_model/cookiecutter-template-{{cookiecutter.modelname}}
Thomas Viehmann 6ed9882ddb
use functional interface for softmax in attention (#14198)
* use functional interface instead of instantiating module and immediately calling it

* fix torch.nn.functional to nn.functional. Thank you Stas!
2021-11-30 11:47:33 -05:00
..
__init__.py Add template for adding flax models (#12441) 2021-09-01 09:49:03 +02:00
{{cookiecutter.lowercase_modelname}}.rst Add template for adding flax models (#12441) 2021-09-01 09:49:03 +02:00
configuration_{{cookiecutter.lowercase_modelname}}.py Make gradient_checkpointing a training argument (#13657) 2021-09-22 07:51:38 -04:00
configuration.json Add template for adding flax models (#12441) 2021-09-01 09:49:03 +02:00
modeling_{{cookiecutter.lowercase_modelname}}.py use functional interface for softmax in attention (#14198) 2021-11-30 11:47:33 -05:00
modeling_flax_{{cookiecutter.lowercase_modelname}}.py Fix Flax params dtype (#13098) 2021-11-11 14:45:20 +05:30
modeling_tf_{{cookiecutter.lowercase_modelname}}.py [Docs] More general docstrings (#14028) 2021-10-16 00:48:37 +02:00
test_modeling_{{cookiecutter.lowercase_modelname}}.py Expose get_config() on ModelTesters (#12812) 2021-07-21 04:13:11 -04:00
test_modeling_flax_{{cookiecutter.lowercase_modelname}}.py Add template for adding flax models (#12441) 2021-09-01 09:49:03 +02:00
test_modeling_tf_{{cookiecutter.lowercase_modelname}}.py Add TFEncoderDecoderModel + Add cross-attention to some TF models (#13222) 2021-10-13 00:10:34 +02:00
to_replace_{{cookiecutter.lowercase_modelname}}.py Add template for adding flax models (#12441) 2021-09-01 09:49:03 +02:00
tokenization_{{cookiecutter.lowercase_modelname}}.py Make get_special_tokens_mask consider all tokens (#11163) 2021-04-09 11:57:44 -04:00
tokenization_fast_{{cookiecutter.lowercase_modelname}}.py Copy tokenizer files in each of their repo (#10624) 2021-03-10 11:26:23 -05:00