Suraj Patil
f6e74a63ca
Add m2m100 ( #10236 )
...
* m2m_100
* no layernorm_embedding
* sinusoidal positional embeddings
* update pos embeddings
* add default config values
* tokenizer
* add conversion script
* fix config
* fix pos embed
* remove _float_tensor
* update tokenizer
* update lang codes
* handle lang codes
* fix pos embeds
* fix spm key
* put embedding weights on device
* remove qa and seq classification heads
* fix convert script
* lang codes pn one line
* fix embeds
* fix tokenizer
* fix tokenizer
* add fast tokenizer
* style
* M2M100MT => M2M100
* fix copyright, style
* tokenizer converter
* vocab file
* remove fast tokenizer
* fix embeds
* fix tokenizer
* fix tests
* add tokenizer tests
* add integration test
* quality
* fix model name
* fix test
* doc
* doc
* fix doc
* add copied from statements
* fix tokenizer tests
* apply review suggestions
* fix urls
* fix shift_tokens_right
* apply review suggestions
* fix
* fix doc
* add lang code to id
* remove unused function
* update checkpoint names
* fix copy
* fix tokenizer
* fix checkpoint names
* fix merge issue
* style
2021-03-06 22:14:16 +05:30
Patrick von Platen
cb38ffcc5e
[PretrainedFeatureExtractor] + Wav2Vec2FeatureExtractor, Wav2Vec2Processor, Wav2Vec2Tokenizer ( #10324 )
...
* push to show
* small improvement
* small improvement
* Update src/transformers/feature_extraction_utils.py
* Update src/transformers/feature_extraction_utils.py
* implement base
* add common tests
* make all tests pass for wav2vec2
* make padding work & add more tests
* finalize feature extractor utils
* add call method to feature extraction
* finalize feature processor
* finish tokenizer
* finish general processor design
* finish tests
* typo
* remove bogus file
* finish docstring
* add docs
* finish docs
* small fix
* correct docs
* save intermediate
* load changes
* apply changes
* apply changes to doc
* change tests
* apply surajs recommend
* final changes
* Apply suggestions from code review
* fix typo
* fix import
* correct docstring
2021-02-25 17:42:46 +03:00
Patrick von Platen
b972125ced
Deprecate Wav2Vec2ForMaskedLM and add Wav2Vec2ForCTC ( #10089 )
...
* add wav2vec2CTC and deprecate for maskedlm
* remove from docs
2021-02-09 03:49:02 -05:00
demSd
00031785a8
BartForCausalLM analogs to ProphetNetForCausalLM
( #9128 )
...
* initiliaze bart4causalLM
* create BartDecoderWrapper, setters/getters
* delete spaces
* forward and additional methods
* update cache function, loss function, remove ngram* params in data class.
* add bartcausallm, bartdecoder testing
* correct bart for causal lm
* remove at
* add mbart as well
* up
* fix typo
* up
* correct
* add pegasusforcausallm
* add blenderbotforcausallm
* add blenderbotsmallforcausallm
* add marianforcausallm
* add test for MarianForCausalLM
* add Pegasus test
* add BlenderbotSmall test
* add blenderbot test
* fix a fail
* fix an import fail
* a fix
* fix
* Update modeling_pegasus.py
* fix models
* fix inputs_embeds setting getter
* adapt tests
* correct repo utils check
* finish test improvement
* fix tf models as well
* make style
* make fix-copies
* fix copies
* run all tests
* last changes
* fix all tests
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
2021-02-04 11:56:12 +03:00
abhishek thakur
f617490e71
ConvBERT Model ( #9717 )
...
* finalize convbert
* finalize convbert
* fix
* fix
* fix
* push
* fix
* tf image patches
* fix torch model
* tf tests
* conversion
* everything aligned
* remove print
* tf tests
* fix tf
* make tf tests pass
* everything works
* fix init
* fix
* special treatment for sepconv1d
* style
* 🙏🏽
* add doc and cleanup
* add electra test again
* fix doc
* fix doc again
* fix doc again
* Update src/transformers/modeling_tf_pytorch_utils.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* Update src/transformers/models/conv_bert/configuration_conv_bert.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* Update docs/source/model_doc/conv_bert.rst
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/auto/configuration_auto.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/conv_bert/configuration_conv_bert.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* conv_bert -> convbert
* more fixes from review
* add conversion script
* dont use pretrained embed
* unused config
* suggestions from julien
* some more fixes
* p -> param
* fix copyright
* fix doc
* Update src/transformers/models/convbert/configuration_convbert.py
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* comments from reviews
* fix-copies
* fix style
* revert shape_list
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
2021-01-27 03:20:09 -05:00
Sylvain Gugger
758ed3332b
Transformers fast import part 2 ( #9446 )
...
* Main init work
* Add version
* Change from absolute to relative imports
* Fix imports
* One more typo
* More typos
* Styling
* Make quality script pass
* Add necessary replace in template
* Fix typos
* Spaces are ignored in replace for some reason
* Forgot one models.
* Fixes for import
Co-authored-by: LysandreJik <lysandre.debut@reseau.eseo.fr>
* Add documentation
* Styling
Co-authored-by: LysandreJik <lysandre.debut@reseau.eseo.fr>
2021-01-07 09:36:14 -05:00
Qbiwan
ecfcac223c
Improve documentation coverage for Phobert ( #9427 )
...
* first commit
* change phobert to phoBERT as per author in overview
* v3 and v4 both runs on same code hence there is no need to differentiate them
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2021-01-06 10:04:32 -05:00
Qbiwan
be898998bb
Improve documentation coverage for Herbert ( #9428 )
...
* first commit
* changed XLMTokenizer to HerbertTokenizer in code example
2021-01-06 09:13:43 -05:00
Patrick von Platen
eef66035a2
[PyTorch Bart] Split Bart into different models ( #9343 )
...
* first try
* remove old template
* finish bart
* finish mbart
* delete unnecessary line
* init pegasus
* save intermediate
* correct pegasus
* finish pegasus
* remove cookie cutter leftover
* add marian
* finish blenderbot
* replace in file
* correctly split blenderbot
* delete "old" folder
* correct "add statement"
* adapt config for tf comp
* correct configs for tf
* remove ipdb
* fix more stuff
* fix mbart
* push pegasus fix
* fix mbart
* more fixes
* fix research projects code
* finish docs for bart, mbart, and marian
* delete unnecessary file
* correct attn typo
* correct configs
* remove pegasus for seq class
* correct peg docs
* correct peg docs
* finish configs
* further improve docs
* add copied from statements to mbart
* fix copied from in mbart
* add copy statements to marian
* add copied from to marian
* add pegasus copied from
* finish pegasus
* finish copied from
* Apply suggestions from code review
* make style
* backward comp blenderbot
* apply lysandres and sylvains suggestions
* apply suggestions
* push last fixes
* fix docs
* fix tok tests
* fix imports code style
* fix doc
2021-01-05 22:00:05 +01:00
Patrick von Platen
189387e9b2
LED ( #9278 )
...
* create model
* add integration
* save current state
* make integration tests pass
* add one more test
* add explanation to tests
* remove from bart
* add padding
* remove unnecessary test
* make all tests pass
* re-add cookie cutter tests
* finish PyTorch
* fix attention test
* Update tests/test_modeling_common.py
* revert change
* remove unused file
* add string to doc
* save intermediate
* make tf integration tests pass
* finish tf
* fix doc
* fix docs again
* add led to doctree
* add to auto tokenizer
* added tips for led
* make style
* apply jplus statements
* correct tf longformer
* apply lysandres suggestions
* apply sylvains suggestions
* Apply suggestions from code review
2021-01-05 13:14:30 +01:00
Qbiwan
086718ac6e
Improve documentation coverage for Bertweet ( #9379 )
...
* bertweet docs coverage
* style doc max len 119
* maxlen style rst
* run main() from style_doc
* changed according to comments
2021-01-04 13:12:59 -05:00
Julien Plu
de29ff9bd2
Fix open ( #9368 )
2021-01-04 10:22:15 -05:00
Patrick von Platen
d5db6c37d4
[Seq2Seq Templates] Fix check_repo.py templates file ( #9277 )
...
* add enc dec pt model to check repo
* fix indent
2020-12-23 11:40:20 +01:00
Sylvain Gugger
1fc7119181
Fix script that check objects are documented ( #9259 )
2020-12-22 11:12:58 -05:00
Patrick von Platen
9cc9f4122e
Make ProphetNetModel really compatible with EncoderDecoder ( #9033 )
...
* improve
* finish
* upload model
* fix lm head
* fix test
2020-12-11 16:59:54 +01:00
Sylvain Gugger
1310e1a758
Enforce all objects in the main init are documented ( #9014 )
2020-12-10 11:57:12 -05:00
Patrick von Platen
06971ac4f9
[Bart] Refactor - fix issues, consistency with the library, naming ( #8900 )
...
* remove make on the fly linear embedding
* start refactor
* big first refactor
* save intermediate
* save intermediat
* correct mask issue
* save tests
* refactor padding masks
* make all tests pass
* further refactor
* make pegasus test pass
* fix bool if
* fix leftover tests
* continue
* bart renaming
* delete torchscript test hack
* fix imports in tests
* correct shift
* fix docs and repo cons
* re-add fix for FSTM
* typo in test
* fix typo
* fix another typo
* continue
* hot fix 2 for tf
* small fixes
* refactor types linting
* continue
* finish refactor
* fix import in tests
* better bart names
* further refactor and add test
* delete hack
* apply sylvains and lysandres commens
* small perf improv
* further perf improv
* improv perf
* fix typo
* make style
* small perf improv
2020-12-09 20:55:24 +01:00
Ahmed Elnaggar
40ecaf0c2b
Add T5 Encoder for Feature Extraction ( #8717 )
...
* Add T5 Encoder class for feature extraction
* fix T5 encoder add_start_docstrings indent
* update init with T5 encoder
* update init with TFT5ModelEncoder
* remove TFT5ModelEncoder
* change T5ModelEncoder order in init
* add T5ModelEncoder to transformers init
* clean T5ModelEncoder
* update init with TFT5ModelEncoder
* add TFModelEncoder for Tensorflow
* update init with TFT5ModelEncoder
* Update src/transformers/models/t5/modeling_t5.py
change output from Seq2SeqModelOutput to BaseModelOutput
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* remove encoder_outputs
1. remove encoder_outputs from the function call.
2. remove the encoder_outputs If statement.
3. remove isinstance from return_dict.
* Authorize missing decoder keys
* remove unnecessary input parameters
remove pask_key_values and use_cache
* remove use_cache
remove use_cache from the forward method
* add doctoring for T5 encoder
add doctoring for T5 encoder with T5_ENCODER_INPUTS_DOCSTRING
* change return_dict to dot access
* add T5_ENCODER_INPUTS_DOCSTRING for TF T5
* change TFT5Encoder output type to BaseModelOutput
* remove unnecessary parameters for TFT5Encoder
* remove unnecessary if statement
* add import BaseModelOutput
* fix BaseModelOutput typo to TFBaseModelOutput
* update T5 doc with T5ModelEncoder
* add T5ModelEncoder to tests
* finish pytorch
* finish docs and mt5
* add mtf to init
* fix init
* remove n_positions
* finish PR
* Update src/transformers/models/mt5/modeling_mt5.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* Update src/transformers/models/t5/modeling_t5.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* Update src/transformers/models/t5/modeling_tf_t5.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* Update src/transformers/models/mt5/modeling_tf_mt5.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* make style
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
2020-11-30 08:34:40 +01:00
Sylvain Gugger
4821ea5aeb
Big model table ( #8774 )
...
* First draft
* Styling
* With all changes staged
* Update docs/source/index.rst
Co-authored-by: Julien Chaumond <chaumond@gmail.com>
* Styling
Co-authored-by: Julien Chaumond <chaumond@gmail.com>
2020-11-25 12:02:15 -05:00
Sylvain Gugger
7f3b41a306
Fix check repo utils ( #8600 )
2020-11-17 14:01:46 -05:00
Patrick von Platen
86822a358b
T5 & mT5 ( #8552 )
...
* add mt5 and t5v1_1 model
* fix tests
* correct some imports
* add tf model
* finish tf t5
* improve examples
* fix copies
* clean doc
2020-11-17 12:23:09 +01:00
Sylvain Gugger
c89bdfbe72
Reorganize repo ( #8580 )
...
* Put models in subfolders
* Styling
* Fix imports in tests
* More fixes in test imports
* Sneaky hidden imports
* Fix imports in doc files
* More sneaky imports
* Finish fixing tests
* Fix examples
* Fix path for copies
* More fixes for examples
* Fix dummy files
* More fixes for example
* More model import fixes
* Is this why you're unhappy GitHub?
* Fix imports in conver command
2020-11-16 21:43:42 -05:00
Lysandre Debut
826f04576f
Model templates encoder only ( #8509 )
...
* Model templates
* TensorFlow
* Remove pooler
* CI
* Tokenizer + Refactoring
* Encoder-Decoder
* Let's go testing
* Encoder-Decoder in TF
* Let's go testing in TF
* Documentation
* README
* Fixes
* Better names
* Style
* Update docs
* Choose to skip either TF or PT
* Code quality fixes
* Add to testing suite
* Update file path
* Cookiecutter path
* Update `transformers` path
* Handle rebasing
* Remove seq2seq from model templates
* Remove s2s config
* Apply Sylvain and Patrick comments
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Last fixes from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2020-11-13 11:59:30 -05:00
Julien Plu
91a67b7506
Use LF instead of os.linesep ( #8491 )
2020-11-12 13:52:40 -05:00
Ratthachat (Jung)
026a2ff225
Add TFDPR ( #8203 )
...
* Create modeling_tf_dpr.py
* Add TFDPR
* Add back TFPegasus, TFMarian, TFMBart, TFBlenderBot
last commit accidentally deleted these 4 lines, so I recover them back
* Add TFDPR
* Add TFDPR
* clean up some comments, add TF input-style doc string
* Add TFDPR
* Make return_dict=False as default
* Fix return_dict bug (in .from_pretrained)
* Add get_input_embeddings()
* Create test_modeling_tf_dpr.py
The current version is already passed all 27 tests!
Please see the test run at :
https://colab.research.google.com/drive/1czS_m9zy5k-iSJbzA_DP1k1xAAC_sdkf?usp=sharing
* fix quality
* delete init weights
* run fix copies
* fix repo consis
* del config_class, load_tf_weights
They shoud be 'pytorch only'
* add config_class back
after removing it, test failed ... so totally only removing "use_tf_weights = None" on Lysandre suggestion
* newline after .. note::
* import tf, np (Necessary for ModelIntegrationTest)
* slow_test from_pretrained with from_pt=True
At the moment we don't have TF weights (since we don't have official official TF model)
Previously, I did not run slow test, so I missed this bug
* Add simple TFDPRModelIntegrationTest
Note that this is just a test that TF and Pytorch gives approx. the same output.
However, I could not test with the official DPR repo's output yet
* upload correct tf model
* remove position_ids as missing keys
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: patrickvonplaten <patrick@huggingface.co>
2020-11-11 12:28:09 -05:00
Julien Plu
8551a99232
Add auto next sentence prediction ( #8432 )
...
* Add auto next sentence prediction
* Fix style
* Add mobilebert next sentence prediction
2020-11-10 11:11:48 -05:00
Sylvain Gugger
a39218b75b
Check all models are in an auto class ( #8425 )
2020-11-09 15:44:54 -05:00
Julien Plu
76e7a44dee
Fix some tooling for windows ( #8359 )
...
* Fix some tooling for windows
* Fix conflict
* Trigger CI
2020-11-09 13:50:38 +01:00
Sam Shleifer
566b083eb1
TFMarian, TFMbart, TFPegasus, TFBlenderbot ( #7987 )
...
* Start plumbing
* Marian close
* Small stubs for all children
* Fixed bart
* marian working
* pegasus test is good, but failing
* Checkin tests
* More model files
* Subtle marian, pegasus integration test failures
* Works well
* rm print
* boom boom
* Still failing model2doc
* merge master
* Equivalence test failing, all others fixed
* cleanup
* Fix embed_scale
* Cleanup marian pipeline test
* Undo extra changes
* Smaller delta
* Cleanup model testers
* undo delta
* fix tests import structure
* cross test decorator
* Cleaner set_weights
* Respect authorized_unexpected_keys
* No warnings
* No warnings
* style
* Nest tf import
* black
* Apply suggestions from code review
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* functional dropout
* fixup
* Fixup
* style_doc
* embs
* shape list
* delete slow force_token_id_to_be_generated func
* fixup
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
2020-10-30 11:23:16 -04:00
Stas Bekman
ca37db0559
[flax] fix repo_check ( #7914 )
...
* [flax] fix repo_check
Unless, this is actually a problem, this adds `modeling_flax_utils` to ignore list. otherwise currently it expects to have a 'tests/test_modeling_flax_utils.py' for it.
for context please see: https://github.com/huggingface/transformers/pull/3722#issuecomment-712360415
* fix 2 more issues
* merge https://github.com/huggingface/transformers/pull/7919/
2020-10-20 07:55:40 -04:00
Weizhen
2422cda01b
ProphetNet ( #7157 )
...
* add new model prophetnet
prophetnet modified
modify codes as suggested v1
add prophetnet test files
* still bugs, because of changed output formats of encoder and decoder
* move prophetnet into the latest version
* clean integration tests
* clean tokenizers
* add xlm config to init
* correct typo in init
* further refactoring
* continue refactor
* save parallel
* add decoder_attention_mask
* fix use_cache vs. past_key_values
* fix common tests
* change decoder output logits
* fix xlm tests
* make common tests pass
* change model architecture
* add tokenizer tests
* finalize model structure
* no weight mapping
* correct n-gram stream attention mask as discussed with qweizhen
* remove unused import
* fix index.rst
* fix tests
* delete unnecessary code
* add fast integration test
* rename weights
* final weight remapping
* save intermediate
* Descriptions for Prophetnet Config File
* finish all models
* finish new model outputs
* delete unnecessary files
* refactor encoder layer
* add dummy docs
* code quality
* fix tests
* add model pages to doctree
* further refactor
* more refactor, more tests
* finish code refactor and tests
* remove unnecessary files
* further clean up
* add docstring template
* finish tokenizer doc
* finish prophetnet
* fix copies
* fix typos
* fix tf tests
* fix fp16
* fix tf test 2nd try
* fix code quality
* add test for each model
* merge new tests to branch
* Update model_cards/microsoft/prophetnet-large-uncased-cnndm/README.md
Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
* Update model_cards/microsoft/prophetnet-large-uncased-cnndm/README.md
Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
* Update src/transformers/modeling_prophetnet.py
Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
* Update utils/check_repo.py
Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
* apply sams and sylvains comments
* make style
* remove unnecessary code
* Update README.md
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update README.md
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/configuration_prophetnet.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* implement lysandres comments
* correct docs
* fix isort
* fix tokenizers
* fix copies
Co-authored-by: weizhen <weizhen@mail.ustc.edu.cn>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
2020-10-19 17:36:09 +02:00
Sylvain Gugger
0611eab5e3
Document RAG again ( #7377 )
...
Do not merge before Monday
2020-09-28 08:31:46 -04:00
Sylvain Gugger
a8e7982f84
Remove mentions of RAG from the docs ( #7376 )
...
* Remove mentions of RAG from the docs
* Deactivate check
2020-09-24 17:07:14 -04:00
Sylvain Gugger
1ff5bd38a3
Check decorator order ( #7326 )
...
* Check decorator order
* Adapt for parametrized decorators
* Fix typos
2020-09-24 04:54:37 -04:00
Sylvain Gugger
e4b94d8e58
Copy code from Bert to Roberta and add safeguard script ( #7219 )
...
* Copy code from Bert to Roberta and add safeguard script
* Fix docstring
* Comment code
* Formatting
* Update src/transformers/modeling_roberta.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* Add test and fix bugs
* Fix style and make new comand
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
2020-09-22 05:02:27 -04:00
Patrick von Platen
7fd1febf38
Add "Leveraging Pretrained Checkpoints for Generation" Seq2Seq models. ( #6594 )
...
* add conversion script
* improve conversion script
* make style
* add tryout files
* fix
* update
* add causal bert
* better names
* add tokenizer file as well
* finish causal_bert
* fix small bugs
* improve generate
* change naming
* renaming
* renaming
* renaming
* remove leftover files
* clean files
* add fix tokenizer
* finalize
* correct slow test
* update docs
* small fixes
* fix link
* adapt check repo
* apply sams and sylvains recommendations
* fix import
* implement Lysandres recommendations
* fix logger warn
2020-09-10 16:40:51 +02:00
Sylvain Gugger
d155b38d6e
Funnel transformer ( #6908 )
...
* Initial model
* Fix upsampling
* Add special cls token id and test
* Formatting
* Test and fist FunnelTokenizerFast
* Common tests
* Fix the check_repo script and document Funnel
* Doc fixes
* Add all models
* Write doc
* Fix test
* Initial model
* Fix upsampling
* Add special cls token id and test
* Formatting
* Test and fist FunnelTokenizerFast
* Common tests
* Fix the check_repo script and document Funnel
* Doc fixes
* Add all models
* Write doc
* Fix test
* Fix copyright
* Forgot some layers can be repeated
* Apply suggestions from code review
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* Update src/transformers/modeling_funnel.py
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
* Address review comments
* Update src/transformers/modeling_funnel.py
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* Address review comments
* Update src/transformers/modeling_funnel.py
Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
* Slow integration test
* Make small integration test
* Formatting
* Add checkpoint and separate classification head
* Formatting
* Expand list, fix link and add in pretrained models
* Styling
* Add the model in all summaries
* Typo fixes
Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
2020-09-08 08:08:08 -04:00
Suraj Patil
680f1337c3
MBartForConditionalGeneration ( #6441 )
...
* add MBartForConditionalGeneration
* style
* rebase and fixes
* add mbart test in TEST_FILES_WITH_NO_COMMON_TESTS
* fix docs
* don't ignore mbart
* doc
* fix mbart fairseq link
* put mbart before bart
* apply doc suggestions
2020-08-14 03:21:16 -04:00
Sylvain Gugger
34fabe1697
Move prediction_loss_only to TrainingArguments ( #6426 )
2020-08-12 08:03:45 -04:00
Lysandre Debut
b99098abc7
Patch models ( #6326 )
...
* TFAlbertFor{TokenClassification, MultipleChoice}
* Patch models
* BERT and TF BERT info
s
* Update check_repo
2020-08-10 10:39:17 -04:00
Sylvain Gugger
6ba540b747
Add a script to check all models are tested and documented ( #6298 )
...
* Add a script to check all models are tested and documented
* Apply suggestions from code review
Co-authored-by: Kevin Canwen Xu <canwenxu@126.com>
* Address comments
Co-authored-by: Kevin Canwen Xu <canwenxu@126.com>
2020-08-07 09:18:37 -04:00