transformers/docs/source
Francesco Saverio Zuppichini d83d22f578
Maskformer (#15682)
* maskformer

* conflicts

* conflicts

* minor fixes

* feature extractor test fix

refactor MaskFormerLoss following conversation

MaskFormer related types should not trigger a module time import error

missed one

removed all the types that are not used

update config mapping

minor updates in the doc

resolved conversation that doesn't need a discussion

minor changes

resolved conversations

fixed DetrDecoder

* minor changes

minor changes

fixed mdx file

test feature_extractor return types

functional losses -> classes

removed the return type test for the feature extractor

minor changes + style + quality

* conflicts?

* rebase master

* readme

* added missing files

* deleded poolformers test that where in the wrong palce

* CI

* minor changes

* Apply suggestions from code review

Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>

* resolved conversations

* minor changes

* conversations

[Unispeech] Fix slow tests (#15818)

* remove soundfile old way of loading audio

* Adapt slow test

[Barthez Tokenizer] Fix saving (#15815)

[TFXLNet] Correct tf xlnet generate (#15822)

* [TFXLNet] Correct tf xlnet

* adapt test comment

Fix the push run (#15807)

Fix semantic segmentation pipeline test (#15826)

Fix dummy_inputs() to dummy_inputs in symbolic_trace doc (#15776)

Add model specific output classes to PoolFormer model docs (#15746)

* Added model specific output classes to poolformer docs

* Fixed Segformer typo in Poolformer docs

Adding the option to return_timestamps on pure CTC ASR models. (#15792)

* Adding the option to return_timestamps on pure CTC ASR models.

* Remove `math.prod` which was introduced in Python 3.8

* int are not floats.

* Reworking the PR to support "char" vs "word" output.

* Fixup!

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Update src/transformers/pipelines/automatic_speech_recognition.py

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* Quality.

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

HFTracer.trace should use/return self.graph to be compatible with torch.fx.Tracer (#15824)

Fix tf.concatenate + test past_key_values for TF models (#15774)

* fix wrong method name tf.concatenate

* add tests related to causal LM / decoder

* make style and quality

* clean-up

* Fix TFBertModel's extended_attention_mask when past_key_values is provided

* Fix tests

* fix copies

* More tf.int8 -> tf.int32 in TF test template

* clean-up

* Update TF test template

* revert the previous commit + update the TF test template

* Fix TF template extended_attention_mask when past_key_values is provided

* Fix some styles manually

* clean-up

* Fix ValueError: too many values to unpack in the test

* Fix more: too many values to unpack in the test

* Add a comment for extended_attention_mask when there is past_key_values

* Fix TFElectra extended_attention_mask when past_key_values is provided

* Add tests to other TF models

* Fix for TF Electra test: add prepare_config_and_inputs_for_decoder

* Fix not passing training arg to lm_head in TFRobertaForCausalLM

* Fix tests (with past) for TF Roberta

* add testing for pask_key_values for TFElectra model

Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>

[examples/summarization and translation] fix readme (#15833)

Add ONNX Runtime quantization for text classification notebook (#15817)

Re-enable doctests for the quicktour (#15828)

* Re-enable doctests for the quicktour

* Re-enable doctests for task_summary (#15830)

* Remove &

Framework split model report (#15825)

Add TFConvNextModel (#15750)

* feat: initial implementation of convnext in tensorflow.

* fix: sample code for the classification model.

* chore: added checked for  from the classification model.

* chore: set bias initializer in the classification head.

* chore: updated license terms.

* chore: removed ununsed imports

* feat: enabled  argument during using drop_path.

* chore: replaced tf.identity with layers.Activation(linear).

* chore: edited default checkpoint.

* fix: minor bugs in the initializations.

* partial-fix: tf model errors for loading pretrained pt weights.

* partial-fix: call method updated

* partial-fix: cross loading of weights (4x3 variables to be matched)

* chore: removed unneeded comment.

* removed playground.py

* rebasing

* rebasing and removing playground.py.

* fix: renaming TFConvNextStage conv and layer norm layers

* chore: added initializers and other minor additions.

* chore: added initializers and other minor additions.

* add: tests for convnext.

* fix: integration tester class.

* fix: issues mentioned in pr feedback (round 1).

* fix: how output_hidden_states arg is propoagated inside the network.

* feat: handling of  arg for pure cnn models.

* chore: added a note on equal contribution in model docs.

* rebasing

* rebasing and removing playground.py.

* feat: encapsulation for the convnext trunk.

* Fix variable naming; Test-related corrections; Run make fixup

* chore: added Joao as a contributor to convnext.

* rebasing

* rebasing and removing playground.py.

* rebasing

* rebasing and removing playground.py.

* chore: corrected copyright year and added comment on NHWC.

* chore: fixed the black version and ran formatting.

* chore: ran make style.

* chore: removed from_pt argument from test, ran make style.

* rebasing

* rebasing and removing playground.py.

* rebasing

* rebasing and removing playground.py.

* fix: tests in the convnext subclass, ran make style.

* rebasing

* rebasing and removing playground.py.

* rebasing

* rebasing and removing playground.py.

* chore: moved convnext test to the correct location

* fix: locations for the test file of convnext.

* fix: convnext tests.

* chore: applied  sgugger's suggestion for dealing w/ output_attentions.

* chore: added comments.

* chore: applied updated quality enviornment style.

* chore: applied formatting with quality enviornment.

* chore: revert to the previous tests/test_modeling_common.py.

* chore: revert to the original test_modeling_common.py

* chore: revert to previous states for test_modeling_tf_common.py and modeling_tf_utils.py

* fix: tests for convnext.

* chore: removed output_attentions argument from convnext config.

* chore: revert to the earlier tf utils.

* fix: output shapes of the hidden states

* chore: removed unnecessary comment

* chore: reverting to the right test_modeling_tf_common.py.

* Styling nits

Co-authored-by: ariG23498 <aritra.born2fly@gmail.com>
Co-authored-by: Joao Gante <joao@huggingface.co>
Co-authored-by: Sylvain Gugger <Sylvain.gugger@gmail.com>

* minor changes

* doc fix in feature extractor

* doc

* typose

* removed detr logic from config

* removed detr logic from config

* removed num_labels

* small fix in the config

* auxilary -> auxiliary

* make style

* some test is failing

* fix a weird char in config prevending doc-builder

* retry to fix the doc-builder issue

* make style

* new try to fix the doc builder

* CI

* change weights to facebook

Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: ariG23498 <aritra.born2fly@gmail.com>
Co-authored-by: Joao Gante <joao@huggingface.co>
Co-authored-by: Sylvain Gugger <Sylvain.gugger@gmail.com>
2022-03-02 15:48:20 +01:00
..
internal TF generate refactor - Greedy Search (#15562) 2022-02-15 17:54:43 +01:00
main_classes Adding ZeroShotImageClassificationPipeline (#12119) 2022-02-23 09:41:42 +01:00
model_doc Maskformer (#15682) 2022-03-02 15:48:20 +01:00
tasks 🧼 NLP task guides (#15731) 2022-02-23 13:58:33 -06:00
_config.py Prevent style_doc from tempering the config file 2021-12-10 15:31:43 -05:00
_toctree.yml Maskformer (#15682) 2022-03-02 15:48:20 +01:00
accelerate.mdx Fix code format for Accelerate doc (#15335) 2022-01-27 13:49:04 -06:00
add_new_model.mdx added link to our writing-doc document (#15756) 2022-02-22 09:57:28 +01:00
add_new_pipeline.mdx Doc styler examples (#14953) 2021-12-27 19:07:46 -05:00
autoclass_tutorial.mdx Update tutorial docs (#15165) 2022-02-01 18:31:35 -06:00
benchmarks.mdx [Benchmark tools] Deprecate all (#15848) 2022-03-01 11:26:20 +01:00
bertology.mdx Convert rst files (#14888) 2021-12-22 16:14:35 -05:00
community.mdx add t5 ner finetuning (#15432) 2022-01-31 17:03:06 +01:00
contributing.md Update installation page and add contributing to the doc (#5084) 2020-06-17 14:01:10 -04:00
converting_tensorflow_models.mdx Convert rst files (#14888) 2021-12-22 16:14:35 -05:00
create_a_model.mdx Create a custom model guide (#15489) 2022-02-07 12:34:56 -06:00
custom_datasets.mdx Added missing code in exemplary notebook - custom datasets fine-tuning (#15300) 2022-01-25 17:26:17 -05:00
custom_models.mdx [doc] custom_models: mention security features of the Hub (#15768) 2022-02-23 11:40:06 -05:00
debugging.mdx add a network debug script and document it (#15652) 2022-02-15 08:48:00 -08:00
examples.md per_device instead of per_gpu/error thrown when argument unknown (#4618) 2020-05-27 11:36:55 -04:00
fast_tokenizers.mdx Convert rst files (#14888) 2021-12-22 16:14:35 -05:00
glossary.mdx Doc styler examples (#14953) 2021-12-27 19:07:46 -05:00
index.mdx Maskformer (#15682) 2022-03-02 15:48:20 +01:00
installation.mdx Get started docs (#15098) 2022-01-28 19:01:37 -06:00
migration.mdx Doc styler examples (#14953) 2021-12-27 19:07:46 -05:00
model_sharing.mdx Update model share tutorial (#15288) 2022-01-28 18:49:26 -06:00
model_summary.mdx Add "open in hf spaces" gradio button issue #73 (#15106) 2022-01-14 10:12:30 -05:00
multilingual.mdx Inference for multilingual models (#15836) 2022-03-01 15:10:31 -06:00
notebooks.md Update notebooks (#3620) 2020-04-06 14:32:39 -04:00
parallelism.mdx [deepspeed docs] Megatron-Deepspeed info (#15488) 2022-02-04 11:15:13 -08:00
performance.mdx add model scaling section (#15119) 2022-02-09 15:27:30 +01:00
perplexity.mdx Doc styler examples (#14953) 2021-12-27 19:07:46 -05:00
philosophy.mdx Convert rst files (#14888) 2021-12-22 16:14:35 -05:00
pipeline_tutorial.mdx Update tutorial docs (#15165) 2022-02-01 18:31:35 -06:00
pr_checks.mdx [docs] fix wrong file name in pr_check (#15380) 2022-01-28 07:52:01 -05:00
preprocessing.mdx Update tutorial docs (#15165) 2022-02-01 18:31:35 -06:00
quicktour.mdx Re-enable doctests for the quicktour (#15828) 2022-02-25 17:46:38 +01:00
sagemaker.mdx Convert rst files (#14888) 2021-12-22 16:14:35 -05:00
serialization.mdx M2M100 support for ONNX export (#15193) 2022-03-02 10:03:14 +01:00
task_summary.mdx Re-enable doctests for the quicktour (#15828) 2022-02-25 17:46:38 +01:00
testing.mdx [doc] normalize HF Transformers string (#15023) 2022-01-10 08:44:33 -08:00
tokenizer_summary.mdx Fix grammar in tokenizer_summary (#15614) 2022-02-11 16:51:30 -05:00
training.mdx Update fine-tune docs (#15259) 2022-02-01 18:28:12 -06:00
troubleshooting.mdx Convert rst files (#14888) 2021-12-22 16:14:35 -05:00