NielsRogge
4ed89d48ab
Add UperNet ( #20648 )
...
* First draft
* More improvements
* Add convnext backbone
* Add conversion script
* Add more improvements
* Comment out to_dict
* Add to_dict method
* Add default config
* Fix config
* Fix backbone
* Fix backbone some more
* Add docs, auto mapping, tests
* Fix some tests
* Fix more tests
* Fix more tests
* Add conversion script
* Improve conversion script
* Add support for getting reshaped undownsampled hidden states
* Fix forward pass
* Add print statements
* Comment out set_shift_and_window_size
* More improvements
* Correct downsampling layers conversion
* Fix style
* First draft
* Fix conversion script
* Remove config attribute
* Fix more tests
* Update READMEs
* Update ConvNextBackbone
* Fix ConvNext tests
* Align ConvNext with Swin
* Remove files
* Fix index
* Improve docs
* Add output_attentions to model forward
* Add backbone mixin, improve tests
* More improvements
* Update init_weights
* Fix interpolation of logits
* Add UperNetImageProcessor
* Improve image processor
* Fix image processor
* Remove print statements
* Remove script
* Update import
* Add image processor tests
* Remove print statements
* Fix test
* Add integration test
* Add convnext integration test
* Update docstring
* Fix README
* Simplify config
* Apply suggestions
* Improve docs
* Rename class
* Fix test_initialization
* Fix import
* Address review
* Fix confg
* Convert all checkpoints
* Fix default backbone
* Usage same processor as segformer
* Apply suggestions
* Fix init_weights, update conversion scripts
* Improve config
* Use Auto API instead of creating a new image processor
* Fix docs
* Add doctests
* Remove ResNetConfig dependency
* Add always_partition argument
* Fix rebaseé
* Improve docs
* Convert checkpoints
Co-authored-by: Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
Co-authored-by: Niels Rogge <nielsrogge@Nielss-MBP.localdomain>
2023-01-16 09:39:13 +01:00
Shogo Hida
7f65d2366a
Add Spanish translation to community.mdx ( #21055 )
...
* Add community to toctree
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Copy English content
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Add some translations
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Add some translations
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Add some translations
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Fix position of community
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Fix translation
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Add translation
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Add translation
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Add translation
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
* Add translation
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
Signed-off-by: Shogo Hida <shogo.hida@gmail.com>
2023-01-14 09:25:05 +01:00
Steven Liu
f58248b824
Update task summary part 1 ( #21014 )
...
* first draft of new task summary
* make style
* review
* apply feedback
* apply feedbacks
* final touches
2023-01-13 11:01:53 -08:00
Steven Liu
8f796960f6
Fix header level ( #21072 )
...
fix header level
2023-01-10 10:24:10 -08:00
Sayak Paul
263fd3c4c7
add: task guide on video classification model fine-tuning. ( #20827 )
...
* add: task guide on video classification model fine-tuning.
* apply make style from hf-formatting.
* add: toc entry.
* chore: address PR comments.
Co-authored-by Maria Khalusova
* Reflect Maria's contributions.
Co-authored-by: Maria Khalusova <1065417+MKhalusova@users.noreply.github.com>
* chore: minor correction.
* Apply suggestions from code review
Co-authored-by: Nathan Raw <nxr9266@g.rit.edu>
* PyTorch Video -> PyTorchVideo.
* Apply suggestions from code review
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* change licensing year.
* minor rewording.
* apply make style.
* address Sylvain's comments.
* replace links.
Co-authored-by: Maria Khalusova <1065417+MKhalusova@users.noreply.github.com>
Co-authored-by: Nathan Raw <nxr9266@g.rit.edu>
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2023-01-05 00:43:40 +05:30
Maria Khalusova
b493fee958
Add: doc page for the object detection task ( #20925 )
...
* Added Object Detection task guide (new branch)
* Polished code examples after running make style
* Update docs/source/en/tasks/object_detection.mdx
Rephrasing suggestion from Sayak
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update docs/source/en/tasks/object_detection.mdx
A rephrasing suggestion from Sayak
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update docs/source/en/tasks/object_detection.mdx
typo
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Applied reviewers suggestions
>
>
Co-authored-by: sayakpaul <spsayakpaul@gmail.com>
Co-authored-by: sgugger <sylvain.gugger@gmail.com>
* polished code examples
* Added a visualization of the inference result. Slightly changed hyperparameters, and updated the results.
* polished code examples
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update docs/source/en/tasks/object_detection.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Applying Steven's review suggestions
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* minor punctuation fix
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
2023-01-04 08:36:37 -05:00
Jongjyh
ce85686a1f
Add AltCLIP ( #20446 )
...
* add altclip
* update
* fix wrong title
* fix the copyright in readme
* add altclip model
* add altclip
* fix test_gradient_checkpointing_enable_disable
* code
* add return class
* add projection_state
* "fix pretrained model bug"
* delete print and fix 2 test instances.
* delete token
* rm xlmr
* one model one file.
* empty commit to trigger CI
* Fix modeling_outputs.py
* Fix __init__
* Fix quality
* Fix modeling file docstring
* Fix README.md
* Fix test file
* add vision model
* empty commit to trigger CI
* fix
* fix
* fix
* fix
* fix
* fix
* fix
* fix
* fix
* del token in mdx file
* fix
* fix
* fix
* remove altrob from test list
* add vision test
* fix fx
* fix
* fix
* fix
* trigger CI
* fix copies
* fix tests
* fix style
* fix quality
* update
* recover import
* recover
* add ,
* recover
* fix copies
* trigger CI
* fix
* some of review
* update
* remove import
* last 2
* fix
* fix style
* fix style
* fix bug
* fix uncomment
* fix
* update
* fix
* second review
* empty commit to trigger CI
* empty commit to trigger CI
* fix position
* fix
* empty commit to trigger CI
* empty commit to trigger CI
* third comment
* Update docs/source/en/model_doc/altclip.mdx
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
* Update docs/source/en/model_doc/altclip.mdx
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
* Update src/transformers/__init__.py
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
* Update src/transformers/models/altclip/configuration_altclip.py
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
* Update src/transformers/models/altclip/modeling_altclip.py
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
* Update src/transformers/models/altclip/processing_altclip.py
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
* Update src/transformers/models/altclip/modeling_altclip.py
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
* fix merge
* fix copies
* update
* update
* empty commit to trigger CI
* fix code example
* empty commit to trigger CI
* fix
* empty commit to trigger CI
* empty commit to trigger CI
Co-authored-by: shunxing1234 <xw747777271@gmail.com>
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Co-authored-by: shunxing1234 <33774367+shunxing1234@users.noreply.github.com>
Co-authored-by: Yih-Dar <2521628+ydshieh@users.noreply.github.com>
2023-01-04 09:18:57 +01:00
Alara Dirik
cd2457809f
Improve OWL-ViT postprocessing ( #20980 )
...
* add post_process_object_detection method
* style changes
2023-01-03 19:25:09 +03:00
NielsRogge
9c6f7485a6
Add GIT (GenerativeImage2Text) ( #20295 )
...
* First draft
* Make model instantiation work
* Fix copied from statement
* More fixes
* Add correct output head
* Improve configuration
* Add conversion script
* Improve conversion script
* Remove token_type_ids
* Fix conversion of projection layers
* Convert all weights
* Use cats image
* Make logits match
* Generate caption on cats image
* Add GITProcessor
* Update conversion script
* Add support for more checkpoints
* Fix conversion script
* Add initial tests
* Remove cross-attention
* More improvements
* Remove is_decoder
* Improve model tests
* Improve tests
* Improve model outputs
* Fix model outputs equivalence
* Fix more tests
* Remove unused code
* Use generate to generate text, no use of cache for now
* Use generate more appropriately
* Fix config tests
* Fix style
* Add support for use_cache
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* Fix style
* Fix GIT vision encoder
* Update README
* Fix integration test
* Set bos and eos token ids
* Improve docs
* Improve code
* Add support for provided attention_mask
* Add copied from statement
* Fix gradient checkpointing test
* Set model_input_names
* Investigate model_input_names
* Remove script
* Fix model inputs
* Fix docstring
* Rename GIT to Git
* Support more models
* Add support for textvqa model
* Add video support
* Extend conversion script for video
* Add support for large variant
* Add support for more models
* Fix config archive map
* Update integration test
* Fix README
* Fix CLIP mean and std
* Update processor
* Fix use_cache for video, thanks @gante
* Remove print statements
* Remove assertion
* Add processor tests
* Fix model_input_names
* Use Auto API for processor
* Fix processor tests
* Fix integration test
* Fix pipeline test
* Make tests faster
* Update conversion script
* Update conversion script
* Convert more checkpoints
* Update conversion script
* Fix typo
* Update docstrings
* Improve code snippets
* Fix doc tests
* Add more code examplesé
* Fix doc tests
* Add integration tests
* Fix unused variable
* revert
* Add GIT to Japanese README
Co-authored-by: Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2023-01-03 14:17:18 +01:00
Konstantin Kotik
367fdf3330
MinNewTokensLengthLogitsProcessor
for .generate
method #20814 ( #20892 )
...
* feat: add min new length logit processor
* test: add min new length logit processor
* docs: add MinNewTokensLengthLogitsProcessor
* feat: import MinNewTokensLengthLogitsProcessor
* fix: update pytorch dummy objects
* refactor & fix: rename attributes and var and get rid of dynamic attribute
* tests: align test with new interface
* docs: fix typo
* docs: minor clarification
* Empty-Commit
* empty commit
* run automated quality edits
Co-authored-by: Joao Gante <joao@huggingface.co>
2023-01-03 06:29:02 -05:00
Alex Hedges
0b686a8a1e
Remove non-breaking spaces ( #20929 )
...
* Remove non-breaking space in comment
It was likely added unintionally.
* Remove remaining non-breaking spaces
2022-12-29 02:12:40 -05:00
Yih-Dar
5fa0b17c3d
[Past CI] 🔥 Leave Past CI failures in the past 🔥 ( #20861 )
...
* torch.jit._state
* Fix past CI
* Fix for perceiver
* Fix REALM
* Fix for Bloom
* Fix for SwinMode
* Fix for TrajectoryTransformerModel
* Fix for test_wav2vec2_with_lm
* make style
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2022-12-27 18:37:25 +01:00
Eli Simhayev
e35bc46af6
fix docs typos in "add_new_model" ( #20900 )
...
fix Jupyter typos
2022-12-27 02:49:15 -05:00
Kamal Raj Kanakarajan
d1b3011292
Update flan-t5 original model link ( #20897 )
...
Update flan-t5.mdx
2022-12-27 02:26:14 -05:00
Nathan Barry
47146721b8
typo fix ( #20891 )
2022-12-26 02:06:23 -05:00
Syed Abdul Gaffar Shakhadri
15bc776fec
Add Onnx Config for PoolFormer ( #20868 )
...
poolformer onnx
Co-authored-by: syed <syed.abdul@sandlogic.com>
2022-12-23 01:30:57 -05:00
Maria Khalusova
04c560225b
Adding evaluate
to the list of libraries required in generated notebooks ( #20850 )
...
Adding `evaluate` to the list of libraries to be installed for every generated notebook in transformers
2022-12-21 14:04:08 +01:00
Younes Belkada
0d284bd574
Add BLIP ( #20716 )
...
* add new model like
* add v1
* v1
* v1
* vision encoder logits match
* v2
* fix
* add docstring
* CI tests pass
* fix tests
* make fixup
* add to `toctree`
* fix processors
* fix processors
* fix doc
* fill title
* add content doc
* remove from tokenization auto
* fix config
* change order
* add `# Copied from`
* few fixes
- add correct license on modeling text
- remove dummy argument
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* replace name
* refactor a bit
* more refactor
* remove unused arg
* make fixup + remove some `# Adapted from ...`
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* more `# Copied from`
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* now `generate` supports no prefix
* remove `FeatureExtractor`
* fix path
* correct dependency
* fix tests
* few fixes
* add integration tests
* add correct conversion script
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* add `blip` to tokenization auto
* fix docstrings
* fix test + add image
* remove processor from uncorrect place
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* clean up a bit
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* clean pixel mask
* clean pixel mask
* fix `F`
* Update src/transformers/models/blip/modeling_blip.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fix output
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fix pad token id
* remove `token_type_ids`
* make fixup
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* make fixup
* Apply suggestions from code review
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* add comments
* Update src/transformers/models/blip/modeling_blip.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* remove `token_type_ids`
* make fixup
* better name
* replace with `image_attention_mask`
* refactor
* make fixup
* better docstring
* replace `answer_xx`
* remove ununsed args
* add `labels`
* add `labels`
* fix processing tests
* make fixup
* make fixup
* put correct repo
* remove `pad`
* remove `crop` and `center_crop`
* Update src/transformers/models/blip/image_processing_blip.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fix
* remove `size_divisor`
* fix weights `init`
* remove unneeded functions
* add suggestions
* minor changes
- change slow test output for PT 1.13
- docstring order
* replace `feature_extractor` by `image_processor`
* fix doctests
* fix weight init order + add fp16 slow test
* add `blip` to doctest
* add correct repo name and fix test
* Update src/transformers/models/blip/processing_blip.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fix tests
* use `convert_to_rgb` from `image_transforms`
* make fixup
* fix large loading issue
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-12-21 09:39:10 +01:00
Steven Liu
3be028bc9d
Embed circle packing chart for model summary ( #20791 )
...
* embed circle packing chart
* trim whitespace from bottom
* explain bubble sizes
2022-12-20 10:26:52 -08:00
stanleycai95
bdb84e2bad
Add model resources for ViT ( #20723 )
...
* Set up overall resources documentation structure
* Update vit.mdx
* Removing irrelevant sections on text models
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
* Update vit.mdx
2022-12-19 10:59:34 -08:00
Andreas Madsen
b4b613b102
Implement Roberta PreLayerNorm ( #20305 )
...
* Copy RoBERTa
* formatting
* implement RoBERTa with prelayer normalization
* update test expectations
* add documentation
* add convertion script for DinkyTrain weights
* update checkpoint repo
Unfortunately the original checkpoints assumes a hacked roberta model
* add to RoBERTa-PreLayerNorm docs to toc
* run utils/check_copies.py
* lint files
* remove unused import
* fix check_repo reporting wrongly a test is missing
* fix import error, caused by rebase
* run make fix-copies
* add RobertaPreLayerNormConfig to ROBERTA_EMBEDDING_ADJUSMENT_CONFIGS
* Fix documentation <Facebook> -> Facebook
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* fixup: Fix documentation <Facebook> -> Facebook
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Add missing Flax header
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* expected_slice -> EXPECTED_SLICE
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* update copies after rebase
* add missing copied from statements
* make fix-copies
* make prelayernorm explicit in code
* fix checkpoint path for the original implementation
* add flax integration tests
* improve docs
* update utils/documentation_tests.txt
* lint files
* Remove Copyright notice
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* make fix-copies
* Remove EXPECTED_SLICE calculation comments
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-12-19 09:30:17 +01:00
NielsRogge
26dd041c6e
Add Swin2SR ( #19784 )
...
* First draft
* Add more improvements
* Improve forward pass
* Fix layernorm
* Add upscaler
* More improvements
* More improvements
* More improvements
* Improve conversion script
* Add preprocessing
* Make output match original implementation
* Add additional attributes
* Add support for more models
* Support more models
* Add support for real world sr
* Add initial Swin2SRFeatureExtractor
* Add ImageSuperResolutionOutput
* Make more tests pass
* Use BaseModelOutput
* Fix one more test
* Fix more tests
* Fix another test
* Fix all tests
* Rename to Swin2SRImageProcessor
* Fix toctree
* Fix toctree
* Fix rebase
* Improve Swin2SRImageProcessor
* Remove feature extractor file
* Improve model
* Improve conversion script
* Fix integration test
* Fix init
* Fix conversion script
* Address comments
* Improve upsampler
* Add NearestConvUpsampler
* Improve pixel shuffle upsampler
* Improve auxiliary upsampler
* Improve conversion script
* Rename conv_last to final_convolution
* Fix rebase
* Improve upsample module
* Add padding to image processor
* Fix bug
* Update padding
* Remove print statement and fix integration test
* Improve docs
* Add image processor tests
* Convert all checkpoints, fix testsé
* Remove print statements
* Fix import
Co-authored-by: Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
2022-12-16 16:24:01 +01:00
NielsRogge
7f99861218
Add Universal Segmentation class + mapping ( #20766 )
...
* Add mapping
* Add mapping to pipeline
* Apply suggestions
* Fix feature extractor tests
* Use ForInstance, add model to universal mapping
* More fixes
* Remove model from deprecated objectsé
Co-authored-by: Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
2022-12-16 14:22:46 +01:00
Joao Gante
4bc723f87d
Generate: use GenerationConfig
as the basis for .generate()
parametrization ( #20388 )
...
* generate from config mvp
* fix failing tests
* max_time test
* Load default gen config at model load time; Update docs
* further documentation; add tests
* adapt rag to the new structure
* handle models not instantiated with from_pretained (like in tests)
* better default generation config
* add can_generate fn
* handle legacy use case of ad hoc model config changes
* initialize gen config from config in individual methods, if gen config is none
* fix _get_decoder_start_token_id when called outside GenerationMixin
* correct model config load order (set attr > model config > decoder config)
* update rag to match latest changes
* Apply suggestions from code review
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* load gen config from model config in model.from_pretrained
* fix can_generate fn
* handle generate calls without a previous from_pretrained (e.g. tests)
* add legacy behavior (and a warning)
* lower logger severity
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-12-15 18:27:20 +00:00
Nicolas Patry
ba9da49aa2
Fixing the pipeline tutorial test ( #20746 )
...
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2022-12-13 19:08:30 +01:00
Hazrul Akmal
f28c918c7e
Add docs xlm roberta ( #20742 )
...
* added model resources for xlm-roberta
* added model resources for xlm-roberta
* resolve suggested changes
* add resources to xlm-roberta
2022-12-13 09:25:55 -08:00
Ariel Ekgren
5f94855dc3
Add gpt-sw3 model to transformers ( #20209 )
...
* Add templates for gpt-sw3
* Add templates for gpt-sw3
* Added sentencepiece tokenizer
* intermediate commit with many changes
* fixed conflicts
* Init commit for tokenization port
* Tokenization progress
* Remove fast tokenizer
* Clean up and rename spm.model -> spiece.model
* Remove TF -> PT conversion script template, Clean up Megatron -> PT script
* Optimize encode & decode performance
* added new attention
* added new attention
* attention for gpt-sw3 working
* attention good
* Cache is now working
* fixed attention mask so that it works with causal attention
* fixed badbmm bug for cpu and caching
* updated config with correct parameters
* Refactor and leave optimizations as separate functions to avoid breaking expected functionality
* Fix special tokens mapping for both tokenizers
* cleaning up of code and comments
* HF compatible attention outputs
* Tokenizer now passing tests, add documentation
* Update documentation
* reverted back to base implementation after checking that it is identical to pretrained model
* updated gpt-sw3 config
* updated conversion script
* aligned parameters with gpt-sw3 config
* changed default scale_attn_by_inverse_layer_idx to true
* removed flag from conversion script
* added temporary model path
* reverted back to functioning convert script
* small changes to default config
* updated tests for gpt-sw3
* make style, make quality, minor cleanup
* Change local paths to testing online repository
* Change name: GptSw3 -> GPTSw3
* Remove GPTSw3TokenizerFast references
* Use official model repository and add more model sizes
* Added reference to 6.7b model
* Add GPTSw3DoubleHeadsModel to IGNORE_NON_AUTO_CONFIGURED, like GPT2DoubleHeadsModel
* Remove pointers to non-existing TFGPTSw3
* Add GPTSw3 to docs/_toctree.yml
* Remove TF artifacts from GPTSw3 in __init__ files
* Update README:s with 'make fix-copies'
* Add 20b model to archive list
* Add documentation for GPT-Sw3
* Fix typo in documentation for GPT-Sw3
* Do 'make fix-copies' again after having updated docs
* Fix some typos in docs
* Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/__init__.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/__init__.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update tests/models/gpt_sw3/test_tokenization_gpt_sw3.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Resolve comments from PR feedback
* Resolve more comments from PR feedback, also set use_cache=True in convert script
* Add '# Copied from' comments for GPTSw3 modeling
* Set 'is_parallelizable = False'
* Remove '# Copied from' where code was modified and add 'with x->y' when appropriate
* Remove parallelize in mdx
* make style, make quality
* Update GPTSw3Config default values and corresponding documentation
* Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/__init__.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Clean up and protect GPTSw3Tokenizer imports with is_sentencepiece_available
* Make style, make quality
* Add dummy object for GPTSw3Tokenizer via 'make fix-copies'
* make fix-copies
* Remove GPTSw3 modeling classes
* make style, make quality
* Add GPTSw3 auto-mappings for other GPT2 heads
* Update docs/source/en/model_doc/gpt-sw3.mdx
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Remove old TODO-comment
* Add example usage to GPTSw3Tokenizer docstring
* make style, make quality
* Add implementation details and example usage to gpt-sw3.mdx
Co-authored-by: JoeyOhman <joeyoh@kth.se>
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-12-12 13:12:13 -05:00
Matt
c1b9a11dd4
Convert tokenizer outputs for Keras in doc example ( #20732 )
...
* Convert tokenizer outputs for Keras in doc example
* Das deutsche Beispiel auch korrigieren
2022-12-12 16:14:04 +00:00
Juanjo do Olmo
0ba94aceb6
Spanish translation of the file debugging.mdx ( #20566 )
...
* Create and translate to Spanish debugging.mdx
* solved typo error in a header
* Update debugging.mdx
* Update debugging.mdx
* Update docs/source/es/debugging.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/debugging.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/debugging.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/debugging.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/debugging.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update _toctree.yml
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-12-12 10:38:56 -05:00
stanleycai95
17c742bbf5
Very small edit to change name to OpenAI GPT ( #20722 )
2022-12-12 09:43:43 -05:00
Alberto Mario Ceballos-Arroyo
8286af6f54
Spanish translation of asr.mdx and add_new_pipeline.mdx ( #20569 )
...
* Fix minor typo in question_answering.mdx
* Fixes minor typo in the english version of tasks/asr.mdx
* Update _toctree.yml
* Translate add_new_pipeline.mdx into Spanish
* Fixes some typos in the English version of add_new_pipeline.mdx
* Translate asr.mdx into Spanish
* Fixes small typos in add_new_pipeline.mdx
* Update docs/source/es/add_new_pipeline.mdx
Suggestion by @osanseviero
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/add_new_pipeline.mdx
Suggestion by @osanseviero: use "biblioteca" instead of "librería."
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/tasks/asr.mdx
Suggestion by @osanseviero.
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/add_new_pipeline.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/add_new_pipeline.mdx
Suggestion by @osanseviero.
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/add_new_pipeline.mdx
Suggestion by @osanseviero.
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/add_new_pipeline.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/tasks/asr.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/tasks/asr.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update docs/source/es/tasks/asr.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
* Update asr.mdx
Co-authored-by: Omar Sanseviero <osanseviero@gmail.com>
2022-12-12 09:23:23 -05:00
Sylvain Gugger
799cea64ac
Fix rendering issue in quicktour ( #20708 )
...
* Fix rendering issue in quicktour
* Separate in two blocks
2022-12-09 13:51:35 -05:00
Michael Benayoun
6a062a3ed9
Change transformers.onnx to use optimum.exporters.onnx ( #20529 )
...
* Change transformers.onnx to use optimum.exporters.onnx
* Update doc
* Remove print
* Fix transformers.onnx cli
* Update documentation
* Update documentation
* Small fixes
* Fix log message
* Apply suggestions
* Update src/transformers/onnx/__main__.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Apply suggestions
* Add missing line break
* Ran make fix-copies
* Update src/transformers/onnx/__main__.py
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
* Update src/transformers/onnx/__main__.py
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
Co-authored-by: Michael Benayoun <michael@huggingface.co>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
2022-12-09 10:42:02 +01:00
Nathan Raw
9e56aff58a
Add video classification pipeline ( #20151 )
...
* 🚧 wip video classification pipeline
* 🚧 wip - add is_decord_available check
* 🐛 add missing import
* ✅ add tests
* 🔧 add decord to setup extras
* 🚧 add is_decord_available
* ✨ add video-classification pipeline
* 📝 add video classification pipe to docs
* 🐛 add missing VideoClassificationPipeline import
* 📌 add decord install in test runner
* ✅ fix url inputs to video-classification pipeline
* ✨ updates from review
* 📝 add video cls pipeline to docs
* 📝 add docstring
* 🔥 remove unused import
* 🔥 remove some code
* 📝 docfix
2022-12-08 16:22:43 -05:00
Sylvain Gugger
9cc65f8701
Migrate torchdynamo to torch.compile ( #20634 )
...
* Migrate torchdynamo to torch.compile
* Add docstring and generic option
* Properly use the function...
* Reorg args
2022-12-08 11:18:52 -05:00
Cole Howard
fc95386ea1
Add TFBartForSequenceClassification ( #20570 )
...
* read to load
* base functionality
* revert init
* fix dummy data
* moving right along
* moving right along
* finally
* cleanup
* pull out comment
* add test
* update docstring for main class
* flake comments and rewriting copies from make repo-consistency`
* remove irrelevant differences/accidental spaces
* put copies back after space removals
* mid
* final test pass
* stray comment
* update test file
* update test file
* fixup
* black
* missed
* black missed one more
* sytle
* add doc update
* fix order of output class
* comment
* Revert "comment"
This reverts commit 03f86b6948
.
* remove redundant function, and redundant reshape
* move change out of common
* style
* put common spaces back
* reorder kwargs in output
* doc style
2022-12-07 18:05:39 +01:00
NielsRogge
d151a8c550
Add BiT + ViT hybrid ( #20550 )
...
* First draft
* More improvements
* Add backbone, first draft of ViT hybrid
* Add AutoBackbone
* More improvements
* Fix bug
* More improvements
* More improvements
* Convert ViT-hybrid
* More improvements
* add patch bit
* Fix style
* Improve code
* cleaned v1
* more cleaning
* more refactoring
* Improve models, add tests
* Add docs and tests
* Make more tests pass
* Improve default backbone config
* Update model_type
* Fix more tests
* Add more copied from statements
* More improvements
* Add push to hub to conversion scripts
* clean
* more cleanup
* clean
* replace to
* fix
* Update src/transformers/models/bit/configuration_bit.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fix base model prefix
* more cleaning
* get rid of stem
* clean
* replace flag
* Update src/transformers/models/bit/configuration_bit.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/bit/configuration_bit.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* add check
* another check
* fix for hybrid vit
* final fix
* update config
* fix class name
* fix `make fix-copies`
* remove `use_activation`
* Update src/transformers/models/bit/configuration_bit.py
* rm unneeded file
* Add BiT image processor
* rm unneeded file
* add doc
* Add image processor to conversion script
* Add ViTHybrid image processor
* Add resources
* Move bit to correct position
* Fix auto mapping
* Rename hybrid to Hybrid
* Fix name in toctree
* Fix READMEs'
* Improve config
* Simplify GroupNormActivation layer
* fix test + make style
* Improve config
* Apply suggestions from code review
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* remove comment
* remove comment
* replace
* replace
* remove all conv_layer
* refactor norm_layer
* revert x
* add copied from
* last changes + integration tests
* make fixup
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* fix name
* fix message
* remove assert and refactor
* refactor + make fixup
* refactor - add + sfety checker
* fix docstring + checkpoint names
* fix merge issues
* fix function name
* fix copies
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* fix model checkpoint
* fix doctest output
* vit name on doc
* fix name on doc
* fix small nits
* fixed integration tests
* final changes - slow tests pass
Co-authored-by: Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
Co-authored-by: younesbelkada <younesbelkada@gmail.com>
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-12-07 11:03:39 +01:00
Samuel Xu
e842e181df
Documentation fixes ( #20607 )
2022-12-06 07:32:46 -05:00
Nicolas Patry
28f3d431d4
Rework the pipeline tutorial ( #20437 )
...
* [WIP] Rework the pipeline tutorial
- Switch to `asr` instead of another NLP task.
- It also has simpler to understand results.
- Added a section with interaction with `datasets`.
- Added a section with writing a simple webserver.
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* Addressing comments.
* Links.
* Fixing docs format.
* Adding pipeline_webserver to _toctree.
* Warnig -> Tip warnings={true}.
* Fix link ?
* Links ?
* Fixing link, adding chunk batching.
* Oops.
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* Update docs/source/en/pipeline_tutorial.mdx
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
2022-12-06 10:47:31 +01:00
Steven Liu
720e9599c1
Split autoclasses on modality ( #20559 )
...
* split autoclasses on modality
* apply review
* auto classes
2022-12-05 12:28:44 -08:00
Steven Liu
7d1c1c5b21
Fix code sample in preprocess ( #20561 )
...
* change to image_processor
* apply review
2022-12-05 11:49:43 -08:00
Francisco Kurucz
ac3bccdc74
Fix link to Swin Model contributor novice03 ( #20557 )
2022-12-05 11:42:29 -05:00
Erin
87282cb73c
Add RemBERT ONNX config ( #20520 )
...
* rembert onnx config
* formatting
Co-authored-by: Ho <erincho@bcd0745f972b.ant.amazon.com>
2022-12-05 11:39:09 -05:00
Kamal Raj Kanakarajan
13e736685a
Add BioGPT ( #20420 )
...
* biogpt initial commit
* updated init
* fix faster decoding with use_cache
* 1. fix input_ids and input_embeds with correct device
2. added _keys_to_ignore_on_load_missing
3. updated prepare_inputs_for_generation
* add activation_dropout and scale_embedding
* replace fsmt attention with bart attention
* added test
* run make fix-copies
* doc init and fix build
* updated README with proper information
* 1. added tips to docs
2. updated BioGptTokenizer func
* 1. added tokenizer test
2. refactor tokenizer
* make fixup
* add biogpt fairseq to hf converter
* updated layer names more
similar to original checkpoints
* config update doc string and set defaults
* added "#copied" from bart model and
updated doc strings
* enable model_input_names in tokenizer
* 1. positionalembedding depending on attention_mask
2. added attention mask to prepare for generation
* added test to verify past and generation
* BioGptLMHeadModel -> BioGptForCausalLM
* fix typo
* tokenization and test
Copyright and updated assertion
* updated Copyright and
one func at time in line
* Copyright updates and
minor doc fix
* replace assertion with ValueError
* rm extra space
* added code syntax
* revert cmnt position change
* add tokenizer to auto
* updated doc string
* tokenizer doc string update
* biogpt hub model update to microsoft/biogpt
* make fixup
* rm cmnt to fix flake8 5.0.4 vs 6 error
2022-12-05 10:12:03 -05:00
szhublox
699e90437f
flan-t5.mdx: fix link to large model ( #20555 )
2022-12-02 19:27:46 +01:00
fatih
cc3d0e1b01
[New Model] Add TimeSformer model ( #18908 )
...
* init timesformer
* apply fix-copies
* reformat style
* revert back some incoorect style updates
* init timesformer
* apply fix-copies
* reformat style
* revert back some incoorect style updates
* update timseformer doc
* add some functions and classes
* add new config params
* implement multiple classes
* update TimeSformerLayer
* update TimeSformerModel, TimeSformerPreTrainedModel, TimeSformerEncoder
* several fixes
* reformat
* temporary update
* fix some typos
* fix weight converter
* more fixes
* fix a typo
* fix typo
* remove redundant params
* fix for latest hf-hub
* merge fix
* fix some checks
* video classification works with einops
* add paper info to docs
* merge fix
* remove redundant line
* remove redundant docstring
* update config
* fix some typos
* fix converter
* update some test constants
* refactor einops functions
* reformat
* fix a comment
* remove redundat imports
* reformat
* fix a typo
* remove comment
* remove unused imports
* remove redundant doc line
* reformat
* add missing line
* fix docs
* fix timesformer auto feat ext
* add unittests
* reformat
* fix docs
* some fixes and updates
* fix readme
* fix modeling
* fix readme
* update index
* revert _toctree.yml changes
* update timseformer.mdx
* update drop_path_prob to drop_path_rate
* add dosctring for drop_path_rate
* update TimeSformerPatchEmbed naming
* remove to_2tuple
* explicit use of nn.functional
* reformat
* many updates from review comments
* fix a typo
* reformat
* remove assert, better variable name
* make variable names more explicit
* add some adapted from
* more explicit variable names
* remove redundant docstring
* fix initilaization
* move permute inside embedding
* update class names
* remove unused imports
* add test for video classification
* update PretrainedModel with PreTrainedModel
* remove double permute
* update based on sylvain's review
* aply auto fix
* update image_processing_auto for timesformer
* update hub urls
* reformat
* remove duplicate import
* update doc link
2022-12-02 09:13:25 +01:00
Younes Belkada
8b486c0310
add doc for ( #20525 )
2022-12-01 16:52:13 +01:00
Yang An
721764028e
Add Chinese-CLIP implementation ( #20368 )
...
* init chinese-clip model from clip
* init model tests and docs
* implement chinese-clip into hf
* implement chinese-clip into hf
* implement chinese-clip into hf
* implement chinese-clip into hf
* implement chinese-clip into hf
* update usecase example in model implementation
* fix codestyle
* fix model_type typo in readme
* add placeholder in doc
* add placeholder in doc
* update the init script
* update usecase
* fix codestyle
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* update testcase
* forward the convert_rgb
* update testcase
* update testcase
* update testcase
* merge the recent update from clip about model_input_name property
* update the doc
* update the doc
* update the doc
* update the doc
* remove unused imports
* reformat code style
* update the doc
* fix isort style
* bypass a weird failed unit test which is unrelated with my PR
* update the doc
* implement independent vision config class
* implement independent vision model class
* fix refactor bug
* fix refactor bug
* fix refactor bug
* make style
* fix refactor bug
* make style
* fix refactor bug
* fix refactor bug
* make style
* fix refactor bug
* fix refactor bug
* doc-build restyle
* implement independent text config class
* implement independent text model class
* implement independent text model class
* make style
* make fix-copies
* fix refactor bug
* fix refactor bug
* fix refactor bug
* fix refactor bug
* fix refactor bug
* fix refactor bug
* fix refactor bug
* fix refactor bug
* fix refactor bug
* fix refactor bug
* make style
* update doc
* black and isort
* update doc
* Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/auto/tokenization_auto.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* modify the model type from chinese-clip to chinese_clip
* format the example comment of ChineseCLIPVisionConfig
* correct the copyright comment
* fix the tokenizer specification
* add copied from for loss function
* remove unused class
* update CHINESE_CLIP_TEXT_INPUTS_DOCSTRING
* update CHINESE_CLIP_INPUTS_DOCSTRING
* update doc
* update doc
* update code comment in config
* update copied from statement
* make style
* rename the doc file
* add copied statement
* remove unused attention_mask, causal_attention_mask in ChineseCLIPVisionEncoder
* remove ChineseCLIPTextPreTrainedModel
* fix bug
* fix bug
* fix bug
* update doc
* make style
* Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/chinese_clip/configuration_chinese_clip.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* update ChineseCLIPImageProcessor in image_processing_auto
* fix config_class of chinesecliptextmodel
* fix the test case
* update the docs
* remove the copied from comment for ChineseCLIPTextModel, since it has diverged from BertModel with customed config_class
* update the testcase
* final fix
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2022-11-30 19:22:23 +01:00
Sylvain Gugger
08b4621899
Repurpose torchdynamo training args towards torch._dynamo ( #20498 )
...
* Repurpose torchdynamo training args towards torch._dynamo
* Add doc
2022-11-30 11:10:45 -05:00
Julian Pollmann
829374e4fc
Fix Typo in Docs for GPU ( #20509 )
2022-11-30 10:41:18 -05:00