* Do not prepare lr scheduler as it as the right number of steps
* Trigger CI
* Trigger CI
* Trigger CI
* Add fake comment
* Remove fake comment
* Trigger CI please!
* Let's see if we can use the smallest possible dummies
* Make GPT-2's dummies a little longer
* Just use (1,2) as the default shape
* Update other dummies in sync
* Correct imports for Keras 2.13
* Shrink the Wav2Vec2 dummies
* Add support for non-rust implemented tokenization for `__getitem__` method.
* Update for error message on adding new sub-branch for `__item__` method.
---------
Co-authored-by: liuyang17 <liuyang17@zhihu.com>
* Fix model load when it has both code on the Hub and locally
* Add input check with timeout
* Add tests
* Apply suggestions from code review
Co-authored-by: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
* Some non-saved stuff
* Add feature extractors
* Add image processor
* Add model
* Add processor and tokenizer
* Reduce timeout
---------
Co-authored-by: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
* Use new parametrization based weight norm if available
See https://github.com/pytorch/pytorch/pull/103001
Signed-off-by: Edward Z. Yang <ezyang@meta.com>
* handle copies
Signed-off-by: Edward Z. Yang <ezyang@meta.com>
* black
Signed-off-by: Edward Z. Yang <ezyang@meta.com>
---------
Signed-off-by: Edward Z. Yang <ezyang@meta.com>
* A fun new PR where I break the entire codebase again
* A fun new PR where I break the entire codebase again
* Handle cross-attention
* Move calls to model(model.dummy_inputs) to the new build() method
* Seeing what fails with the build context thing
* make fix-copies
* Let's see what fails with new build methods
* Fix the pytorch crossload build calls
* Fix the overridden build methods in vision_text_dual_encoder
* Make sure all our build methods set self.built or call super().build(), which also sets it
* make fix-copies
* Remove finished TODO
* Tentatively remove unneeded (?) line
* Transpose b in deberta correctly and remove unused threading local
* Get rid of build_with_dummies and all it stands for
* Rollback some changes to TF-PT crossloading
* Correctly call super().build()
* Add test_backbone for convnext
* Add TimmBackbone model
* Add check for backbone type
* Tidying up - config checks
* Update convnextv2
* Tidy up
* Fix indices & clearer comment
* Exceptions for config checks
* Correclty update config for tests
* Safer imports
* Safer safer imports
* Fix where decorators go
* Update import logic and backbone tests
* More import fixes
* Fixup
* Only import all_models if torch available
* Fix kwarg updates in from_pretrained & main rebase
* Tidy up
* Add tests for AutoBackbone
* Tidy up
* Fix import error
* Fix up
* Install nattan in doc_test_job
* Revert back to setting self._out_xxx directly
* Bug fix - out_indices mapping from out_features
* Fix tests
* Dont accept output_loading_info for Timm models
* Set out_xxx and don't remap
* Use smaller checkpoint for test
* Don't remap timm indices - check out_indices based on stage names
* Skip test as it's n/a
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Cleaner imports / spelling is hard
---------
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* docs: ko: tasks_explained.mdx
* feat: nmt and manual edit `tasks_explained.mdx`
* revised: resolve suggestions task_explained.mdx
* fixed: added draft of reference docs
Co-Authored-By: Kihoon Son <75935546+KIHOON71@users.noreply.github.com>
Co-Authored-By: Nayeon Han <nayeon2.han@gmail.com>
* revised: resolve suggestions(voca, spell check) task_explained.mdx
Co-Authored-By: Sohyun Sim <96299403+sim-so@users.noreply.github.com>
* revised: remove duplicate sentence in task_explained.mdx
* fixed: remove draft of reference docs
- I think it will be confusing in the translation process.
- This issue is included in #23971.
---------
Co-authored-by: Kihoon Son <75935546+KIHOON71@users.noreply.github.com>
Co-authored-by: Nayeon Han <nayeon2.han@gmail.com>
Co-authored-by: Sohyun Sim <96299403+sim-so@users.noreply.github.com>
expose safe_serialization argument of PreTrainedModel and TFPreTrainedModel in the save_pretrained of the pipeline api
Co-authored-by: Yessen Kanapin <yessen@deepinfra.com>