Sylvain Gugger
7e7f62bfa7
Fix pipeline tests for Roberta-like tokenizers ( #19365 )
...
* Fix pipeline tests for Roberta-like tokenizers
* Fix fix
2022-10-05 17:48:14 -04:00
Alara Dirik
bad353cebf
Fix DETR segmentation postprocessing output ( #19363 )
...
Ensures post_process_instance_segmentation and post_process_panoptic_segmentation methods return a tensor of shape (target_height, target_width) filled with -1 values if no segment with score > threshold is found.
2022-10-06 00:16:36 +03:00
Arthur
45e14038f2
Add WhisperModel to transformers ( #19166 )
...
* simplify loop
* add featur extractor
* add model
* start conversion
* add dropout
* initial commit of test files
* copnversion for all models
* update processor for correct padding
* update feature extraction
* update integration test logits match
* fmnt: off for the logits
* on the fly mel bank
* small nit
* update test
* update tokenizer
* nit feature extraction
* update
* update tokenizer test
* adds logit processor and update tokenizer to get supress tokens
* style
* clean convert
* revert to original modeling tf utils
* Update
* update
* nit
* clean convert file
* update tests and nits
* quality
* slow generation test
* ffn_dim to allow customization
* update readme
* add to toctreee
* start fixing integration tests
* update tests and code
* fix feature extractor
* fix config tests common
* update code to fix tests
* fix feature exctractor
* nit feature extraction
* update test for new feature extractor
* style
* add absrtact
* large logits wioth custom decoder input ids
* wraap around is otrch available
* fix feature extractor
* correct logits for whisper small.en
* nit
* fix encoder_attentino_mask
* some fixes
* remove unnecessary inputs
* nits
* add normalizer file
* update etst tokenization
* fix attention mask not defined
* Add model to README
* Fix doc tests
* fix generate
* remove uncoder attention mask useless
* update test modeling whisper
* update condfig to add second non supress tokens
* nits on feature exrtactor
* nit for test tokenizers
* update etsts
* update tests
* update tokenization test
* fixup
* invalidated hf token. Clean convert openai to whisper
* fix logit tests
* fixup
* clean merge
* revert toc_tree changes
* remove useless LogitProcessor
* Update whisper .mdx
* update config file doc
* update configuration docstring
* update test tokenization
* update test tokenization
* update tokenization whisper
Added copied from where needed
* update feature extraction
* nit test name
* style
* quality
* remove get suppress tokens and update non_speech tokens global variables
* Update src/transformers/models/whisper/feature_extraction_whisper.py
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* clean modeling whisper and test
Removed the attention mask arguments that are deprecated
* fix large test
* Add multilingual audio test, and translate test
* style
* fix larg multilingual test
* nits
* Update docs/source/en/model_doc/whisper.mdx
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* add copied from for attention layer
* remove attention masks in doc
* add english normalizer
* update tokenization test
* remove copied from in whisper attention : no bias in k_proj only
* wrap around dependencies in english normalizer
* style
* correct import generation logits
* for now, wrap feature extractor with torch
* Update src/transformers/models/whisper/convert_openai_whisper_to_tfms.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update docs/source/en/model_doc/whisper.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* remove torch depencies for feature extraction and style
* fixup
* nit
* update logitds
* style
* nit
* nits and fix final tests
* add `is_more_itertools_available` to utils
* quality
* add begin supress tokens, supress tokens to generate args and config
* clean supressTokensLogitProcessor in generation logits
* Nit naming
* add supressTokensAtBegin
* udpate tests, supress tokens to None or correct values
* nit and style
* update RAG to fit test and generate_logit
* add copy pasted statment on english normalizer
* add arguments to config_common_kwargs
* Update src/transformers/generation_utils.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/generation_logits_process.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* revert changes based on reviews
* update doc and nits
* more nits
* last nits
* update test configuration common
* add BART name in decoder attention mask documentation
* Update src/transformers/models/whisper/modeling_whisper.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* style
* nit
* nit
* add english.json file to git
* nits on documentation
* nit
* nits
* last styling
* add main toctree file
* remove sentence piece dependency
* clean init file
* fix tokenizer that has no dependencies on sentencepiece
* update whisper init file, nit
* remove english.json file
* add get decoder prompt id
* revert changes and add forced logit processor
* nit
* clean normalizer
* remove protected
* update
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* update based on review
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* add batched tests
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: NielsRogge <niels.rogge1@gmail.com>
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-10-05 22:28:31 +02:00
Alara Dirik
7598791c09
Fix MaskFormer failing postprocess tests ( #19354 )
...
Ensures post_process_instance_segmentation and post_process_panoptic_segmentation methods return a tensor of shape (target_height, target_width) filled with -1 values if no segment with score > threshold is found.
2022-10-05 23:25:58 +03:00
Zachary Mueller
ad98642a82
Fix gather for metrics ( #19360 )
2022-10-05 14:52:01 -04:00
Harsha
d9101b71bc
Removes Roberta and Bert config dependencies from Longformer ( #19343 )
...
* removes roberta and bert config dependencies from longformer
* adds copied from statements
* fixes style
* removes excessive comments and replace bert with longformer in a couple places
* fixes style
2022-10-05 13:50:15 -04:00
Paula Isabel
226b8ef063
correct typos in README ( #19304 )
2022-10-05 10:40:38 -07:00
Matt
071df6eb13
Call _set_save_spec() when creating TF models ( #19321 )
...
* Add a build_from_serving_sig_and_dummies method and replace all calls like model(model.dummy_inputs) with it.
* make fixup
* Remove the overridden save() as this is no longer necessary
* Also call _set_save_spec(), the last missing piece
* Ensure we set the save spec when loading from config too
* Turn this whole thing into a one-line PR
* Turn this whole thing into a one-line PR
* Turn this whole thing into a one-line PR
Co-authored-by: Your Name <you@example.com>
2022-10-05 18:03:49 +01:00
Sylvain Gugger
c875a96eb1
Test failing test while we resolve the issue. ( #19355 )
2022-10-05 12:23:48 -04:00
Younes Belkada
4cbc797b27
Change BloomConfig
docstring ( #19336 )
...
* change `BloomConfig` docstring
- slightly change the docstring of the `BloomConfig`
- Use correct default vocab size
- Use correct default `hidden_dim`, `n_head`
* Update src/transformers/models/bloom/configuration_bloom.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/bloom/configuration_bloom.py
Co-authored-by: SaulLu <55560583+SaulLu@users.noreply.github.com>
* make style
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: SaulLu <55560583+SaulLu@users.noreply.github.com>
2022-10-05 18:12:13 +02:00
Harsha
e794ca5b16
Frees LongformerTokenizer of the Roberta dependency ( #19346 )
...
* copies over roberta tokenizer to longformertokenizer since they are both identical
* adds Copied from patterns to pass copy check
2022-10-05 11:49:14 -04:00
r-terada
2f53ab5745
Add sudachi and jumanpp tokenizers for bert_japanese ( #19043 )
...
* add sudachipy and jumanpp tokenizers for bert_japanese
* use ImportError instead of ModuleNotFoundError in SudachiTokenizer and JumanppTokenizer
* put test cases of test_tokenization_bert_japanese in one line
* add require_sudachi and require_jumanpp decorator for testing
* add sudachi and pyknp(jumanpp) to dependencies
* remove sudachi_dict_small and sudachi_dict_full from dependencies
* empty commit for ci
2022-10-05 11:41:37 -04:00
mustapha ajeghrir
60db81ff60
Making camembert independent from roberta, clean ( #19337 )
...
Co-authored-by: Mustapha AJEGHRIR <mustapha.ajeghrir@kleegroup.com>
2022-10-05 09:31:33 -04:00
Druhin Abrol
c54bb1ad79
[WIP]remove XLMTokenizer inheritance from FlaubertTokenizer ( #19330 )
...
* remove XLMTokenizer inheritance from FlaubertTokenizer
* remove XLMTokenizer inheritance from FlaubertTokenizer
* remove XLMTokenizer inheritance from FlaubertTokenizer
* remove XLMTokenizer inheritance from FlaubertTokenizer: fixed styling
* removed repo-consistensy issue
2022-10-05 09:19:04 -04:00
Shyam Sudhakaran
e12bbe3b4d
Remove bert interdependency from clip tokenizer ( #19332 )
2022-10-05 09:15:14 -04:00
Divyanshu Kumar
512fa41c53
Removed interdependency of BERT's Tokenizer in tokenization of prophetnet ( #19331 )
...
* removed interdependency of BERTTokenizer in tokenization of prophetnet
* fix: style
2022-10-05 09:12:47 -04:00
Alara Dirik
07e94bf159
Maskformer post-processing fixes and improvements ( #19172 )
...
- Improves MaskFormer docs, corrects minor typos
- Restructures MaskFormerFeatureExtractor.post_process_panoptic_segmentation for better readability, adds target_sizes argument for optional resizing
- Adds post_process_semantic_segmentation and post_process_instance_segmentation methods.
- Adds a deprecation warning to post_process_segmentation method in favour of post_process_instance_segmentation
2022-10-05 15:27:15 +03:00
Druhin Abrol
6268694e27
removing XLMConfig inheritance from FlaubertConfig ( #19326 )
...
* removing XLMConfig inheritance from FlaubertConfig
* removing XLMConfig inheritance from FlaubertConfig
* Fixed styling issue
* Update configuration_flaubert.py
Co-authored-by: Druhin Abrol <druhinabrol@192.168.1.6>
2022-10-04 19:39:47 -04:00
Erin
bf7eb0c9b3
Remove interdependency from OpenAI tokenizer ( #19327 )
...
* Remove interdependency from OpenAI tokenizer
* Adjust import order for linter
2022-10-04 17:51:55 -04:00
Samuel Arcadinho
971da2e6ec
Clamping hidden state values to allow FP16 ( #19229 )
...
* Clamping hidden state values to allow FP16
* Reformating
* Adding missing if condition
* Update src/transformers/models/longt5/modeling_longt5.py
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
* Update src/transformers/models/longt5/modeling_longt5.py
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
* Update src/transformers/models/longt5/modeling_longt5.py
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
* Formating file
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
2022-10-04 20:28:28 +02:00
Younes Belkada
587d84b178
Add BloomForQuestionAnswering
( #19310 )
...
* add bloom for question answering
- attempt to add Bloom for question answering
- adapted from `GPTJForQuestionAnswering`
- Fixed `num_labels` to `2` for common tests
- Added a bit of docstring
- All common tests pass
* Update src/transformers/models/bloom/modeling_bloom.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* revert changes related to `num_labels`
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-10-04 17:52:13 +02:00
Sushrut1101
6dce9e0cdd
docker-build: Update actions/checkout to v3 ( #19288 )
2022-10-04 16:26:52 +02:00
Arnaud Stiegler
6fd254a37d
Removing BertConfig inheritance from LayoutLMConfig ( #19307 )
...
* removing BertConfig inheritance
* fix missing arguments
2022-10-04 10:24:07 -04:00
Partho
a9782881a4
wrap forward passes with torch.no_grad() ( #19273 )
2022-10-04 16:13:22 +02:00
Partho
d6e920449e
wrap forward passes with torch.no_grad() ( #19274 )
2022-10-04 16:12:03 +02:00
Partho
2403dbd607
wrap forward passes with torch.no_grad() ( #19278 )
2022-10-04 16:09:23 +02:00
Partho
f134d38553
wrap forward passes with torch.no_grad() ( #19279 )
2022-10-04 16:08:29 +02:00
Oscar Dominguez
cd024da6f8
ci(workflows): update actions/checkout to v3 ( #19280 )
...
in stale.yml
2022-10-04 16:07:53 +02:00
Oscar Dominguez
ca3ebc44e0
ci(stale.yml): upgrade actions/setup-python to v4 ( #19281 )
2022-10-04 16:07:33 +02:00
gouqi_nju
cc263e9bb4
alter retrived to retrieved ( #18863 )
2022-10-04 16:00:47 +02:00
Debjit Bhowal
9b630168a9
Added type hints for TF: rag model ( #19284 )
...
* Added type hints for TF: rag model
* TFModelInputType added in place of TF.Tensor
* reformatting by black
2022-10-04 14:56:35 +01:00
IMvision12
ac5ea74ee8
Added Type hints for LED TF ( #19315 )
...
* Update modeling_tf_led.py
* Update modeling_tf_led.py
2022-10-04 14:55:15 +01:00
Matt
3a1a56a8fe
Fix for sequence regression fit() in TF ( #19316 )
...
Co-authored-by: Your Name <you@example.com>
2022-10-04 14:48:27 +01:00
Patrick von Platen
fe10796f4f
[Docs] Fix link ( #19313 )
2022-10-04 09:00:52 -04:00
Shubham S Jagtap
534cd8ff94
Update README.md ( #19309 )
2022-10-04 07:46:50 -04:00
dependabot[bot]
4c962d5e79
Bump joblib in /examples/research_projects/visual_bert ( #19269 )
...
Bumps [joblib](https://github.com/joblib/joblib ) from 0.16.0 to 1.2.0.
- [Release notes](https://github.com/joblib/joblib/releases )
- [Changelog](https://github.com/joblib/joblib/blob/master/CHANGES.rst )
- [Commits](https://github.com/joblib/joblib/compare/0.16.0...1.2.0 )
---
updated-dependencies:
- dependency-name: joblib
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2022-10-03 23:57:50 +02:00
dependabot[bot]
c7ec0afce0
Bump joblib in /examples/research_projects/decision_transformer ( #19270 )
...
Bumps [joblib](https://github.com/joblib/joblib ) from 1.1.0 to 1.2.0.
- [Release notes](https://github.com/joblib/joblib/releases )
- [Changelog](https://github.com/joblib/joblib/blob/master/CHANGES.rst )
- [Commits](https://github.com/joblib/joblib/compare/1.1.0...1.2.0 )
---
updated-dependencies:
- dependency-name: joblib
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2022-10-03 23:57:37 +02:00
dependabot[bot]
ca26277e33
Bump joblib from 0.16.0 to 1.2.0 in /examples/research_projects/lxmert ( #19268 )
...
Bumps [joblib](https://github.com/joblib/joblib ) from 0.16.0 to 1.2.0.
- [Release notes](https://github.com/joblib/joblib/releases )
- [Changelog](https://github.com/joblib/joblib/blob/master/CHANGES.rst )
- [Commits](https://github.com/joblib/joblib/compare/0.16.0...1.2.0 )
---
updated-dependencies:
- dependency-name: joblib
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2022-10-03 23:49:35 +02:00
Andrew Sansom
008531c18a
Update Protobuf dependency version to fix known vulnerability ( #19247 )
...
* Update protobuf dependency to fix vulnerability
* Update `dependency_versions_table.py` to include updated protobuf.
2022-10-03 23:37:09 +02:00
Steven Liu
68f50f3453
Breakup export guide ( #19271 )
...
* split onnx and torchscript docs
* make style
* apply reviews
2022-10-03 13:18:29 -07:00
Sylvain Gugger
18c06208c4
Don't automatically add bug label ( #19302 )
2022-10-03 12:42:04 -04:00
Divyanshu Kumar
c28d04e9e2
Update no_trainer script for summarization ( #19277 )
...
* Update no_trainer script for summarization
* removed unnecessary import
* fixes notation mistake
* removed: unused variable
2022-10-03 09:21:51 -04:00
Alara Dirik
36f52e9593
Restructure DETR post-processing, return prediction scores ( #19262 )
...
* Restructure DetrFeatureExtractor post-processing methods
* Update post_process_instance_segmentation and post_process_panoptic_segmentation methods to return prediction scores
* Update DETR models docs
2022-10-03 12:02:51 +03:00
Kashif Rasul
5cd16f01db
time series forecasting model ( #17965 )
...
* initial files
* initial model via cli
* typos
* make a start on the model config
* ready with configuation
* remove tokenizer ref.
* init the transformer
* added initial model forward to return dec_output
* require gluonts
* update dep. ver table and add as extra
* fixed typo
* add type for prediction_length
* use num_time_features
* use config
* more config
* typos
* opps another typo
* freq can be none
* default via transformation is 1
* initial transformations
* fix imports
* added transform_start_field
* add helper to create pytorch dataloader
* added inital val and test data loader
* added initial distr head and loss
* training working
* remove TimeSeriesTransformerTokenizer
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/__init__.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/__init__.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fixed copyright
* removed docs
* remove time series tokenizer
* fixed docs
* fix text
* fix second
* fix default
* fix order
* use config directly
* undo change
* fix comment
* fix year
* fix import
* add additional arguments for training vs. test
* initial greedy inference loop
* fix inference
* comment out token inputs to enc dec
* Use HF encoder/decoder
* fix inference
* Use Seq2SeqTSModelOutput output
* return Seq2SeqTSPredictionOutput
* added default arguments
* fix return_dict true
* scale is a tensor
* output static_features for inference
* clean up some unused bits
* fixed typo
* set return_dict if none
* call model once for both train/predict
* use cache if future_target is none
* initial generate func
* generate arguments
* future_time_feat is required
* return SampleTSPredictionOutput
* removed unneeded classes
* fix when params is none
* fix return dict
* fix num_attention_heads
* fix arguments
* remove unused shift_tokens_right
* add different dropout configs
* implement FeatureEmbedder, Scaler and weighted_average
* remove gluonts dependency
* fix class names
* avoid _variable names
* remove gluonts dependency
* fix imports
* remove gluonts from configuration
* fix docs
* fixed typo
* move utils to examples
* add example requirements
* config has no freq
* initial run_ts_no_trainer
* remove from ignore
* fix output_attentions and removed unsued getters/setters
* removed unsed tests
* add dec seq len
* add test_attention_outputs
* set has_text_modality=False
* add config attribute_map
* make style
* make fix-copies
* add encoder_outputs to TimeSeriesTransformerForPrediction forward
* Improve docs, add model to README
* added test_forward_signature
* More improvements
* Add more copied from
* Fix README
* Fix remaining quality issues
* updated encoder and decoder
* fix generate
* output_hidden_states and use_cache are optional
* past key_values returned too
* initialize weights of distribution_output module
* fixed more tests
* update test_forward_signature
* fix return_dict outputs
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* removed commented out tests
* added neg. bin and normal output
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* move to one line
* Add docstrings
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* add try except for assert and raise
* try and raise exception
* fix the documentation formatting
* fix assert call
* fix docstring formatting
* removed input_ids from DOCSTRING
* Update input docstring
* Improve variable names
* Update order of inputs
* Improve configuration
* Improve variable names
* Improve docs
* Remove key_length from tests
* Add extra docs
* initial unittests
* added test_inference_no_head test
* added test_inference_head
* add test_seq_to_seq_generation
* make style
* one line
* assert mean prediction
* removed comments
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fix order of args
* make past_observed_mask optional as well
* added Amazon license header
* updated utils with new fieldnames
* make style
* cleanup
* undo position of past_observed_mask
* fix import
* typo
* more typo
* rename example files
* remove example for now
* Update docs/source/en/_toctree.yml
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update modeling_time_series_transformer.py
fix style
* fixed typo
* fix typo and grammer
* fix style
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: NielsRogge <niels.rogge1@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-09-30 15:32:59 -04:00
Joao Gante
cfb777f27c
Docs - Guide to add a new TensorFlow model ( #19256 )
...
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
Co-authored-by: Matt <Rocketknight1@users.noreply.github.com>
2022-09-30 20:30:38 +01:00
Keith Kjer
6a08162ad4
Fix cached lookup filepath on windows for hub ( #19178 )
...
* Update hub.py commit_hash extraction
Add safety mechanism for windows systems to unify logic (replace double backslashes with /)
* Fix string quotetype
* Aaaa circleci is messing with me.
* Switch to using as_posix() method from pathlib
* Update src/transformers/utils/hub.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/utils/hub.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-09-30 15:13:39 -04:00
Yih-Dar
f33858d18a
Fix Encoder-Decoder testing issue about repo. names ( #19250 )
...
* Change "../gpt2" to "gpt2"
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2022-09-30 18:15:07 +02:00
Yih-Dar
2fba98e585
Add beautifulsoup4
to the dependency list ( #19253 )
...
* Add `beautifulsoup4` to extras["testing"]
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2022-09-30 18:14:01 +02:00
Sylvain Gugger
3e2dd7f92d
Poc to use safetensors ( #19175 )
...
* Poc to use safetensors
* Typo
* Final version
* Add tests
* Save with the right name!
* Update tests/test_modeling_common.py
Co-authored-by: Julien Chaumond <julien@huggingface.co>
* Support for sharded checkpoints
* Test from Hub part 1
* Test from hub part 2
* Fix regular checkpoint sharding
* Bump for fixes
Co-authored-by: Julien Chaumond <julien@huggingface.co>
2022-09-30 10:58:04 -04:00
Jingya HUANG
dad578e4c3
Add notebooks ( #19259 )
2022-09-30 10:04:36 -04:00