transformers/docs/source/main_classes
Cheng Li c83fbc5f2d
[Deepspeed] Allow HF optimizer and scheduler to be passed to deepspeed (#10464)
* pass hf optimizer and scheduler to deepspeed if not specified in ds config

* pass hf optimizer and scheduler to deepspeed if not specified in ds config

* update

* make init_deepspeed support config dict

* fix docstring formatting

* clean up trainer's comments

* add new tests

* fix type

* composit argparse doesn't work

* style

* add a new test, rename others

* document new functionality

* complete tests, add docs

* style

* correct level

* Apply suggestions from code review

Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>

* add new methods to the doc

* must tell DS we are using a non-native optimizer

* add protection against cpu_offload + HF optimizer combo

* fix the cli overrides

* sync docs + tests

* restore AdamW

* better docs

* need new version

* no longer needed

* remove outdate information

* refactor duplicated code

Co-authored-by: Stas Bekman <stas@stason.org>
Co-authored-by: Stas Bekman <stas00@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2021-03-16 15:51:09 -07:00
..
callback.rst Copyright (#8970) 2020-12-07 18:36:34 -05:00
configuration.rst Copyright (#8970) 2020-12-07 18:36:34 -05:00
feature_extractor.rst [FeatureExtractorSavingUtils] Refactor PretrainedFeatureExtractor (#10594) 2021-03-09 12:16:59 +03:00
logging.rst Logging propagation (#10092) 2021-02-09 10:27:49 -05:00
model.rst [Flax] Align FlaxBertForMaskedLM with BertForMaskedLM, implement from_pretrained, init (#9054) 2020-12-16 13:03:32 +01:00
optimizer_schedules.rst Seq2seq trainer (#9241) 2020-12-22 11:33:44 -05:00
output.rst Remove unsupported methods from ModelOutput doc (#10505) 2021-03-03 14:55:18 -05:00
pipelines.rst TableQuestionAnsweringPipeline (#9145) 2020-12-16 12:31:50 -05:00
processors.rst Fix documentation links always pointing to master. (#9217) 2021-01-05 06:18:48 -05:00
tokenizer.rst A few fixes in the documentation (#10033) 2021-02-08 05:02:01 -05:00
trainer.rst [Deepspeed] Allow HF optimizer and scheduler to be passed to deepspeed (#10464) 2021-03-16 15:51:09 -07:00