transformers/docs/source
Cyril Vallez 6604764007
add Glm (#33823)
* Create modular_glm.py

* Update modular_glm.py

* Finalize architecture without all attentions

* Add all attentions modules

* Finalize modular

* Update given last version

* Last update

* Finalize model

* Finalize converter

* Update convert_glm_weights_to_hf.py

* style

* style

* Create __init__.py

* Aff all inits

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Correct the rotary embeddings

* Remove apply_residual_connection_post_layernorm (always false)

* remove use_rms_norm (always true)

* remove past_layer_norm (always true)

* Update __init__.py

* Update config and license

* start adding tests and doc

* Add doc + style

* Update test_modeling_glm.py

* Add dummies

* Apply correct modeling

* Refactor attention to follow llama

* Update __init__.py

* Update convert_glm_weights_to_hf.py

* Correct bias

* remove linear_bias and pdrop (never used)

* apply modular

* Simplify converter

* remove dummies + style

* add model_input_names

* Add pretraining_tp to config for when eager attention is used

* Update modular to remove all pretraining_tp

* Update test_modeling_glm.py

* Update the __all__

* Update __all__

* Update __init__.py

* Update test_modeling_glm.py

* add revisions

* Add the correct repos and revisions

* style

* Update __init__.py

* update exports

* remove import of modular files

* style

* Apply Llama changes + refine converter

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* Update convert_glm_weights_to_hf.py

* style

* Use new modular converter

* add pretrainedmodel to init

* style

* Update test_modeling_glm.py

* Move config outside modular to please CI about docstrings

* Add dummies to please CI

* Update glm.md

* Update glm.md
2024-10-18 17:41:12 +02:00
..
ar Add Translate docs into Arabic - section files CONCEPTUAL GUIDES (#33982) 2024-10-09 14:51:19 -07:00
de Skip tests properly (#31308) 2024-06-26 21:59:08 +01:00
en add Glm (#33823) 2024-10-18 17:41:12 +02:00
es Trainer - deprecate tokenizer for processing_class (#32385) 2024-10-02 14:08:46 +01:00
fr Trainer - deprecate tokenizer for processing_class (#32385) 2024-10-02 14:08:46 +01:00
hi More fixes for doctest (#30265) 2024-04-16 11:58:55 +02:00
it Updated the custom_models.md changed cross_entropy code (#33118) 2024-08-26 13:15:43 +02:00
ja Add auto model for image-text-to-text (#32472) 2024-10-08 14:26:43 +02:00
ko 🌐 [i18n-KO] Translated blip-2.md to Korean (#33516) 2024-10-15 11:21:22 -07:00
ms Remove old TF port docs (#30426) 2024-04-23 16:06:20 +01:00
pt Trainer - deprecate tokenizer for processing_class (#32385) 2024-10-02 14:08:46 +01:00
te Trainer - deprecate tokenizer for processing_class (#32385) 2024-10-02 14:08:46 +01:00
tr Translate index.md to Turkish (#27093) 2023-11-08 08:35:20 -05:00
zh Trainer - deprecate tokenizer for processing_class (#32385) 2024-10-02 14:08:46 +01:00
_config.py [#29174] ImportError Fix: Trainer with PyTorch requires accelerate>=0.20.1 Fix (#29888) 2024-04-08 14:21:16 +01:00