Kirill
|
a60010566a
|
llama docs: fix conversion script url (#22514)
|
2023-04-03 10:28:40 -04:00 |
|
Sylvain Gugger
|
786092a35e
|
Rework a bit the LLaMA conversion script (#22236)
* Update LLaMA conversion script
* Doc
* Fix the weight size for the 13B checkpoint
* Update src/transformers/models/llama/convert_llama_weights_to_hf.py
Co-authored-by: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
---------
Co-authored-by: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
|
2023-03-20 11:30:36 -04:00 |
|
lewtun
|
f251441387
|
Add LlamaForSequenceClassification (#22209)
* Add LlamaForSequenceClassification
* Update src/transformers/models/llama/modeling_llama.py
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
* Update src/transformers/models/llama/modeling_llama.py
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
* Add docstring
* Add test
* Add input embedding getter and setter
* Remove dead code
---------
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
|
2023-03-17 14:39:26 +01:00 |
|
Sylvain Gugger
|
00934026a4
|
LLaMA house-keeping (#22216)
* LLaMA house-keeping
* Doc links
|
2023-03-17 08:55:15 -04:00 |
|
Kevin Turner
|
33d033d694
|
fix typos in llama.mdx (#22223)
|
2023-03-17 08:43:18 +00:00 |
|
Jason Phang
|
0041be5b3d
|
LLaMA Implementation (#21955)
* LLaMA
* sharding and docs
* tweak
* black
* inits
* ruff
* LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP
* init
* no checkpoint
* docs
* ruff
* type_vocab_size
* tokenizer fixes
* tokenizer fixes
* Update tokenization_llama.py
* Update tokenization_llama.py
* Update configuration_llama.py
* Update modeling_llama.py
* tokenizer add_bos by default
* licenses
* remove decoder
* norms and mlp
* rope overhaul
* tweaks
* black
* mention OPT implementation
* off-by-one naming
* typo
* fix
* tokenization fix and slicing bug
* padding config
* cleanup
* black
* update tests
* undo typo
* fix vocab caching logic
* ruff
* docbuilder
* attn fix from BlackSamorez
* initial feedback
* typo
* docs
* llama case
* llama case
* load checkpoint docs
* comment about tokenizer
* tokenizer defaults
* clear past_key_values if use_cache=False
* last tweaks
* last tweaks
* last tweaks
* last tweaks
---------
Co-authored-by: Stella Biderman <stellabiderman@gmail.com>
|
2023-03-16 09:00:53 -04:00 |
|