mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-19 12:38:23 +06:00
![]() * Initial commit * update modeling code * update doc * add functions necessary * fix impotrs * revert changes * fixup * more styling to get going * remove standalone encoder * update code * styling * fix config and model * update code and some refactoring * make more tests pass * Adding NLLB-200 - MoE - 54.5B for no language left behind Fixes #21300 * fix mor common tests * styke * update testing file * update * update * Router2 doc * update check config with sparse layer * add dummy router * update current conversion script * create on the fly conversion script * Fixup * style * style 2 * fix empty return * fix return * Update default config sparse layers * easier to create sparse layers * update * update conversion script * update modeling * add to toctree * styling * make ruff happy * update docstring * update conversion script * update, will break tests but impelemting top2 * update * ❗local groups are supported here * ⚠️ Support for local groups is now removed ⚠️ This is because it has to work with model parallelism that we do not support * finish simplificaiton * Fix forward * style * fixup * Update modelling and test, refactoring * update tests * remove final layer)norm as it is done in the FF * routing works! Logits test added * nit in test * remove top1router * style * make sure sparse are tested. Had to change route_tokens a liottle bit * add support for unslip models when converting * fixup * style * update test s * update test * REFACTOR * encoder outputs match! * style * update testing * 🎉encoder and decoder logits match 🎉 * styleing * update tests * cleanup tests * fix router test and CIs * cleanup * cleanup test styling * fix tests * Finally the generation tests match! * cleanup * update test * style testing file * remove script * cleanup * more cleanup * nits * update * NLLB tokenizer is wrong and will be fixed soon * use LongTensors * update tests * revert some small changes * fix second expert sampling and batch prioritized routing * update tests * finish last tests * make ruff happy * update * ruff again * style * Update docs/source/en/model_doc/nllb-moe.mdx Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Updates based on review * style and fix import issue * nit * more nits * cleanup * styling * update test_seconde_expert_policy * fix name * last nit on the markdown examples --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> |
||
---|---|---|
.. | ||
internal | ||
main_classes | ||
model_doc | ||
tasks | ||
_config.py | ||
_toctree.yml | ||
accelerate.mdx | ||
add_new_model.mdx | ||
add_new_pipeline.mdx | ||
add_tensorflow_model.mdx | ||
attention.mdx | ||
autoclass_tutorial.mdx | ||
benchmarks.mdx | ||
bertology.mdx | ||
big_models.mdx | ||
community.mdx | ||
contributing.md | ||
converting_tensorflow_models.mdx | ||
create_a_model.mdx | ||
custom_models.mdx | ||
debugging.mdx | ||
fast_tokenizers.mdx | ||
generation_strategies.mdx | ||
glossary.mdx | ||
hpo_train.mdx | ||
index.mdx | ||
installation.mdx | ||
migration.mdx | ||
model_sharing.mdx | ||
model_summary.mdx | ||
multilingual.mdx | ||
notebooks.md | ||
pad_truncation.mdx | ||
perf_hardware.mdx | ||
perf_infer_cpu.mdx | ||
perf_infer_gpu_many.mdx | ||
perf_infer_gpu_one.mdx | ||
perf_infer_special.mdx | ||
perf_train_cpu_many.mdx | ||
perf_train_cpu.mdx | ||
perf_train_gpu_many.mdx | ||
perf_train_gpu_one.mdx | ||
perf_train_special.mdx | ||
perf_train_tpu_tf.mdx | ||
perf_train_tpu.mdx | ||
performance.mdx | ||
perplexity.mdx | ||
philosophy.mdx | ||
pipeline_tutorial.mdx | ||
pipeline_webserver.mdx | ||
pr_checks.mdx | ||
preprocessing.mdx | ||
quicktour.mdx | ||
run_scripts.mdx | ||
sagemaker.mdx | ||
serialization.mdx | ||
task_summary.mdx | ||
tasks_explained.mdx | ||
testing.mdx | ||
tf_xla.mdx | ||
tokenizer_summary.mdx | ||
torchscript.mdx | ||
training.mdx | ||
troubleshooting.mdx |