transformers/docker
Ita Zaporozhets e48e5f1f13
Support reading tiktoken tokenizer.model file (#31656)
* use existing TikTokenConverter to read tiktoken tokenizer.model file

* del test file

* create titktoken integration file

* adding tiktoken llama test

* ALTNATIVE IMPLEMENTATION: supports llama 405B

* fix one char

* remove redundant line

* small fix

* rm unused import

* flag for converting from tiktokeng

* remove unneeded file

* ruff

* remove llamatiktokenconverter, stick to general converter

* tiktoken support v2

* update test

* remove stale changes

* udpate doc

* protect import

* use is_protobuf_available

* add templateprocessor in tiktokenconverter

* reverting templateprocessor from tiktoken support

* update test

* add require_tiktoken

* dev-ci

* trigger build

* trigger build again

* dev-ci

* [build-ci-image] tiktoken

* dev-ci

* dev-ci

* dev-ci

* dev-ci

* change tiktoken file name

* feedback review

* feedback rev

* applying feedback, removing tiktoken converters

* conform test

* adding docs for review

* add doc file for review

* add doc file for review

* add doc file for review

* support loading model without config.json file

* Revert "support loading model without config.json file"

This reverts commit 2753602e51c34cef2f184eb11f36d2ad1b02babb.

* remove dev var

* updating docs

* safely import protobuf

* fix protobuf import error

* fix protobuf import error

* trying isort to fix ruff error

* fix ruff error

* try to fix ruff again

* try to fix ruff again

* try to fix ruff again

* doc table of contents

* add fix for consistency.dockerfile torchaudio

* ruff

* applying feedback

* minor typo

* merging with push-ci-image

* clean up imports

* revert dockerfile consistency
2024-09-06 14:24:02 +02:00
..
transformers-all-latest-gpu use torch 2.4 in 2 CI jobs (#32302) 2024-07-29 22:12:21 +02:00
transformers-doc-builder Use python 3.10 for docbuild (#28399) 2024-01-11 14:39:49 +01:00
transformers-gpu TF: TF 2.10 unpin + related onnx test skips (#18995) 2022-09-12 19:30:27 +01:00
transformers-past-gpu Byebye pytorch 1.9 (#24080) 2023-06-16 16:38:23 +02:00
transformers-pytorch-amd-gpu CI: update to ROCm 6.0.2 and test MI300 (#30266) 2024-05-13 18:14:36 +02:00
transformers-pytorch-deepspeed-amd-gpu fix: Fixed pydantic required version in dockerfiles to make it compatible with DeepSpeed (#33105) 2024-08-26 17:10:36 +02:00
transformers-pytorch-deepspeed-latest-gpu fix: Fixed pydantic required version in dockerfiles to make it compatible with DeepSpeed (#33105) 2024-08-26 17:10:36 +02:00
transformers-pytorch-deepspeed-nightly-gpu Update CUDA versions for DeepSpeed (#27853) 2023-12-05 16:15:21 -05:00
transformers-pytorch-gpu use torch 2.4 in 2 CI jobs (#32302) 2024-07-29 22:12:21 +02:00
transformers-pytorch-tpu Rename master to main for notebooks links and leftovers (#16397) 2022-03-25 09:12:23 -04:00
transformers-quantization-latest-gpu Quantization / TST: Fix remaining quantization tests (#31000) 2024-05-24 14:35:59 +02:00
transformers-tensorflow-gpu Use torch 2.3 for CI (#30837) 2024-05-15 19:31:52 +02:00
consistency.dockerfile Support reading tiktoken tokenizer.model file (#31656) 2024-09-06 14:24:02 +02:00
custom-tokenizers.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
examples-tf.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
examples-torch.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
exotic-models.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
jax-light.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
pipeline-tf.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
pipeline-torch.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
quality.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
tf-light.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
torch-jax-light.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00
torch-light.dockerfile Support reading tiktoken tokenizer.model file (#31656) 2024-09-06 14:24:02 +02:00
torch-tf-light.dockerfile unpin uv (#31055) 2024-05-27 13:47:47 +02:00