mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-04 21:30:07 +06:00
![]() * Enable non-safetensor serialization and deserialization for TorchAoConfig quantized model Summary: After https://github.com/huggingface/huggingface_hub/pull/2440 we added non-safetensor serialization and deserialization in huggingface, with this we can now add the support in transformers Note that we don't plan to add safetensor serialization due to different goals of wrapper tensor subclass and safetensor see README for more details Test Plan: tested locally Reviewers: Subscribers: Tasks: Tags: * formatting * formatting * minor fix * formatting * address comments * comments * minor fix * update doc * refactor compressed tensor quantizer |
||
---|---|---|
.. | ||
aqlm.md | ||
awq.md | ||
bitsandbytes.md | ||
compressed_tensors.md | ||
contribute.md | ||
eetq.md | ||
fbgemm_fp8.md | ||
gptq.md | ||
hqq.md | ||
optimum.md | ||
overview.md | ||
quanto.md | ||
torchao.md |