mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
![]() * Add autoquant support for torchao quantizer Summary: att, also verified that autoquantized model can be saved and loaded: save: https://gist.github.com/jerryzh168/01d367aaf44dbbbfd4068a4a10a00061 load: https://gist.github.com/jerryzh168/d5c6c401b2abdf18e0b6771341f1525c Test Plan: tested locally with above script model uploaded to https://huggingface.co/jerryzh168/llama3-8b-autoquant Reviewers: Subscribers: Tasks: Tags: * add test * ruff fix * ruff reformat * add docs and min_sqnr support * format * format * fix test * update doc * format * remove disable_compile * format |
||
---|---|---|
.. | ||
aqlm_integration | ||
autoawq | ||
bitnet_integration | ||
bnb | ||
compressed_tensors | ||
eetq_integration | ||
fbgemm_fp8 | ||
finegrained_fp8 | ||
ggml | ||
gptq | ||
higgs | ||
hqq | ||
quanto_integration | ||
spqr_integration | ||
torchao_integration | ||
vptq_integration |