mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-14 10:08:29 +06:00
![]() * Add autoquant support for torchao quantizer Summary: att, also verified that autoquantized model can be saved and loaded: save: https://gist.github.com/jerryzh168/01d367aaf44dbbbfd4068a4a10a00061 load: https://gist.github.com/jerryzh168/d5c6c401b2abdf18e0b6771341f1525c Test Plan: tested locally with above script model uploaded to https://huggingface.co/jerryzh168/llama3-8b-autoquant Reviewers: Subscribers: Tasks: Tags: * add test * ruff fix * ruff reformat * add docs and min_sqnr support * format * format * fix test * update doc * format * remove disable_compile * format |
||
---|---|---|
.. | ||
aqlm.md | ||
awq.md | ||
bitnet.md | ||
bitsandbytes.md | ||
compressed_tensors.md | ||
contribute.md | ||
eetq.md | ||
fbgemm_fp8.md | ||
finegrained_fp8.md | ||
gptq.md | ||
higgs.md | ||
hqq.md | ||
optimum.md | ||
overview.md | ||
quanto.md | ||
spqr.md | ||
torchao.md | ||
vptq.md |