mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-04 21:30:07 +06:00
![]() * add quark quantizer * add quark doc * clean up doc * fix tests * make style * more style fixes * cleanup imports * cleaning * precise install * Update docs/source/en/quantization/quark.md Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com> * Update tests/quantization/quark_integration/test_quark.py Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com> * Update src/transformers/utils/quantization_config.py Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com> * remove import guard as suggested * update copyright headers * add quark to transformers-quantization-latest-gpu Dockerfile * make tests pass on transformers main + quark==0.7 * add missing F8_E4M3 and F8_E5M2 keys from str_to_torch_dtype --------- Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com> Co-authored-by: Bowen Bao <bowenbao@amd.com> Co-authored-by: Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com> |
||
---|---|---|
.. | ||
aqlm.md | ||
awq.md | ||
bitnet.md | ||
bitsandbytes.md | ||
compressed_tensors.md | ||
contribute.md | ||
eetq.md | ||
fbgemm_fp8.md | ||
finegrained_fp8.md | ||
gptq.md | ||
higgs.md | ||
hqq.md | ||
optimum.md | ||
overview.md | ||
quanto.md | ||
quark.md | ||
spqr.md | ||
torchao.md | ||
vptq.md |