transformers/tests/quantization/aqlm_integration
Joao Gante cf32ee1753
Cache: use batch_size instead of max_batch_size (#32657)
* more precise name

* better docstrings

* Update src/transformers/cache_utils.py

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

---------

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2024-08-16 11:48:45 +01:00
..
__init__.py
test_aqlm.py Cache: use batch_size instead of max_batch_size (#32657) 2024-08-16 11:48:45 +01:00