mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-02 19:21:31 +06:00
[Blip2
] Add int8 support for blip2-flan-t5-xxl
(#21574)
add int8 support
This commit is contained in:
parent
b47a16743b
commit
75a208ef66
@ -285,6 +285,7 @@ class Blip2PreTrainedModel(PreTrainedModel):
|
||||
r"language_model.decoder.embed_tokens.weight",
|
||||
]
|
||||
_no_split_modules = ["Blip2Attention", "T5Block", "OPTDecoderLayer"]
|
||||
_keep_in_fp32_modules = ["wo"]
|
||||
|
||||
def _init_weights(self, module):
|
||||
"""Initialize the weights"""
|
||||
|
Loading…
Reference in New Issue
Block a user