mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
Enable device map (#30870)
* added_no_split_modules * added LlavaNextVisionAttention to _no_split_modules
This commit is contained in:
parent
57c965a8f1
commit
3802e786ef
@ -124,6 +124,7 @@ class VideoLlavaPreTrainedModel(PreTrainedModel):
|
||||
supports_gradient_checkpointing = True
|
||||
_skip_keys_device_placement = "past_key_values"
|
||||
_supports_flash_attn_2 = True
|
||||
_no_split_modules = ["VideoLlavaVisionAttention"]
|
||||
|
||||
def _init_weights(self, module):
|
||||
# important: this ported version of VideoLlava isn't meant for training from scratch - only
|
||||
|
Loading…
Reference in New Issue
Block a user