mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-04 13:20:12 +06:00
parent
74f5e4a1fa
commit
07aab1af1e
@ -14,12 +14,6 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
"""ALIGN model configuration"""
|
"""ALIGN model configuration"""
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
pass
|
|
||||||
|
|
||||||
from ...configuration_utils import PretrainedConfig
|
from ...configuration_utils import PretrainedConfig
|
||||||
from ...utils import logging
|
from ...utils import logging
|
||||||
|
|
||||||
|
@ -52,13 +52,10 @@ from ...utils import (
|
|||||||
can_return_tuple,
|
can_return_tuple,
|
||||||
logging,
|
logging,
|
||||||
)
|
)
|
||||||
from ...utils.import_utils import is_causal_conv1d_available, is_flash_attn_2_available, is_mamba_2_ssm_available
|
from ...utils.import_utils import is_causal_conv1d_available, is_mamba_2_ssm_available
|
||||||
from .configuration_bamba import BambaConfig
|
from .configuration_bamba import BambaConfig
|
||||||
|
|
||||||
|
|
||||||
if is_flash_attn_2_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
if is_mamba_2_ssm_available():
|
if is_mamba_2_ssm_available():
|
||||||
from mamba_ssm.ops.triton.selective_state_update import selective_state_update
|
from mamba_ssm.ops.triton.selective_state_update import selective_state_update
|
||||||
from mamba_ssm.ops.triton.ssd_combined import mamba_chunk_scan_combined, mamba_split_conv1d_scan_combined
|
from mamba_ssm.ops.triton.ssd_combined import mamba_chunk_scan_combined, mamba_split_conv1d_scan_combined
|
||||||
|
@ -15,11 +15,7 @@
|
|||||||
"""CLVP model configuration"""
|
"""CLVP model configuration"""
|
||||||
|
|
||||||
import os
|
import os
|
||||||
from typing import TYPE_CHECKING, Union
|
from typing import Union
|
||||||
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
pass
|
|
||||||
|
|
||||||
from ...configuration_utils import PretrainedConfig
|
from ...configuration_utils import PretrainedConfig
|
||||||
from ...utils import logging
|
from ...utils import logging
|
||||||
|
@ -33,13 +33,10 @@ from transformers.models.sam.modeling_sam import SamMLPBlock, SamVisionAttention
|
|||||||
from ...configuration_utils import PretrainedConfig
|
from ...configuration_utils import PretrainedConfig
|
||||||
from ...modeling_flash_attention_utils import FlashAttentionKwargs
|
from ...modeling_flash_attention_utils import FlashAttentionKwargs
|
||||||
from ...processing_utils import Unpack
|
from ...processing_utils import Unpack
|
||||||
from ...utils import auto_docstring, can_return_tuple, is_vision_available, logging
|
from ...utils import auto_docstring, can_return_tuple, logging
|
||||||
from ..auto import CONFIG_MAPPING, AutoConfig
|
from ..auto import CONFIG_MAPPING, AutoConfig
|
||||||
|
|
||||||
|
|
||||||
if is_vision_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
logger = logging.get_logger(__name__)
|
logger = logging.get_logger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
@ -14,11 +14,7 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
"""Mpt configuration"""
|
"""Mpt configuration"""
|
||||||
|
|
||||||
from typing import TYPE_CHECKING, Optional, Union
|
from typing import Optional, Union
|
||||||
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
pass
|
|
||||||
|
|
||||||
from ...configuration_utils import PretrainedConfig
|
from ...configuration_utils import PretrainedConfig
|
||||||
from ...utils import logging
|
from ...utils import logging
|
||||||
|
@ -41,16 +41,12 @@ from ...image_utils import (
|
|||||||
validate_preprocess_arguments,
|
validate_preprocess_arguments,
|
||||||
)
|
)
|
||||||
from ...utils import TensorType, filter_out_non_signature_kwargs, logging
|
from ...utils import TensorType, filter_out_non_signature_kwargs, logging
|
||||||
from ...utils.import_utils import is_cv2_available, is_vision_available
|
from ...utils.import_utils import is_vision_available
|
||||||
|
|
||||||
|
|
||||||
logger = logging.get_logger(__name__)
|
logger = logging.get_logger(__name__)
|
||||||
|
|
||||||
|
|
||||||
if is_cv2_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
if is_vision_available():
|
if is_vision_available():
|
||||||
import PIL
|
import PIL
|
||||||
|
|
||||||
|
@ -25,7 +25,7 @@ from ...activations import ACT2FN
|
|||||||
from ...cache_utils import Cache, DynamicCache
|
from ...cache_utils import Cache, DynamicCache
|
||||||
from ...generation import GenerationMixin
|
from ...generation import GenerationMixin
|
||||||
from ...modeling_attn_mask_utils import AttentionMaskConverter
|
from ...modeling_attn_mask_utils import AttentionMaskConverter
|
||||||
from ...modeling_flash_attention_utils import FlashAttentionKwargs, is_flash_attn_available
|
from ...modeling_flash_attention_utils import FlashAttentionKwargs
|
||||||
from ...modeling_outputs import (
|
from ...modeling_outputs import (
|
||||||
BaseModelOutputWithPast,
|
BaseModelOutputWithPast,
|
||||||
CausalLMOutputWithPast,
|
CausalLMOutputWithPast,
|
||||||
@ -44,10 +44,6 @@ if is_torch_flex_attn_available():
|
|||||||
from ...integrations.flex_attention import make_flex_block_causal_mask
|
from ...integrations.flex_attention import make_flex_block_causal_mask
|
||||||
|
|
||||||
|
|
||||||
if is_flash_attn_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
logger = logging.get_logger(__name__)
|
logger = logging.get_logger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
@ -14,12 +14,6 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
"""OWLv2 model configuration"""
|
"""OWLv2 model configuration"""
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
pass
|
|
||||||
|
|
||||||
from ...configuration_utils import PretrainedConfig
|
from ...configuration_utils import PretrainedConfig
|
||||||
from ...utils import logging
|
from ...utils import logging
|
||||||
|
|
||||||
|
@ -32,15 +32,12 @@ from ...image_utils import (
|
|||||||
to_numpy_array,
|
to_numpy_array,
|
||||||
valid_images,
|
valid_images,
|
||||||
)
|
)
|
||||||
from ...utils import TensorType, is_torch_available, is_vision_available, logging, requires_backends
|
from ...utils import TensorType, is_torch_available, logging, requires_backends
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available():
|
if is_torch_available():
|
||||||
import torch
|
import torch
|
||||||
|
|
||||||
if is_vision_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
logger = logging.get_logger(__name__)
|
logger = logging.get_logger(__name__)
|
||||||
|
|
||||||
|
@ -22,7 +22,6 @@ from transformers import (
|
|||||||
AutoProcessor,
|
AutoProcessor,
|
||||||
AyaVisionConfig,
|
AyaVisionConfig,
|
||||||
is_torch_available,
|
is_torch_available,
|
||||||
is_vision_available,
|
|
||||||
)
|
)
|
||||||
from transformers.testing_utils import (
|
from transformers.testing_utils import (
|
||||||
Expectations,
|
Expectations,
|
||||||
@ -51,10 +50,6 @@ if is_torch_available():
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
if is_vision_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
class AyaVisionVisionText2TextModelTester:
|
class AyaVisionVisionText2TextModelTester:
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
|
@ -17,14 +17,11 @@ import unittest
|
|||||||
|
|
||||||
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
||||||
from transformers.testing_utils import require_torch, require_vision
|
from transformers.testing_utils import require_torch, require_vision
|
||||||
from transformers.utils import is_torch_available, is_torchvision_available, is_vision_available
|
from transformers.utils import is_torchvision_available, is_vision_available
|
||||||
|
|
||||||
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
if is_vision_available():
|
if is_vision_available():
|
||||||
if is_torchvision_available():
|
if is_torchvision_available():
|
||||||
from transformers import InternVLVideoProcessor
|
from transformers import InternVLVideoProcessor
|
||||||
|
@ -20,15 +20,10 @@ import unittest
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
from transformers import AutoProcessor, AutoTokenizer, JanusProcessor
|
from transformers import AutoProcessor, AutoTokenizer, JanusProcessor
|
||||||
from transformers.utils import is_vision_available
|
|
||||||
|
|
||||||
from ...test_processing_common import ProcessorTesterMixin
|
from ...test_processing_common import ProcessorTesterMixin
|
||||||
|
|
||||||
|
|
||||||
if is_vision_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
class JanusProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
class JanusProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
||||||
processor_class = JanusProcessor
|
processor_class = JanusProcessor
|
||||||
|
|
||||||
|
@ -16,14 +16,11 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from transformers.testing_utils import require_torch, require_vision
|
from transformers.testing_utils import require_torch, require_vision
|
||||||
from transformers.utils import is_torch_available, is_torchvision_available, is_vision_available
|
from transformers.utils import is_torchvision_available, is_vision_available
|
||||||
|
|
||||||
from ...test_image_processing_common import ImageProcessingTestMixin, prepare_image_inputs
|
from ...test_image_processing_common import ImageProcessingTestMixin, prepare_image_inputs
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
if is_vision_available() and is_torchvision_available():
|
if is_vision_available() and is_torchvision_available():
|
||||||
from transformers import Llama4ImageProcessorFast
|
from transformers import Llama4ImageProcessorFast
|
||||||
|
|
||||||
|
@ -18,7 +18,7 @@ import unittest
|
|||||||
|
|
||||||
from transformers import AutoProcessor, AutoTokenizer, LlamaTokenizerFast, LlavaProcessor
|
from transformers import AutoProcessor, AutoTokenizer, LlamaTokenizerFast, LlavaProcessor
|
||||||
from transformers.testing_utils import require_vision
|
from transformers.testing_utils import require_vision
|
||||||
from transformers.utils import is_torch_available, is_vision_available
|
from transformers.utils import is_vision_available
|
||||||
|
|
||||||
from ...test_processing_common import ProcessorTesterMixin
|
from ...test_processing_common import ProcessorTesterMixin
|
||||||
|
|
||||||
@ -26,9 +26,6 @@ from ...test_processing_common import ProcessorTesterMixin
|
|||||||
if is_vision_available():
|
if is_vision_available():
|
||||||
from transformers import CLIPImageProcessor
|
from transformers import CLIPImageProcessor
|
||||||
|
|
||||||
if is_torch_available:
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
@require_vision
|
@require_vision
|
||||||
class LlavaProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
class LlavaProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
||||||
|
@ -21,7 +21,7 @@ import torch
|
|||||||
|
|
||||||
from transformers import AutoProcessor, LlamaTokenizerFast, LlavaNextVideoProcessor
|
from transformers import AutoProcessor, LlamaTokenizerFast, LlavaNextVideoProcessor
|
||||||
from transformers.testing_utils import require_vision
|
from transformers.testing_utils import require_vision
|
||||||
from transformers.utils import is_torch_available, is_torchvision_available, is_vision_available
|
from transformers.utils import is_torchvision_available, is_vision_available
|
||||||
|
|
||||||
from ...test_processing_common import ProcessorTesterMixin
|
from ...test_processing_common import ProcessorTesterMixin
|
||||||
|
|
||||||
@ -32,9 +32,6 @@ if is_vision_available():
|
|||||||
if is_torchvision_available():
|
if is_torchvision_available():
|
||||||
from transformers import LlavaNextVideoVideoProcessor
|
from transformers import LlavaNextVideoVideoProcessor
|
||||||
|
|
||||||
if is_torch_available:
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
@require_vision
|
@require_vision
|
||||||
class LlavaNextVideoProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
class LlavaNextVideoProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
||||||
|
@ -17,14 +17,11 @@ import unittest
|
|||||||
|
|
||||||
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
||||||
from transformers.testing_utils import require_torch, require_vision
|
from transformers.testing_utils import require_torch, require_vision
|
||||||
from transformers.utils import is_torch_available, is_torchvision_available, is_vision_available
|
from transformers.utils import is_torchvision_available, is_vision_available
|
||||||
|
|
||||||
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
if is_vision_available():
|
if is_vision_available():
|
||||||
if is_torchvision_available():
|
if is_torchvision_available():
|
||||||
from transformers import LlavaNextVideoVideoProcessor
|
from transformers import LlavaNextVideoVideoProcessor
|
||||||
|
@ -20,7 +20,7 @@ import unittest
|
|||||||
import torch
|
import torch
|
||||||
|
|
||||||
from transformers.testing_utils import require_torch, require_vision
|
from transformers.testing_utils import require_torch, require_vision
|
||||||
from transformers.utils import is_torch_available, is_torchvision_available, is_vision_available
|
from transformers.utils import is_torchvision_available, is_vision_available
|
||||||
|
|
||||||
from ...test_processing_common import ProcessorTesterMixin
|
from ...test_processing_common import ProcessorTesterMixin
|
||||||
|
|
||||||
@ -36,9 +36,6 @@ if is_vision_available():
|
|||||||
if is_torchvision_available():
|
if is_torchvision_available():
|
||||||
from transformers import LlavaOnevisionVideoProcessor
|
from transformers import LlavaOnevisionVideoProcessor
|
||||||
|
|
||||||
if is_torch_available:
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
@require_vision
|
@require_vision
|
||||||
@require_torch
|
@require_torch
|
||||||
|
@ -17,14 +17,11 @@ import unittest
|
|||||||
|
|
||||||
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
||||||
from transformers.testing_utils import require_torch, require_vision
|
from transformers.testing_utils import require_torch, require_vision
|
||||||
from transformers.utils import is_torch_available, is_torchvision_available, is_vision_available
|
from transformers.utils import is_torchvision_available, is_vision_available
|
||||||
|
|
||||||
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
if is_vision_available():
|
if is_vision_available():
|
||||||
if is_torchvision_available():
|
if is_torchvision_available():
|
||||||
from transformers import LlavaOnevisionVideoProcessor
|
from transformers import LlavaOnevisionVideoProcessor
|
||||||
|
@ -20,7 +20,7 @@ import numpy as np
|
|||||||
|
|
||||||
from transformers import PixtralProcessor
|
from transformers import PixtralProcessor
|
||||||
from transformers.testing_utils import require_vision
|
from transformers.testing_utils import require_vision
|
||||||
from transformers.utils import is_torch_available, is_vision_available
|
from transformers.utils import is_torch_available
|
||||||
|
|
||||||
from ...test_processing_common import ProcessorTesterMixin
|
from ...test_processing_common import ProcessorTesterMixin
|
||||||
|
|
||||||
@ -29,10 +29,6 @@ if is_torch_available():
|
|||||||
import torch
|
import torch
|
||||||
|
|
||||||
|
|
||||||
if is_vision_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
@require_vision
|
@require_vision
|
||||||
class Mistral3ProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
class Mistral3ProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
||||||
"""This tests Pixtral processor with the new `spatial_merge_size` argument in Mistral3."""
|
"""This tests Pixtral processor with the new `spatial_merge_size` argument in Mistral3."""
|
||||||
|
@ -23,7 +23,6 @@ from transformers import (
|
|||||||
PaliGemmaConfig,
|
PaliGemmaConfig,
|
||||||
PaliGemmaForConditionalGeneration,
|
PaliGemmaForConditionalGeneration,
|
||||||
is_torch_available,
|
is_torch_available,
|
||||||
is_vision_available,
|
|
||||||
)
|
)
|
||||||
from transformers.testing_utils import (
|
from transformers.testing_utils import (
|
||||||
is_flaky,
|
is_flaky,
|
||||||
@ -40,10 +39,6 @@ if is_torch_available():
|
|||||||
import torch
|
import torch
|
||||||
|
|
||||||
|
|
||||||
if is_vision_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
class PaliGemma2VisionText2TextModelTester:
|
class PaliGemma2VisionText2TextModelTester:
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
|
@ -17,15 +17,10 @@ import unittest
|
|||||||
|
|
||||||
from transformers import AutoProcessor, AutoTokenizer, Qwen2AudioProcessor, WhisperFeatureExtractor
|
from transformers import AutoProcessor, AutoTokenizer, Qwen2AudioProcessor, WhisperFeatureExtractor
|
||||||
from transformers.testing_utils import require_torch, require_torchaudio
|
from transformers.testing_utils import require_torch, require_torchaudio
|
||||||
from transformers.utils import is_torch_available
|
|
||||||
|
|
||||||
from ...test_processing_common import ProcessorTesterMixin
|
from ...test_processing_common import ProcessorTesterMixin
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available:
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
@require_torch
|
@require_torch
|
||||||
@require_torchaudio
|
@require_torchaudio
|
||||||
class Qwen2AudioProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
class Qwen2AudioProcessorTest(ProcessorTesterMixin, unittest.TestCase):
|
||||||
|
@ -21,15 +21,11 @@ import torch
|
|||||||
|
|
||||||
from transformers import TimesFmConfig, is_torch_available
|
from transformers import TimesFmConfig, is_torch_available
|
||||||
from transformers.testing_utils import require_torch, slow, torch_device
|
from transformers.testing_utils import require_torch, slow, torch_device
|
||||||
from transformers.utils import is_torch_fx_available
|
|
||||||
|
|
||||||
from ...test_configuration_common import ConfigTester
|
from ...test_configuration_common import ConfigTester
|
||||||
from ...test_modeling_common import ModelTesterMixin
|
from ...test_modeling_common import ModelTesterMixin
|
||||||
|
|
||||||
|
|
||||||
if is_torch_fx_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
if is_torch_available():
|
if is_torch_available():
|
||||||
from transformers import TimesFmModelForPrediction
|
from transformers import TimesFmModelForPrediction
|
||||||
|
|
||||||
|
@ -17,14 +17,11 @@ import unittest
|
|||||||
|
|
||||||
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
from transformers.image_utils import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD
|
||||||
from transformers.testing_utils import require_torch, require_vision
|
from transformers.testing_utils import require_torch, require_vision
|
||||||
from transformers.utils import is_torch_available, is_torchvision_available, is_vision_available
|
from transformers.utils import is_torchvision_available, is_vision_available
|
||||||
|
|
||||||
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
from ...test_video_processing_common import VideoProcessingTestMixin, prepare_video_inputs
|
||||||
|
|
||||||
|
|
||||||
if is_torch_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
if is_vision_available():
|
if is_vision_available():
|
||||||
if is_torchvision_available():
|
if is_torchvision_available():
|
||||||
from transformers import VideoLlavaVideoProcessor
|
from transformers import VideoLlavaVideoProcessor
|
||||||
|
@ -18,7 +18,7 @@ import unittest
|
|||||||
|
|
||||||
from transformers import VitPoseBackboneConfig
|
from transformers import VitPoseBackboneConfig
|
||||||
from transformers.testing_utils import require_torch, torch_device
|
from transformers.testing_utils import require_torch, torch_device
|
||||||
from transformers.utils import is_torch_available, is_vision_available
|
from transformers.utils import is_torch_available
|
||||||
|
|
||||||
from ...test_backbone_common import BackboneTesterMixin
|
from ...test_backbone_common import BackboneTesterMixin
|
||||||
from ...test_configuration_common import ConfigTester
|
from ...test_configuration_common import ConfigTester
|
||||||
@ -31,10 +31,6 @@ if is_torch_available():
|
|||||||
from transformers import VitPoseBackbone
|
from transformers import VitPoseBackbone
|
||||||
|
|
||||||
|
|
||||||
if is_vision_available():
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
class VitPoseBackboneModelTester:
|
class VitPoseBackboneModelTester:
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
|
Loading…
Reference in New Issue
Block a user