From 69632aadb70cbe3121ee3852b461fb1ccb6cbc37 Mon Sep 17 00:00:00 2001 From: Yih-Dar <2521628+ydshieh@users.noreply.github.com> Date: Tue, 25 Mar 2025 16:16:06 +0100 Subject: [PATCH] Update after #36962 (#36965) update Co-authored-by: ydshieh --- src/transformers/cache_utils.py | 4 +++- src/transformers/models/qwen2_audio/processing_qwen2_audio.py | 2 +- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/src/transformers/cache_utils.py b/src/transformers/cache_utils.py index 8fd8e963967..416bc2564ec 100644 --- a/src/transformers/cache_utils.py +++ b/src/transformers/cache_utils.py @@ -631,8 +631,10 @@ class OffloadedCache(DynamicCache): def prefetch_layer(self, layer_idx: int): "Starts prefetching the next layer cache" if layer_idx < len(self): - with self.prefetch_stream if is_torch_greater_or_equal("2.7", accept_dev=True) else torch.cuda.stream( + with ( self.prefetch_stream + if is_torch_greater_or_equal("2.7", accept_dev=True) + else torch.cuda.stream(self.prefetch_stream) ): # Prefetch next layer tensors to GPU device = self.original_device[layer_idx] diff --git a/src/transformers/models/qwen2_audio/processing_qwen2_audio.py b/src/transformers/models/qwen2_audio/processing_qwen2_audio.py index 9e65c6d6634..0daa90c5643 100644 --- a/src/transformers/models/qwen2_audio/processing_qwen2_audio.py +++ b/src/transformers/models/qwen2_audio/processing_qwen2_audio.py @@ -129,7 +129,7 @@ class Qwen2AudioProcessor(ProcessorMixin): if audio is not None: # ensure we have as much audios as audio tokens num_audio_tokens = sum(sample.count(self.audio_token) for sample in text) - num_audios = 1 if type(audio) == np.ndarray else len(audio) + num_audios = 1 if type(audio) is np.ndarray else len(audio) if num_audio_tokens != num_audios: raise ValueError( f"Found {num_audio_tokens} {self.audio_token} token{'s' if num_audio_tokens > 1 else ''} in provided text but received {num_audios} audio{'s' if num_audios > 1 else ''}"