From 84a6570e7bce91ba7d18c0782186241c5f1fde75 Mon Sep 17 00:00:00 2001 From: Youssef Adarrab <104783077+youssefadr@users.noreply.github.com> Date: Tue, 18 Apr 2023 23:31:59 +0000 Subject: [PATCH] Make ClipSeg compatible with model parallelism (#22844) --- src/transformers/models/clipseg/modeling_clipseg.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/src/transformers/models/clipseg/modeling_clipseg.py b/src/transformers/models/clipseg/modeling_clipseg.py index 3ec81b33fb7..1dcf2f47446 100644 --- a/src/transformers/models/clipseg/modeling_clipseg.py +++ b/src/transformers/models/clipseg/modeling_clipseg.py @@ -1480,6 +1480,8 @@ class CLIPSegForImageSegmentation(CLIPSegPreTrainedModel): loss = None if labels is not None: + # move labels to the correct device to enable PP + labels = labels.to(logits.device) loss_fn = nn.BCEWithLogitsLoss() loss = loss_fn(logits, labels)