Spaces:
Running
on
Zero
Running
on
Zero
Update pipeline_stable_diffusion_3_ipa.py
Browse files
pipeline_stable_diffusion_3_ipa.py
CHANGED
@@ -1161,7 +1161,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
|
|
1161 |
clip_image_embeds_2 = clip_image_embeds_2.to(device, dtype=dtype)
|
1162 |
clip_image_embeds_2 = self.image_encoder(clip_image_embeds_2, output_hidden_states=True).hidden_states[-2]
|
1163 |
clip_image_embeds_2 = clip_image_embeds_2 * scale_2
|
1164 |
-
image_prompt_embeds_list.append(
|
1165 |
if clip_image_3 != None:
|
1166 |
print('Using tertiary image.')
|
1167 |
clip_image_3 = clip_image_3.resize((max(clip_image_3.size), max(clip_image_3.size)))
|
@@ -1170,7 +1170,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
|
|
1170 |
clip_image_embeds_3 = clip_image_embeds_3.to(device, dtype=dtype)
|
1171 |
clip_image_embeds_3 = self.image_encoder(clip_image_embeds_3, output_hidden_states=True).hidden_states[-2]
|
1172 |
clip_image_embeds_3 = clip_image_embeds_3 * scale_3
|
1173 |
-
image_prompt_embeds_list.append(
|
1174 |
if clip_image_4 != None:
|
1175 |
print('Using quaternary image.')
|
1176 |
clip_image_4 = clip_image_4.resize((max(clip_image_4.size), max(clip_image_4.size)))
|
@@ -1179,7 +1179,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
|
|
1179 |
clip_image_embeds_4 = clip_image_embeds_4.to(device, dtype=dtype)
|
1180 |
clip_image_embeds_2 = self.image_encoder(clip_image_embeds_4, output_hidden_states=True).hidden_states[-2]
|
1181 |
clip_image_embeds_4 = clip_image_embeds_4 * scale_4
|
1182 |
-
image_prompt_embeds_list.append(
|
1183 |
if clip_image_5 != None:
|
1184 |
print('Using quinary image.')
|
1185 |
clip_image_5 = clip_image_5.resize((max(clip_image_5.size), max(clip_image_5.size)))
|
@@ -1188,7 +1188,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
|
|
1188 |
clip_image_embeds_5 = clip_image_embeds_5.to(device, dtype=dtype)
|
1189 |
clip_image_embeds_5 = self.image_encoder(clip_image_embeds_5, output_hidden_states=True).hidden_states[-2]
|
1190 |
clip_image_embeds_5 = clip_image_embeds_5 * scale_5
|
1191 |
-
image_prompt_embeds_list.append(
|
1192 |
|
1193 |
# Concatenate the image embeddings
|
1194 |
## clip_image_embeds = torch.mean(torch.stack(image_prompt_embeds_list), dim=0)
|
|
|
1161 |
clip_image_embeds_2 = clip_image_embeds_2.to(device, dtype=dtype)
|
1162 |
clip_image_embeds_2 = self.image_encoder(clip_image_embeds_2, output_hidden_states=True).hidden_states[-2]
|
1163 |
clip_image_embeds_2 = clip_image_embeds_2 * scale_2
|
1164 |
+
image_prompt_embeds_list.append(clip_image_embeds_2)
|
1165 |
if clip_image_3 != None:
|
1166 |
print('Using tertiary image.')
|
1167 |
clip_image_3 = clip_image_3.resize((max(clip_image_3.size), max(clip_image_3.size)))
|
|
|
1170 |
clip_image_embeds_3 = clip_image_embeds_3.to(device, dtype=dtype)
|
1171 |
clip_image_embeds_3 = self.image_encoder(clip_image_embeds_3, output_hidden_states=True).hidden_states[-2]
|
1172 |
clip_image_embeds_3 = clip_image_embeds_3 * scale_3
|
1173 |
+
image_prompt_embeds_list.append(clip_image_embeds_3)
|
1174 |
if clip_image_4 != None:
|
1175 |
print('Using quaternary image.')
|
1176 |
clip_image_4 = clip_image_4.resize((max(clip_image_4.size), max(clip_image_4.size)))
|
|
|
1179 |
clip_image_embeds_4 = clip_image_embeds_4.to(device, dtype=dtype)
|
1180 |
clip_image_embeds_2 = self.image_encoder(clip_image_embeds_4, output_hidden_states=True).hidden_states[-2]
|
1181 |
clip_image_embeds_4 = clip_image_embeds_4 * scale_4
|
1182 |
+
image_prompt_embeds_list.append(clip_image_embeds_4)
|
1183 |
if clip_image_5 != None:
|
1184 |
print('Using quinary image.')
|
1185 |
clip_image_5 = clip_image_5.resize((max(clip_image_5.size), max(clip_image_5.size)))
|
|
|
1188 |
clip_image_embeds_5 = clip_image_embeds_5.to(device, dtype=dtype)
|
1189 |
clip_image_embeds_5 = self.image_encoder(clip_image_embeds_5, output_hidden_states=True).hidden_states[-2]
|
1190 |
clip_image_embeds_5 = clip_image_embeds_5 * scale_5
|
1191 |
+
image_prompt_embeds_list.append(clip_image_embeds_5)
|
1192 |
|
1193 |
# Concatenate the image embeddings
|
1194 |
## clip_image_embeds = torch.mean(torch.stack(image_prompt_embeds_list), dim=0)
|