1inkusFace commited on
Commit
154abbe
·
verified ·
1 Parent(s): e1b3316

Update pipeline_stable_diffusion_3_ipa.py

Browse files
Files changed (1) hide show
  1. pipeline_stable_diffusion_3_ipa.py +4 -4
pipeline_stable_diffusion_3_ipa.py CHANGED
@@ -1161,7 +1161,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
1161
  clip_image_embeds_2 = clip_image_embeds_2.to(device, dtype=dtype)
1162
  clip_image_embeds_2 = self.image_encoder(clip_image_embeds_2, output_hidden_states=True).hidden_states[-2]
1163
  clip_image_embeds_2 = clip_image_embeds_2 * scale_2
1164
- image_prompt_embeds_list.append(image_prompt_embeds_2)
1165
  if clip_image_3 != None:
1166
  print('Using tertiary image.')
1167
  clip_image_3 = clip_image_3.resize((max(clip_image_3.size), max(clip_image_3.size)))
@@ -1170,7 +1170,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
1170
  clip_image_embeds_3 = clip_image_embeds_3.to(device, dtype=dtype)
1171
  clip_image_embeds_3 = self.image_encoder(clip_image_embeds_3, output_hidden_states=True).hidden_states[-2]
1172
  clip_image_embeds_3 = clip_image_embeds_3 * scale_3
1173
- image_prompt_embeds_list.append(image_prompt_embeds_3)
1174
  if clip_image_4 != None:
1175
  print('Using quaternary image.')
1176
  clip_image_4 = clip_image_4.resize((max(clip_image_4.size), max(clip_image_4.size)))
@@ -1179,7 +1179,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
1179
  clip_image_embeds_4 = clip_image_embeds_4.to(device, dtype=dtype)
1180
  clip_image_embeds_2 = self.image_encoder(clip_image_embeds_4, output_hidden_states=True).hidden_states[-2]
1181
  clip_image_embeds_4 = clip_image_embeds_4 * scale_4
1182
- image_prompt_embeds_list.append(image_prompt_embeds_4)
1183
  if clip_image_5 != None:
1184
  print('Using quinary image.')
1185
  clip_image_5 = clip_image_5.resize((max(clip_image_5.size), max(clip_image_5.size)))
@@ -1188,7 +1188,7 @@ class StableDiffusion3Pipeline(DiffusionPipeline, SD3LoraLoaderMixin, FromSingle
1188
  clip_image_embeds_5 = clip_image_embeds_5.to(device, dtype=dtype)
1189
  clip_image_embeds_5 = self.image_encoder(clip_image_embeds_5, output_hidden_states=True).hidden_states[-2]
1190
  clip_image_embeds_5 = clip_image_embeds_5 * scale_5
1191
- image_prompt_embeds_list.append(image_prompt_embeds_5)
1192
 
1193
  # Concatenate the image embeddings
1194
  ## clip_image_embeds = torch.mean(torch.stack(image_prompt_embeds_list), dim=0)
 
1161
  clip_image_embeds_2 = clip_image_embeds_2.to(device, dtype=dtype)
1162
  clip_image_embeds_2 = self.image_encoder(clip_image_embeds_2, output_hidden_states=True).hidden_states[-2]
1163
  clip_image_embeds_2 = clip_image_embeds_2 * scale_2
1164
+ image_prompt_embeds_list.append(clip_image_embeds_2)
1165
  if clip_image_3 != None:
1166
  print('Using tertiary image.')
1167
  clip_image_3 = clip_image_3.resize((max(clip_image_3.size), max(clip_image_3.size)))
 
1170
  clip_image_embeds_3 = clip_image_embeds_3.to(device, dtype=dtype)
1171
  clip_image_embeds_3 = self.image_encoder(clip_image_embeds_3, output_hidden_states=True).hidden_states[-2]
1172
  clip_image_embeds_3 = clip_image_embeds_3 * scale_3
1173
+ image_prompt_embeds_list.append(clip_image_embeds_3)
1174
  if clip_image_4 != None:
1175
  print('Using quaternary image.')
1176
  clip_image_4 = clip_image_4.resize((max(clip_image_4.size), max(clip_image_4.size)))
 
1179
  clip_image_embeds_4 = clip_image_embeds_4.to(device, dtype=dtype)
1180
  clip_image_embeds_2 = self.image_encoder(clip_image_embeds_4, output_hidden_states=True).hidden_states[-2]
1181
  clip_image_embeds_4 = clip_image_embeds_4 * scale_4
1182
+ image_prompt_embeds_list.append(clip_image_embeds_4)
1183
  if clip_image_5 != None:
1184
  print('Using quinary image.')
1185
  clip_image_5 = clip_image_5.resize((max(clip_image_5.size), max(clip_image_5.size)))
 
1188
  clip_image_embeds_5 = clip_image_embeds_5.to(device, dtype=dtype)
1189
  clip_image_embeds_5 = self.image_encoder(clip_image_embeds_5, output_hidden_states=True).hidden_states[-2]
1190
  clip_image_embeds_5 = clip_image_embeds_5 * scale_5
1191
+ image_prompt_embeds_list.append(clip_image_embeds_5)
1192
 
1193
  # Concatenate the image embeddings
1194
  ## clip_image_embeds = torch.mean(torch.stack(image_prompt_embeds_list), dim=0)