Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -169,7 +169,7 @@ pipe = load_and_prepare_model()
|
|
169 |
checkpoint = "ford442/Phi-3.5-mini-instruct-bf16"
|
170 |
#captioner = pipeline(model="ydshieh/vit-gpt2-coco-en",device='cuda:0', task="image-to-text")
|
171 |
#captioner_2 = pipeline(model="Salesforce/blip-image-captioning-base",device='cuda', task="image-to-text")
|
172 |
-
captioner_2 = pipeline(model="ford442/blip-image-to-text-large-bf16",device='cuda', task="image-to-text")
|
173 |
#model5 = Blip2ForConditionalGeneration.from_pretrained("ford442/blip2-image-to-text-bf16").to('cuda')
|
174 |
#processor5 = Blip2Processor.from_pretrained("ford442/blip2-image-to-text-bf16", device_map='cuda')
|
175 |
#txt_tokenizer = AutoTokenizer.from_pretrained(checkpoint, device_map='cuda', add_prefix_space=False)
|
@@ -408,7 +408,7 @@ def generate_30(
|
|
408 |
#sd_image_b.resize((height,width), Image.LANCZOS)
|
409 |
sd_image_b.resize((224,224), Image.LANCZOS)
|
410 |
#caption.append(captioner(sd_image_b))
|
411 |
-
caption.append(captioner2(sd_image_b))
|
412 |
#caption.append(captioner_3(sd_image_b))
|
413 |
caption_2.append(captioning(sd_image_b))
|
414 |
else:
|
@@ -418,7 +418,7 @@ def generate_30(
|
|
418 |
#sd_image_c.resize((height,width), Image.LANCZOS)
|
419 |
sd_image_c.resize((224,224), Image.LANCZOS)
|
420 |
#caption.append(captioner(sd_image_c))
|
421 |
-
caption.append(captioner2(sd_image_c))
|
422 |
#caption.append(captioner_3(sd_image_c))
|
423 |
caption_2.append(captioning(sd_image_c))
|
424 |
else:
|
@@ -428,7 +428,7 @@ def generate_30(
|
|
428 |
#sd_image_d.resize((height,width), Image.LANCZOS)
|
429 |
sd_image_d.resize((224,224), Image.LANCZOS)
|
430 |
#caption.append(captioner(sd_image_d))
|
431 |
-
caption.append(captioner2(sd_image_d))
|
432 |
#caption.append(captioner_3(sd_image_d))
|
433 |
caption_2.append(captioning(sd_image_d))
|
434 |
else:
|
@@ -438,7 +438,7 @@ def generate_30(
|
|
438 |
#sd_image_e.resize((height,width), Image.LANCZOS)
|
439 |
sd_image_e.resize((224,224), Image.LANCZOS)
|
440 |
#caption.append(captioner(sd_image_e))
|
441 |
-
caption.append(captioner2(sd_image_e))
|
442 |
#caption.append(captioner_3(sd_image_e))
|
443 |
caption_2.append(captioning(sd_image_e))
|
444 |
else:
|
@@ -559,7 +559,7 @@ def generate_60(
|
|
559 |
#sd_image_b.resize((height,width), Image.LANCZOS)
|
560 |
sd_image_b.resize((224,224), Image.LANCZOS)
|
561 |
#caption.append(captioner(sd_image_b))
|
562 |
-
caption.append(captioner2(sd_image_b))
|
563 |
#caption.append(captioner_3(sd_image_b))
|
564 |
caption_2.append(captioning(sd_image_b))
|
565 |
else:
|
@@ -711,7 +711,7 @@ def generate_90(
|
|
711 |
#sd_image_b.resize((height,width), Image.LANCZOS)
|
712 |
sd_image_b.resize((224,224), Image.LANCZOS)
|
713 |
#caption.append(captioner(sd_image_b))
|
714 |
-
caption.append(captioner2(sd_image_b))
|
715 |
#caption.append(captioner_3(sd_image_b))
|
716 |
caption_2.append(captioning(sd_image_b))
|
717 |
else:
|
|
|
169 |
checkpoint = "ford442/Phi-3.5-mini-instruct-bf16"
|
170 |
#captioner = pipeline(model="ydshieh/vit-gpt2-coco-en",device='cuda:0', task="image-to-text")
|
171 |
#captioner_2 = pipeline(model="Salesforce/blip-image-captioning-base",device='cuda', task="image-to-text")
|
172 |
+
#captioner_2 = pipeline(model="ford442/blip-image-to-text-large-bf16",device='cuda', task="image-to-text")
|
173 |
#model5 = Blip2ForConditionalGeneration.from_pretrained("ford442/blip2-image-to-text-bf16").to('cuda')
|
174 |
#processor5 = Blip2Processor.from_pretrained("ford442/blip2-image-to-text-bf16", device_map='cuda')
|
175 |
#txt_tokenizer = AutoTokenizer.from_pretrained(checkpoint, device_map='cuda', add_prefix_space=False)
|
|
|
408 |
#sd_image_b.resize((height,width), Image.LANCZOS)
|
409 |
sd_image_b.resize((224,224), Image.LANCZOS)
|
410 |
#caption.append(captioner(sd_image_b))
|
411 |
+
##caption.append(captioner2(sd_image_b))
|
412 |
#caption.append(captioner_3(sd_image_b))
|
413 |
caption_2.append(captioning(sd_image_b))
|
414 |
else:
|
|
|
418 |
#sd_image_c.resize((height,width), Image.LANCZOS)
|
419 |
sd_image_c.resize((224,224), Image.LANCZOS)
|
420 |
#caption.append(captioner(sd_image_c))
|
421 |
+
#caption.append(captioner2(sd_image_c))
|
422 |
#caption.append(captioner_3(sd_image_c))
|
423 |
caption_2.append(captioning(sd_image_c))
|
424 |
else:
|
|
|
428 |
#sd_image_d.resize((height,width), Image.LANCZOS)
|
429 |
sd_image_d.resize((224,224), Image.LANCZOS)
|
430 |
#caption.append(captioner(sd_image_d))
|
431 |
+
#caption.append(captioner2(sd_image_d))
|
432 |
#caption.append(captioner_3(sd_image_d))
|
433 |
caption_2.append(captioning(sd_image_d))
|
434 |
else:
|
|
|
438 |
#sd_image_e.resize((height,width), Image.LANCZOS)
|
439 |
sd_image_e.resize((224,224), Image.LANCZOS)
|
440 |
#caption.append(captioner(sd_image_e))
|
441 |
+
#caption.append(captioner2(sd_image_e))
|
442 |
#caption.append(captioner_3(sd_image_e))
|
443 |
caption_2.append(captioning(sd_image_e))
|
444 |
else:
|
|
|
559 |
#sd_image_b.resize((height,width), Image.LANCZOS)
|
560 |
sd_image_b.resize((224,224), Image.LANCZOS)
|
561 |
#caption.append(captioner(sd_image_b))
|
562 |
+
#caption.append(captioner2(sd_image_b))
|
563 |
#caption.append(captioner_3(sd_image_b))
|
564 |
caption_2.append(captioning(sd_image_b))
|
565 |
else:
|
|
|
711 |
#sd_image_b.resize((height,width), Image.LANCZOS)
|
712 |
sd_image_b.resize((224,224), Image.LANCZOS)
|
713 |
#caption.append(captioner(sd_image_b))
|
714 |
+
#caption.append(captioner2(sd_image_b))
|
715 |
#caption.append(captioner_3(sd_image_b))
|
716 |
caption_2.append(captioning(sd_image_b))
|
717 |
else:
|