1inkusFace commited on
Commit
58309fe
·
verified ·
1 Parent(s): b4eb9ed

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -169,7 +169,7 @@ pipe = load_and_prepare_model()
169
  checkpoint = "ford442/Phi-3.5-mini-instruct-bf16"
170
  #captioner = pipeline(model="ydshieh/vit-gpt2-coco-en",device='cuda:0', task="image-to-text")
171
  #captioner_2 = pipeline(model="Salesforce/blip-image-captioning-base",device='cuda', task="image-to-text")
172
- captioner_2 = pipeline(model="ford442/blip-image-to-text-large-bf16",device='cuda', task="image-to-text")
173
  #model5 = Blip2ForConditionalGeneration.from_pretrained("ford442/blip2-image-to-text-bf16").to('cuda')
174
  #processor5 = Blip2Processor.from_pretrained("ford442/blip2-image-to-text-bf16", device_map='cuda')
175
  #txt_tokenizer = AutoTokenizer.from_pretrained(checkpoint, device_map='cuda', add_prefix_space=False)
@@ -408,7 +408,7 @@ def generate_30(
408
  #sd_image_b.resize((height,width), Image.LANCZOS)
409
  sd_image_b.resize((224,224), Image.LANCZOS)
410
  #caption.append(captioner(sd_image_b))
411
- caption.append(captioner2(sd_image_b))
412
  #caption.append(captioner_3(sd_image_b))
413
  caption_2.append(captioning(sd_image_b))
414
  else:
@@ -418,7 +418,7 @@ def generate_30(
418
  #sd_image_c.resize((height,width), Image.LANCZOS)
419
  sd_image_c.resize((224,224), Image.LANCZOS)
420
  #caption.append(captioner(sd_image_c))
421
- caption.append(captioner2(sd_image_c))
422
  #caption.append(captioner_3(sd_image_c))
423
  caption_2.append(captioning(sd_image_c))
424
  else:
@@ -428,7 +428,7 @@ def generate_30(
428
  #sd_image_d.resize((height,width), Image.LANCZOS)
429
  sd_image_d.resize((224,224), Image.LANCZOS)
430
  #caption.append(captioner(sd_image_d))
431
- caption.append(captioner2(sd_image_d))
432
  #caption.append(captioner_3(sd_image_d))
433
  caption_2.append(captioning(sd_image_d))
434
  else:
@@ -438,7 +438,7 @@ def generate_30(
438
  #sd_image_e.resize((height,width), Image.LANCZOS)
439
  sd_image_e.resize((224,224), Image.LANCZOS)
440
  #caption.append(captioner(sd_image_e))
441
- caption.append(captioner2(sd_image_e))
442
  #caption.append(captioner_3(sd_image_e))
443
  caption_2.append(captioning(sd_image_e))
444
  else:
@@ -559,7 +559,7 @@ def generate_60(
559
  #sd_image_b.resize((height,width), Image.LANCZOS)
560
  sd_image_b.resize((224,224), Image.LANCZOS)
561
  #caption.append(captioner(sd_image_b))
562
- caption.append(captioner2(sd_image_b))
563
  #caption.append(captioner_3(sd_image_b))
564
  caption_2.append(captioning(sd_image_b))
565
  else:
@@ -711,7 +711,7 @@ def generate_90(
711
  #sd_image_b.resize((height,width), Image.LANCZOS)
712
  sd_image_b.resize((224,224), Image.LANCZOS)
713
  #caption.append(captioner(sd_image_b))
714
- caption.append(captioner2(sd_image_b))
715
  #caption.append(captioner_3(sd_image_b))
716
  caption_2.append(captioning(sd_image_b))
717
  else:
 
169
  checkpoint = "ford442/Phi-3.5-mini-instruct-bf16"
170
  #captioner = pipeline(model="ydshieh/vit-gpt2-coco-en",device='cuda:0', task="image-to-text")
171
  #captioner_2 = pipeline(model="Salesforce/blip-image-captioning-base",device='cuda', task="image-to-text")
172
+ #captioner_2 = pipeline(model="ford442/blip-image-to-text-large-bf16",device='cuda', task="image-to-text")
173
  #model5 = Blip2ForConditionalGeneration.from_pretrained("ford442/blip2-image-to-text-bf16").to('cuda')
174
  #processor5 = Blip2Processor.from_pretrained("ford442/blip2-image-to-text-bf16", device_map='cuda')
175
  #txt_tokenizer = AutoTokenizer.from_pretrained(checkpoint, device_map='cuda', add_prefix_space=False)
 
408
  #sd_image_b.resize((height,width), Image.LANCZOS)
409
  sd_image_b.resize((224,224), Image.LANCZOS)
410
  #caption.append(captioner(sd_image_b))
411
+ ##caption.append(captioner2(sd_image_b))
412
  #caption.append(captioner_3(sd_image_b))
413
  caption_2.append(captioning(sd_image_b))
414
  else:
 
418
  #sd_image_c.resize((height,width), Image.LANCZOS)
419
  sd_image_c.resize((224,224), Image.LANCZOS)
420
  #caption.append(captioner(sd_image_c))
421
+ #caption.append(captioner2(sd_image_c))
422
  #caption.append(captioner_3(sd_image_c))
423
  caption_2.append(captioning(sd_image_c))
424
  else:
 
428
  #sd_image_d.resize((height,width), Image.LANCZOS)
429
  sd_image_d.resize((224,224), Image.LANCZOS)
430
  #caption.append(captioner(sd_image_d))
431
+ #caption.append(captioner2(sd_image_d))
432
  #caption.append(captioner_3(sd_image_d))
433
  caption_2.append(captioning(sd_image_d))
434
  else:
 
438
  #sd_image_e.resize((height,width), Image.LANCZOS)
439
  sd_image_e.resize((224,224), Image.LANCZOS)
440
  #caption.append(captioner(sd_image_e))
441
+ #caption.append(captioner2(sd_image_e))
442
  #caption.append(captioner_3(sd_image_e))
443
  caption_2.append(captioning(sd_image_e))
444
  else:
 
559
  #sd_image_b.resize((height,width), Image.LANCZOS)
560
  sd_image_b.resize((224,224), Image.LANCZOS)
561
  #caption.append(captioner(sd_image_b))
562
+ #caption.append(captioner2(sd_image_b))
563
  #caption.append(captioner_3(sd_image_b))
564
  caption_2.append(captioning(sd_image_b))
565
  else:
 
711
  #sd_image_b.resize((height,width), Image.LANCZOS)
712
  sd_image_b.resize((224,224), Image.LANCZOS)
713
  #caption.append(captioner(sd_image_b))
714
+ #caption.append(captioner2(sd_image_b))
715
  #caption.append(captioner_3(sd_image_b))
716
  caption_2.append(captioning(sd_image_b))
717
  else: