Spaces:
Paused
Paused
File size: 2,000 Bytes
ef3a17c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 |
import gradio as gr
import cv2
import torch
import numpy as np
from torchvision import transforms
import torch
from pytorch_lightning import seed_everything
from torchvision.utils import save_image
from model_lib.modules import MoMA_main_modal
from model_lib.utils import parse_args
import os
os.environ["CUDA_VISIBLE_DEVICES"]="0"
title = "MoMA"
description = "This model has to run on GPU"
article = "<p style='text-align: center'><a href='https://news.machinelearning.sg/posts/beautiful_profile_pics_remove_background_image_with_deeplabv3/'>Blog</a> | <a href='https://github.com/eugenesiow/practical-ml'>Github Repo</a></p>"
def MoMA_demo(rgb, mask, subject, prompt):
# move the input and model to GPU for speed if available
with torch.no_grad():
generated_image = model.generate_images(rgb, mask, subject, prompt, strength=1.0, seed=2)
return generated_image
def inference(rgb, mask, subject, prompt):
result = MoMA_demo(rgb, mask, subject, prompt)
return result
seed_everything(0)
args = parse_args()
#load MoMA from HuggingFace. Auto download
model = MoMA_main_modal(args).to(args.device, dtype=torch.float16)
################ change texture ##################
# prompt = "A wooden sculpture of a car on the table."
# generated_image = model.generate_images(rgb_path, mask_path, subject, prompt, strength=0.4, seed=4, return_mask=True) # set strength to 0.4 for better prompt fidelity
# save_image(generated_image,f"{args.output_path}/{subject}_{prompt}.jpg")
gr.Interface(
inference,
[gr.Image(type="pil", label="Input RGB"),
gr.Image(type="pil", label="Input Mask"),
gr.Textbox(lines=1, label="subject"),
gr.Textbox(lines=5, label="Prompt")],
gr.Image(type="pil", label="Output"),
title=title,
description=description,
article=article,
examples=[["example_images/newImages/3.jpg",'example_images/newImages/3_mask.jpg','car','A car in autumn with falling leaves.']],
# enable_queue=True
).launch(debug=False) |