mGPT-13B / app.py
0x7o
Update app.py
99dc2e8 verified
raw
history blame
560 Bytes
import spaces
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
import gradio as gr
import torch
if torch.cuda.is_available():
tokenizer = AutoTokenizer.from_pretrained("ai-forever/mGPT-13B")
model = AutoModelForCausalLM.from_pretrained("ai-forever/mGPT-13B", load_in_8bit=True, device_map="auto")
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, device=0)
@spaces.GPU
def predict(text):
return pipe(text)
demo = gr.Interface(
fn=greet,
inputs=["text"],
outputs=["text"],
)
demo.launch()