futranbg's picture
Update app.py
4a44055
import os
import gradio as gr
from langchain.llms import HuggingFaceHub
llama_repo = os.getenv('HF_MODEL_LLAMA_REPO')
starchat_repo = os.getenv('HF_MODEL_STARCHAT_REPO')
llamma_template = """<s>[INST]<<SYS>>I want you to act as document language translator. You do translation {source} texts in document into then you return to me the translated document AND DO NOTHING ELSE.<</SYS>>[/INST]
[INST]Begin of the document:
{query}
End of the document.[/INST]
{target} translated document:
"""
starchat_template = """<|system|>I want you to act as document language translator. You do translation {source} texts in document into then you return to me the translated document AND DO NOTHING ELSE.<</SYS>>
Begin of the document:
{query}
End of the document<|end|>
<|assistant|>
{target} translated document:
"""
model_kwargs={
"max_new_tokens":2048,
"temperature": 0.01,
"truncate": 4096,
"seed" : 256,
"stop" : ["</s>","<|endoftext|>","<|end|>"],
}
llm1 = HuggingFaceHub(repo_id=llama_repo, task="text-generation", model_kwargs=model_kwargs)
llm2 = HuggingFaceHub(repo_id=starchat_repo, task="text-generation", model_kwargs=model_kwargs)
def translation(source, target, text):
response = text
try:
input_prompt = llamma_template.replace("{source}", source)
input_prompt = input_prompt.replace("{target}", target)
input_prompt = input_prompt.replace("{query}", text)
response=llm1(input_prompt)
except Exception as e:
print(f"ERROR: LLM show {e}")
input_prompt = starchat_template.replace("{source}", source)
input_prompt = input_prompt.replace("{target}", target)
input_prompt = input_prompt.replace("{query}", text)
response=llm2(input_prompt).replace("<|end|>","")
return response
gr.Interface(translation, inputs=["text","text","text"], outputs="text").launch()