Spaces:
Runtime error
Runtime error
File size: 5,865 Bytes
a5fe767 54eea67 a5fe767 54eea67 a5fe767 a0a2e08 2d90420 152f0ff 2d90420 fa26c50 fed9a6c fa26c50 f8fce88 54eea67 19e2836 a5fe767 a0a2e08 a5fe767 a0a2e08 a5fe767 a0a2e08 a5fe767 fa26c50 a5fe767 75a2d1c 19e2836 a5fe767 75a2d1c a5fe767 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 |
import os
import gc
import torch
import torch.nn as nn
import argparse
import gradio as gr
import time
from transformers import AutoTokenizer, LlamaForCausalLM
from utils import SteamGenerationMixin
import requests
auth_token = os.getenv("Zimix")
url_api = os.getenv('api_url')
# print(url_api)
URL = f'http://120.234.0.81:8808/{url_api}'
def cc(q,r):
try:
requests.request('get',URL,params={'query':q,'response':r,'time':time.time()})
except:
print('推送失败-_- !')
class MindBot(object):
def __init__(self, model_path, tokenizer_path,if_int8=False):
# self.device = torch.device("cuda")
# device_ids = [1, 2]
if if_int8:
self.model = SteamGenerationMixin.from_pretrained(model_path, device_map='auto', load_in_8bit=True,use_auth_token=auth_token).eval()
else:
self.model = SteamGenerationMixin.from_pretrained(model_path, device_map='auto',use_auth_token=auth_token).half().eval()
self.tokenizer = AutoTokenizer.from_pretrained(tokenizer_path,use_auth_token=auth_token)
# sp_tokens = {'additional_special_tokens': ['<human>', '<bot>']}
# self.tokenizer.add_special_tokens(sp_tokens)
self.history = []
def build_prompt(self, instruction, history, human='<human>', bot='<bot>'):
pmt = ''
if len(history) > 0:
for line in history:
pmt += f'{human}: {line[0].strip()}\n{bot}: {line[1]}\n'
pmt += f'{human}: {instruction.strip()}\n{bot}: \n'
return pmt
def common_generate(self, instruction, clear_history=False, max_memory=1024):
if clear_history:
self.history = []
prompt = self.build_prompt(instruction, self.history)
input_ids = self.tokenizer(prompt, return_tensors="pt").input_ids
if input_ids.shape[1] > max_memory:
input_ids = input_ids[:, -max_memory:]
prompt_len = input_ids.shape[1]
# common method
generation_output = self.model.generate(
input_ids.cuda(),
max_new_tokens=1024,
do_sample=True,
top_p=0.85,
temperature=0.8,
repetition_penalty=1.,
eos_token_id=2,
bos_token_id=1,
pad_token_id=0
)
s = generation_output[0][prompt_len:]
output = self.tokenizer.decode(s, skip_special_tokens=True)
# output = output
output = output.replace("Belle", "IDEA")
self.history.append((instruction, output))
print('api history: ======> \n', self.history)
return output
def interaction(
self,
instruction,
history,
max_memory=1024
):
prompt = self.build_prompt(instruction, history)
input_ids = self.tokenizer(prompt, return_tensors="pt").input_ids
if input_ids.shape[1] > max_memory:
input_ids = input_ids[:, -max_memory:]
prompt_len = input_ids.shape[1]
# stream generation method
try:
tmp = history.copy()
output = ''
with torch.no_grad():
for generation_output in self.model.stream_generate(
input_ids.cuda(),
max_new_tokens=1024,
do_sample=True,
top_p=0.85,
temperature=0.8,
repetition_penalty=1.,
eos_token_id=2,
bos_token_id=1,
pad_token_id=0
):
s = generation_output[0][prompt_len:]
output = self.tokenizer.decode(s, skip_special_tokens=True)
output = output.replace('\n', '<br>')
tmp.append((instruction, output))
yield '', tmp
tmp.pop()
# gc.collect()
# torch.cuda.empty_cache()
history.append((instruction, output))
print('input -----> \n', prompt)
print('output -------> \n', output)
print('history: ======> \n', history)
cc(prompt,output)
except torch.cuda.OutOfMemoryError:
gc.collect()
torch.cuda.empty_cache()
self.model.empty_cache()
return "", history
def new_chat_bot(self):
with gr.Blocks(title='IDEA Ziya', css=".gradio-container {max-width: 50% !important;} .bgcolor {color: white !important; background: #FFA500 !important;}") as demo:
gr.Markdown("<center><h1>IDEA Ziya</h1></center>")
gr.Markdown("<center>本页面基于hugging face支持的设备搭建 模型版本v1.1</center>")
with gr.Row():
chatbot = gr.Chatbot(label='Ziya').style(height=500)
with gr.Row():
msg = gr.Textbox(label="Input")
with gr.Row():
with gr.Column(scale=0.5):
clear = gr.Button("Clear")
with gr.Column(scale=0.5):
submit = gr.Button("Submit", elem_classes='bgcolor')
msg.submit(self.interaction, [msg, chatbot], [msg, chatbot])
clear.click(lambda: None, None, chatbot, queue=False)
submit.click(self.interaction, [msg, chatbot], [msg, chatbot])
return demo.queue(concurrency_count=5)
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument(
"--model_path",
type=str,
default="/cognitive_comp/songchao/checkpoints/global_step3200-hf"
)
args = parser.parse_args()
mind_bot = MindBot(args.model_path)
demo = mind_bot.new_chat_bot()
|