izh97 commited on
Commit
1580888
·
verified ·
1 Parent(s): 9602507

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -11
app.py CHANGED
@@ -18,17 +18,17 @@ generation_config = GenerationConfig(
18
  pad_token_id=tokenizer.unk_token_id
19
  )
20
  def ask(text):
21
- tokenizer = AutoTokenizer.from_pretrained("izh97/zephyr-beta-climate-change-assistant")
22
- model = AutoPeftModelForCausalLM.from_pretrained(
23
- "izh97/zephyr-beta-climate-change-assistant",
24
- low_cpu_mem_usage=True,
25
- return_dict=True,
26
- torch_dtype=torch.float16,
27
- device_map="cuda")
28
-
29
- inputs = tokenizer.apply_chat_template(text, tokenize=True, add_generation_prompt=True, return_tensors="pt").to("cuda")
30
- outputs = model.generate(**inputs, generation_config=generation_config,
31
  return_dict_in_generate=True)
32
- return tokenizer.decode(outputs[0], skip_special_tokens=True)
 
33
  iface = gr.Interface(fn=ask, inputs="text", outputs="text")
34
  iface.launch()
 
18
  pad_token_id=tokenizer.unk_token_id
19
  )
20
  def ask(text):
21
+ messages = [
22
+ {
23
+ "role": "user",
24
+ "content": str(text),
25
+ },
26
+ ]
27
+ inputs = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt").to("cuda")
28
+ inputs_length = inputs.shape[1]
29
+ outputs = model.generate(inputs, generation_config=generation_config,
 
30
  return_dict_in_generate=True)
31
+ outputs = outputs.sequences[0, inputs_length:]
32
+ return tokenizer.decode(outputs, skip_special_tokens=True)
33
  iface = gr.Interface(fn=ask, inputs="text", outputs="text")
34
  iface.launch()