Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -9,6 +9,7 @@ checkpoint = "microsoft/phi-2"
|
|
9 |
# Download and load model and tokenizer
|
10 |
tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
|
11 |
model = AutoModelForCausalLM.from_pretrained(checkpoint, torch_dtype=torch.float32, device_map="cpu", trust_remote_code=True)
|
|
|
12 |
|
13 |
# Text generation pipeline
|
14 |
phi2 = pipeline("text-generation", tokenizer=tokenizer, model=model)
|
@@ -27,10 +28,10 @@ def generate(prompt, chat_history, max_new_tokens):
|
|
27 |
final_prompt += "Output:"
|
28 |
|
29 |
generated_text = phi2(final_prompt, max_new_tokens=max_new_tokens)[0]["generated_text"]
|
30 |
-
response = generated_text.
|
31 |
|
32 |
if "User:" in response:
|
33 |
-
response = response.split("User:")[0]
|
34 |
|
35 |
if "Assistant:" in response:
|
36 |
response = response.split("Assistant:")[1].strip()
|
|
|
9 |
# Download and load model and tokenizer
|
10 |
tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
|
11 |
model = AutoModelForCausalLM.from_pretrained(checkpoint, torch_dtype=torch.float32, device_map="cpu", trust_remote_code=True)
|
12 |
+
model.config.eos_token_id = tokenizer.eos_token_id
|
13 |
|
14 |
# Text generation pipeline
|
15 |
phi2 = pipeline("text-generation", tokenizer=tokenizer, model=model)
|
|
|
28 |
final_prompt += "Output:"
|
29 |
|
30 |
generated_text = phi2(final_prompt, max_new_tokens=max_new_tokens)[0]["generated_text"]
|
31 |
+
response = generated_text[len(final_prompt):].strip()
|
32 |
|
33 |
if "User:" in response:
|
34 |
+
response = response.split("User:")[0].strip()
|
35 |
|
36 |
if "Assistant:" in response:
|
37 |
response = response.split("Assistant:")[1].strip()
|