Corrected the Chat template
#17
by
SalmanFaroz
- opened
README.md
CHANGED
@@ -305,9 +305,9 @@ device = "cuda" # the device to load the model onto
|
|
305 |
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
|
306 |
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
|
307 |
|
308 |
-
text = "<s>[INST] What is your favourite condiment? [/INST]
|
309 |
-
|
310 |
-
|
311 |
|
312 |
encodeds = tokenizer(text, return_tensors="pt", add_special_tokens=False)
|
313 |
|
|
|
305 |
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
|
306 |
tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
|
307 |
|
308 |
+
text = """<s>[INST] What is your favourite condiment? [/INST]
|
309 |
+
Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s>
|
310 |
+
[INST] Do you have mayonnaise recipes? [/INST]"""
|
311 |
|
312 |
encodeds = tokenizer(text, return_tensors="pt", add_special_tokens=False)
|
313 |
|