Update README.md
#3
by
Studeni
- opened
README.md
CHANGED
@@ -107,13 +107,14 @@ pip3 install git+https://github.com/casper-hansen/AutoAWQ.git@1c5ccc791fa2cb0697
|
|
107 |
```python
|
108 |
from awq import AutoAWQForCausalLM
|
109 |
from transformers import AutoTokenizer
|
|
|
110 |
|
111 |
model_name_or_path = "TheBloke/Mistral-7B-v0.1-AWQ"
|
112 |
|
113 |
# Load model
|
114 |
model = AutoAWQForCausalLM.from_quantized(model_name_or_path, fuse_layers=True,
|
115 |
-
trust_remote_code=False, safetensors=True)
|
116 |
-
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=False)
|
117 |
|
118 |
prompt = "Tell me about AI"
|
119 |
prompt_template=f'''{prompt}
|
@@ -154,7 +155,8 @@ pipe = pipeline(
|
|
154 |
temperature=0.7,
|
155 |
top_p=0.95,
|
156 |
top_k=40,
|
157 |
-
repetition_penalty=1.1
|
|
|
158 |
)
|
159 |
|
160 |
print(pipe(prompt_template)[0]['generated_text'])
|
|
|
107 |
```python
|
108 |
from awq import AutoAWQForCausalLM
|
109 |
from transformers import AutoTokenizer
|
110 |
+
import torch
|
111 |
|
112 |
model_name_or_path = "TheBloke/Mistral-7B-v0.1-AWQ"
|
113 |
|
114 |
# Load model
|
115 |
model = AutoAWQForCausalLM.from_quantized(model_name_or_path, fuse_layers=True,
|
116 |
+
trust_remote_code=False, safetensors=True, device=device)
|
117 |
+
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=False, device=device)
|
118 |
|
119 |
prompt = "Tell me about AI"
|
120 |
prompt_template=f'''{prompt}
|
|
|
155 |
temperature=0.7,
|
156 |
top_p=0.95,
|
157 |
top_k=40,
|
158 |
+
repetition_penalty=1.1,
|
159 |
+
device=device,
|
160 |
)
|
161 |
|
162 |
print(pipe(prompt_template)[0]['generated_text'])
|