lrl-modelcloud
commited on
Commit
•
dd045e7
1
Parent(s):
64e6765
Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ You can use [GPTQModel](https://github.com/ModelCloud/GPTQModel) for model infer
|
|
21 |
import torch
|
22 |
from transformers import AutoTokenizer, GenerationConfig
|
23 |
from gptqmodel import GPTQModel
|
24 |
-
model_name = "/
|
25 |
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
26 |
# `max_memory` should be set based on your devices
|
27 |
max_memory = {i: "75GB" for i in range(2)}
|
|
|
21 |
import torch
|
22 |
from transformers import AutoTokenizer, GenerationConfig
|
23 |
from gptqmodel import GPTQModel
|
24 |
+
model_name = "ModelCloud/DeepSeek-V2-Chat-0628-gptq-4bit"
|
25 |
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
26 |
# `max_memory` should be set based on your devices
|
27 |
max_memory = {i: "75GB" for i in range(2)}
|