lrl-modelcloud commited on
Commit
dd045e7
1 Parent(s): 64e6765

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -21,7 +21,7 @@ You can use [GPTQModel](https://github.com/ModelCloud/GPTQModel) for model infer
21
  import torch
22
  from transformers import AutoTokenizer, GenerationConfig
23
  from gptqmodel import GPTQModel
24
- model_name = "/monster/data/model/DeepSeek-V2-Chat-0628/gptq_gptq_4_0719/"
25
  tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
26
  # `max_memory` should be set based on your devices
27
  max_memory = {i: "75GB" for i in range(2)}
 
21
  import torch
22
  from transformers import AutoTokenizer, GenerationConfig
23
  from gptqmodel import GPTQModel
24
+ model_name = "ModelCloud/DeepSeek-V2-Chat-0628-gptq-4bit"
25
  tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
26
  # `max_memory` should be set based on your devices
27
  max_memory = {i: "75GB" for i in range(2)}