|
--- |
|
license: mit |
|
language: |
|
- zh |
|
--- |
|
|
|
|
|
--- |
|
基于baichuan-inc/Baichuan-13B-Chat 做的GPTQ的量化,可直接加载,占用GPU约12G左右,用起来效果不错 |
|
--- |
|
|
|
|
|
调用代码: |
|
|
|
---------------- |
|
import torch |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
from transformers.generation.utils import GenerationConfig |
|
|
|
model_dir = 'yinfupai/Baichuan-13B-Chat-GPTQ' |
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_dir,trust_remote_code=True) |
|
model = AutoModelForCausalLM.from_pretrained(model_dir,device_map="auto",torch_dtype=torch.float16,trust_remote_code=True) |
|
|
|
model.generation_config = GenerationConfig.from_pretrained(model_dir) |
|
model.eval() |
|
|
|
messages = [] |
|
|
|
#按baichuan要求的格式 |
|
messages.append({"role": "user", "content": "列举一下先天八卦的卦象"}) |
|
|
|
response = model.chat(tokenizer, messages) |
|
|
|
print(response) |
|
|
|
----------------------- |
|
|
|
请注意模型的商用授权,请遵照baichuan-inc/Baichuan-13B-Chat的页面中的声明 |