shanchen commited on
Commit
b422017
·
verified ·
1 Parent(s): 475bf6c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -16
README.md CHANGED
@@ -42,22 +42,31 @@ dtype: bfloat16
42
  ```python
43
  !pip install -qU transformers accelerate
44
 
45
- from transformers import AutoTokenizer
46
- import transformers
47
- import torch
48
-
49
- model = "shanchen/llama3-8B-slerp-biomed-chat-chinese"
50
- messages = [{"role": "user", "content": "What is a large language model?"}]
51
-
52
- tokenizer = AutoTokenizer.from_pretrained(model)
53
- prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
54
- pipeline = transformers.pipeline(
55
- "text-generation",
56
- model=model,
57
- torch_dtype=torch.float16,
58
- device_map="auto",
 
 
 
 
 
 
 
 
 
59
  )
 
 
60
 
61
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
62
- print(outputs[0]["generated_text"])
63
  ```
 
42
  ```python
43
  !pip install -qU transformers accelerate
44
 
45
+ from transformers import AutoTokenizer, AutoModelForCausalLM
46
+
47
+ model_id = "shenzhi-wang/Llama3-8B-Chinese-Chat"
48
+
49
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
50
+ model = AutoModelForCausalLM.from_pretrained(
51
+ model_id, torch_dtype="auto", device_map="auto"
52
+ )
53
+
54
+ messages = [
55
+ {"role": "user", "content": "Tell me about yourself"},
56
+ ]
57
+
58
+ input_ids = tokenizer.apply_chat_template(
59
+ messages, add_generation_prompt=True, return_tensors="pt"
60
+ ).to(model.device)
61
+
62
+ outputs = model.generate(
63
+ input_ids,
64
+ max_new_tokens=8192,
65
+ do_sample=True,
66
+ temperature=0.6,
67
+ top_p=0.9,
68
  )
69
+ response = outputs[0][input_ids.shape[-1]:]
70
+ print(tokenizer.decode(response, skip_special_tokens=True))
71
 
 
 
72
  ```