hysts HF staff commited on
Commit
8a09ee7
ยท
1 Parent(s): c49fb1b
Files changed (2) hide show
  1. README.md +3 -1
  2. app.py +2 -2
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- title: Zephyr 7b Alpha
3
  emoji: ๐Ÿ“‰
4
  colorFrom: green
5
  colorTo: yellow
@@ -11,4 +11,6 @@ license: mit
11
  suggested-hardware: t4-small
12
  ---
13
 
 
 
14
  Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
1
  ---
2
+ title: Zephyr 7b
3
  emoji: ๐Ÿ“‰
4
  colorFrom: green
5
  colorTo: yellow
 
11
  suggested-hardware: t4-small
12
  ---
13
 
14
+ https://arxiv.org/abs/2310.16944
15
+
16
  Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
app.py CHANGED
@@ -9,7 +9,7 @@ import spaces
9
  import torch
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
11
 
12
- DESCRIPTION = "# Zephyr-7B-alpha"
13
 
14
  if not torch.cuda.is_available():
15
  DESCRIPTION += "\n<p>Running on CPU ๐Ÿฅถ This demo does not work on CPU.</p>"
@@ -19,7 +19,7 @@ DEFAULT_MAX_NEW_TOKENS = 1024
19
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
20
 
21
  if torch.cuda.is_available():
22
- model_id = "HuggingFaceH4/zephyr-7b-alpha"
23
  model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto")
24
  tokenizer = AutoTokenizer.from_pretrained(model_id)
25
 
 
9
  import torch
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
11
 
12
+ DESCRIPTION = "# Zephyr-7B beta"
13
 
14
  if not torch.cuda.is_available():
15
  DESCRIPTION += "\n<p>Running on CPU ๐Ÿฅถ This demo does not work on CPU.</p>"
 
19
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
20
 
21
  if torch.cuda.is_available():
22
+ model_id = "HuggingFaceH4/zephyr-7b-beta"
23
  model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto")
24
  tokenizer = AutoTokenizer.from_pretrained(model_id)
25