aspctu commited on
Commit
d5d3a72
·
verified ·
1 Parent(s): bdff9c2

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. config.json +50 -0
  2. rank0.safetensors +3 -0
config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "producer": {
3
+ "name": "modelopt",
4
+ "version": "0.11.2"
5
+ },
6
+ "architecture": "MedusaForCausalLM",
7
+ "dtype": "float16",
8
+ "num_hidden_layers": 32,
9
+ "num_attention_heads": 32,
10
+ "num_key_value_heads": 8,
11
+ "hidden_size": 4096,
12
+ "norm_epsilon": 1e-05,
13
+ "vocab_size": 32000,
14
+ "max_position_embeddings": 32768,
15
+ "hidden_act": "silu",
16
+ "use_parallel_embedding": true,
17
+ "embedding_sharding_dim": 0,
18
+ "quantization": {
19
+ "quant_algo": "FP8",
20
+ "kv_cache_quant_algo": "FP8",
21
+ "exclude_modules": [
22
+ "lm_head",
23
+ "*router",
24
+ "*vocab_embedding",
25
+ "*position_embedding",
26
+ "*block_embedding",
27
+ "*medusa_heads*"
28
+ ]
29
+ },
30
+ "mapping": {
31
+ "world_size": 1,
32
+ "tp_size": 1,
33
+ "pp_size": 1
34
+ },
35
+ "head_size": 128,
36
+ "intermediate_size": 14336,
37
+ "position_embedding_type": "rope_gpt_neox",
38
+ "share_embedding_table": false,
39
+ "residual_mlp": false,
40
+ "bias": false,
41
+ "rotary_pct": 1.0,
42
+ "rank": 0,
43
+ "decoder": "llama",
44
+ "rmsnorm": true,
45
+ "lm_head_bias": false,
46
+ "rotary_base": 1000000.0,
47
+ "max_draft_len": 63,
48
+ "num_medusa_heads": 3,
49
+ "num_medusa_layers": 1
50
+ }
rank0.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f432787029bf52a0d3f6af9ffca333ac1b020a1e064093203e8fc6ebb8fb463
3
+ size 8391330768