kmfoda commited on
Commit
a1323d8
·
verified ·
1 Parent(s): c5237a9

Epoch 0. Batch Size 0. Peers 0.

Browse files
config.json CHANGED
@@ -18,8 +18,8 @@
18
  ],
19
  "attn_pdrop": 0.1,
20
  "auto_map": {
21
- "AutoConfig": "configuration_gpt_optimized.GPTOptimConfig",
22
- "AutoModelForCausalLM": "modeling_gpt_optimized.GPTOptim"
23
  },
24
  "block_size": 1024,
25
  "bos_token_id": 50256,
 
18
  ],
19
  "attn_pdrop": 0.1,
20
  "auto_map": {
21
+ "AutoConfig": "distributed/optimized-gpt2-2b-vtestnet-v1--configuration_gpt_optimized.GPTOptimConfig",
22
+ "AutoModelForCausalLM": "distributed/optimized-gpt2-2b-vtestnet-v1--modeling_gpt_optimized.GPTOptim"
23
  },
24
  "block_size": 1024,
25
  "bos_token_id": 50256,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:53d8d1ae1200f86b8201a0e7790020d9a420bb6b137709ed38e6487d78273840
3
- size 4975499944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b36dc8257a6c07c5f196b635e30f623412515deb5cf305bc8396a6aa8b431f53
3
+ size 4975530824
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbce586ebbf8b9a2e9e141a04035a0c4881f6ca5c7b6f68de2b3bb802230f8a0
3
  size 3075942944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a17503fd56303509c8465d2e176d23b1c2f4c1ac208dc4e34574eb9495bdeb
3
  size 3075942944
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 8051383872
4
  },
5
  "weight_map": {
6
  "model.transformer.h.0.attn.c_attn.bias": "model-00001-of-00002.safetensors",
@@ -533,7 +533,11 @@
533
  "model.transformer.h.9.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
534
  "model.transformer.ln_f.bias": "model-00002-of-00002.safetensors",
535
  "model.transformer.ln_f.weight": "model-00002-of-00002.safetensors",
 
 
536
  "model.transformer.wpe.weight": "model-00001-of-00002.safetensors",
 
 
537
  "model.transformer.wte.weight": "model-00001-of-00002.safetensors"
538
  }
539
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 8051414336
4
  },
5
  "weight_map": {
6
  "model.transformer.h.0.attn.c_attn.bias": "model-00001-of-00002.safetensors",
 
533
  "model.transformer.h.9.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
534
  "model.transformer.ln_f.bias": "model-00002-of-00002.safetensors",
535
  "model.transformer.ln_f.weight": "model-00002-of-00002.safetensors",
536
+ "model.transformer.wpe.norm.bias": "model-00001-of-00002.safetensors",
537
+ "model.transformer.wpe.norm.weight": "model-00001-of-00002.safetensors",
538
  "model.transformer.wpe.weight": "model-00001-of-00002.safetensors",
539
+ "model.transformer.wte.norm.bias": "model-00001-of-00002.safetensors",
540
+ "model.transformer.wte.norm.weight": "model-00001-of-00002.safetensors",
541
  "model.transformer.wte.weight": "model-00001-of-00002.safetensors"
542
  }
543
  }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c1c3e9b7f72886b3928354c660c77cc64217de83bab4aece73321c1d91dfecc
3
+ size 2536