Epoch 0. Batch Size 0. Peers 0.
Browse files- config.json +2 -2
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +1 -1
- model.safetensors.index.json +5 -1
- optimizer.pt +3 -0
config.json
CHANGED
@@ -18,8 +18,8 @@
|
|
18 |
],
|
19 |
"attn_pdrop": 0.1,
|
20 |
"auto_map": {
|
21 |
-
"AutoConfig": "configuration_gpt_optimized.GPTOptimConfig",
|
22 |
-
"AutoModelForCausalLM": "modeling_gpt_optimized.GPTOptim"
|
23 |
},
|
24 |
"block_size": 1024,
|
25 |
"bos_token_id": 50256,
|
|
|
18 |
],
|
19 |
"attn_pdrop": 0.1,
|
20 |
"auto_map": {
|
21 |
+
"AutoConfig": "distributed/optimized-gpt2-2b-vtestnet-v1--configuration_gpt_optimized.GPTOptimConfig",
|
22 |
+
"AutoModelForCausalLM": "distributed/optimized-gpt2-2b-vtestnet-v1--modeling_gpt_optimized.GPTOptim"
|
23 |
},
|
24 |
"block_size": 1024,
|
25 |
"bos_token_id": 50256,
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b36dc8257a6c07c5f196b635e30f623412515deb5cf305bc8396a6aa8b431f53
|
3 |
+
size 4975530824
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3075942944
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01a17503fd56303509c8465d2e176d23b1c2f4c1ac208dc4e34574eb9495bdeb
|
3 |
size 3075942944
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.transformer.h.0.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
@@ -533,7 +533,11 @@
|
|
533 |
"model.transformer.h.9.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
534 |
"model.transformer.ln_f.bias": "model-00002-of-00002.safetensors",
|
535 |
"model.transformer.ln_f.weight": "model-00002-of-00002.safetensors",
|
|
|
|
|
536 |
"model.transformer.wpe.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
537 |
"model.transformer.wte.weight": "model-00001-of-00002.safetensors"
|
538 |
}
|
539 |
}
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 8051414336
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.transformer.h.0.attn.c_attn.bias": "model-00001-of-00002.safetensors",
|
|
|
533 |
"model.transformer.h.9.mlp.c_proj.weight": "model-00001-of-00002.safetensors",
|
534 |
"model.transformer.ln_f.bias": "model-00002-of-00002.safetensors",
|
535 |
"model.transformer.ln_f.weight": "model-00002-of-00002.safetensors",
|
536 |
+
"model.transformer.wpe.norm.bias": "model-00001-of-00002.safetensors",
|
537 |
+
"model.transformer.wpe.norm.weight": "model-00001-of-00002.safetensors",
|
538 |
"model.transformer.wpe.weight": "model-00001-of-00002.safetensors",
|
539 |
+
"model.transformer.wte.norm.bias": "model-00001-of-00002.safetensors",
|
540 |
+
"model.transformer.wte.norm.weight": "model-00001-of-00002.safetensors",
|
541 |
"model.transformer.wte.weight": "model-00001-of-00002.safetensors"
|
542 |
}
|
543 |
}
|
optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c1c3e9b7f72886b3928354c660c77cc64217de83bab4aece73321c1d91dfecc
|
3 |
+
size 2536
|