Saving weights and log at step 150000
Browse files
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"activation_function": "gelu_new",
|
4 |
"architectures": [
|
5 |
"GPTNeoForCausalLM"
|
@@ -46,6 +46,7 @@
|
|
46 |
"summary_proj_to_labels": true,
|
47 |
"summary_type": "cls_index",
|
48 |
"summary_use_proj": true,
|
|
|
49 |
"transformers_version": "4.13.0",
|
50 |
"use_cache": true,
|
51 |
"vocab_size": 50257,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": ".",
|
3 |
"activation_function": "gelu_new",
|
4 |
"architectures": [
|
5 |
"GPTNeoForCausalLM"
|
|
|
46 |
"summary_proj_to_labels": true,
|
47 |
"summary_type": "cls_index",
|
48 |
"summary_use_proj": true,
|
49 |
+
"torch_dtype": "float32",
|
50 |
"transformers_version": "4.13.0",
|
51 |
"use_cache": true,
|
52 |
"vocab_size": 50257,
|
flax_model.msgpack
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5301955d8ac14e97de94a79bdf99064e9359af712d0c1c56fb35d0819fdfd676
|
3 |
+
size 500800120
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae07c9cf8bfdbf25cf4bbbdc712c1bbafc9c53071d88725c0f1e42f15325a193
|
3 |
+
size 551192465
|
runs/events.out.tfevents.1641165068.t1v-n-d7290f31-w-0.30243.0.v2
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b744f69290d42d941df38ffe35b17fab3cc794951ace974cadb2c31f25b38443
|
3 |
+
size 22570467
|