Add files using large-upload tool
Browse files- all_results.json +8 -0
- checkpoint-2000/global_step2000/zero_pp_rank_20_mp_rank_00_model_states.pt +3 -0
- checkpoint-2000/rng_state_10.pth +3 -0
- checkpoint-2000/rng_state_12.pth +3 -0
- checkpoint-2000/rng_state_14.pth +3 -0
- checkpoint-2000/rng_state_15.pth +3 -0
- checkpoint-2000/rng_state_18.pth +3 -0
- checkpoint-2000/rng_state_19.pth +3 -0
- checkpoint-2000/rng_state_26.pth +3 -0
- checkpoint-2000/rng_state_27.pth +3 -0
- checkpoint-2000/rng_state_3.pth +3 -0
- checkpoint-2000/rng_state_38.pth +3 -0
- checkpoint-2000/rng_state_4.pth +3 -0
- checkpoint-2000/rng_state_46.pth +3 -0
- checkpoint-2000/rng_state_48.pth +3 -0
- checkpoint-2000/rng_state_51.pth +3 -0
- checkpoint-2000/rng_state_54.pth +3 -0
- checkpoint-2000/rng_state_55.pth +3 -0
- checkpoint-2000/rng_state_59.pth +3 -0
- checkpoint-2000/rng_state_60.pth +3 -0
- checkpoint-2000/rng_state_62.pth +3 -0
- checkpoint-2000/rng_state_7.pth +3 -0
- special_tokens_map.json +30 -0
- tokenizer_config.json +43 -0
- train_results.json +8 -0
all_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 2.999650716032134,
|
3 |
+
"total_flos": 8645827988029440.0,
|
4 |
+
"train_loss": 0.35461137654308095,
|
5 |
+
"train_runtime": 104325.7524,
|
6 |
+
"train_samples_per_second": 63.226,
|
7 |
+
"train_steps_per_second": 0.123
|
8 |
+
}
|
checkpoint-2000/global_step2000/zero_pp_rank_20_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74222b34b48fd4459bb0fc32e6b2d5f94969bd373d27483222b1de2b1409e691
|
3 |
+
size 151116
|
checkpoint-2000/rng_state_10.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d4c67c1d0ec6b889370f634ff29f584f829acaeaf69196a69304fdba936f9d7
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_12.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a9d1c7f561c5512193e8053ecf371e4fef647a8503481f2dad76332ddbe164fd
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_14.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d984186a3bc06445d6c077fcbf64da96d47d3ed704e68f0212045efb4b91dbb
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_15.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b3945ba3535361fa13a8460a70ebbd44ece2514a1c9dcb76abfe6f54a775c60
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_18.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e167524b369e20a3a17ee7f6124322da680ccf2ae553b4adc0e7462d2dfba82b
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_19.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a0915a78eba7c1d2768ecd03818e3c01afd7f3168a3318b1212413bd2c62f7f1
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_26.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6b29094f6dcc92be7d58da39df551c573728836c5aec67bef90790810c0d3882
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_27.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b64ff98dfb30bbfa08760a0a79d9be7407917ff0e32f22b03d6528bf7e292e44
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac55dba0b79d5fa4699d239da2f966d52040d576d31234ac8d4632e6956481bc
|
3 |
+
size 15984
|
checkpoint-2000/rng_state_38.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eae0a9bd1151c39106890b996fd767e0569650a669126b6513dbbf0513900d10
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af2d0c015100768ffa23faf3b6c2d54ea89eb045603e30e55cd211e06ff34972
|
3 |
+
size 15984
|
checkpoint-2000/rng_state_46.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1d075e9c27d02dbb9e2a6dddb02dd39b943708e5e4bf03d03996614e5265fb38
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_48.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1f0ab6144c506e09c95b742f41deb186593d88524f1c50e69d2856c4ba23697f
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_51.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6abf2565abbb33088dd3547734e27c1d8b840d31677119e167880f3f38cf78e9
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_54.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:326ed1e97b4c78dc80c69932991f5732cb347f8847a4b9ff237df98dd72764f3
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_55.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b723bb3b70aee1abb8abb975a19dcdaab9ecd3443d374918a5789711f3045db
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_59.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80f5d05d5236f8c5cd58c93f8e304f862f6c665e75f39a0f6e24734fc561ac6c
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_60.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9cbfdc077a0836fc06ab3fd73bbabbe35c4db41f4318e20835410444fe1702b2
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_62.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df375bc22957348c6eaebfa0b8cd9dc556deb98c8a760ba820df320268db26a1
|
3 |
+
size 15997
|
checkpoint-2000/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:38bc23a138cc800b22881742c0f3f9a71731a9a7111c6058a0077e6274d21773
|
3 |
+
size 15984
|
special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "</s>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"unk_token": {
|
24 |
+
"content": "<unk>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
tokenizer_config.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"0": {
|
6 |
+
"content": "<unk>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"1": {
|
14 |
+
"content": "<s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"2": {
|
22 |
+
"content": "</s>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
}
|
29 |
+
},
|
30 |
+
"additional_special_tokens": [],
|
31 |
+
"bos_token": "<s>",
|
32 |
+
"chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
|
33 |
+
"clean_up_tokenization_spaces": false,
|
34 |
+
"eos_token": "</s>",
|
35 |
+
"legacy": true,
|
36 |
+
"model_max_length": 1000000000000000019884624838656,
|
37 |
+
"pad_token": null,
|
38 |
+
"sp_model_kwargs": {},
|
39 |
+
"spaces_between_special_tokens": false,
|
40 |
+
"tokenizer_class": "LlamaTokenizer",
|
41 |
+
"unk_token": "<unk>",
|
42 |
+
"use_default_system_prompt": false
|
43 |
+
}
|
train_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 2.999650716032134,
|
3 |
+
"total_flos": 8645827988029440.0,
|
4 |
+
"train_loss": 0.35461137654308095,
|
5 |
+
"train_runtime": 104325.7524,
|
6 |
+
"train_samples_per_second": 63.226,
|
7 |
+
"train_steps_per_second": 0.123
|
8 |
+
}
|