mdroth commited on
Commit
eb2922c
·
verified ·
1 Parent(s): 4183fcf

End of training

Browse files
README.md CHANGED
@@ -13,9 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # codeparrot-ds
15
 
16
- This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
17
- It achieves the following results on the evaluation set:
18
- - Loss: 1.6432
19
 
20
  ## Model description
21
 
@@ -48,9 +46,6 @@ The following hyperparameters were used during training:
48
 
49
  ### Training results
50
 
51
- | Training Loss | Epoch | Step | Validation Loss |
52
- |:-------------:|:-----:|:----:|:---------------:|
53
- | 2.498 | 0.94 | 5000 | 1.6432 |
54
 
55
 
56
  ### Framework versions
 
13
 
14
  # codeparrot-ds
15
 
16
+ This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
 
 
17
 
18
  ## Model description
19
 
 
46
 
47
  ### Training results
48
 
 
 
 
49
 
50
 
51
  ### Framework versions
config.json CHANGED
@@ -33,7 +33,7 @@
33
  }
34
  },
35
  "torch_dtype": "float32",
36
- "transformers_version": "4.41.2",
37
  "use_cache": true,
38
  "vocab_size": 50000
39
  }
 
33
  }
34
  },
35
  "torch_dtype": "float32",
36
+ "transformers_version": "4.35.2",
37
  "use_cache": true,
38
  "vocab_size": 50000
39
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c69df4b3e5f96e99ef24b2c204189af3462dc59989f19e682380957669de374
3
  size 496984704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f1f524b36916f08ceac4a946ea688787fd5bd401efb5b5235121348976c7690
3
  size 496984704
special_tokens_map.json CHANGED
@@ -1,24 +1,6 @@
1
  {
2
- "bos_token": {
3
- "content": "<|endoftext|>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
  "pad_token": "<|endoftext|>",
17
- "unk_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
  }
 
1
  {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
4
  "pad_token": "<|endoftext|>",
5
+ "unk_token": "<|endoftext|>"
 
 
 
 
 
 
6
  }
tokenizer.json CHANGED
@@ -45,7 +45,6 @@
45
  "end_of_word_suffix": "",
46
  "fuse_unk": false,
47
  "byte_fallback": false,
48
- "ignore_merges": false,
49
  "vocab": {
50
  "<|endoftext|>": 0,
51
  "!": 1,
 
45
  "end_of_word_suffix": "",
46
  "fuse_unk": false,
47
  "byte_fallback": false,
 
48
  "vocab": {
49
  "<|endoftext|>": 0,
50
  "!": 1,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:196aec0ebc4fa28fd55fa5a8a0cbea771c07c10e096572fba5d0fc910769232b
3
- size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e24546a63098f7a280a7dffa8737ba6baf0e9a4690b018b02959f3c581054f37
3
+ size 4536