nurcan commited on
Commit
10e27e6
·
1 Parent(s): c6ebf2b

Training in progress epoch 0

Browse files
README.md CHANGED
@@ -4,20 +4,20 @@ base_model: gpt2
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
7
- - name: turkishReviews-ds-mini
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
12
  probably proofread and complete it, then remove this comment. -->
13
 
14
- # turkishReviews-ds-mini
15
 
16
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 9.1926
19
- - Validation Loss: 9.2656
20
- - Epoch: 2
21
 
22
  ## Model description
23
 
@@ -43,9 +43,7 @@ The following hyperparameters were used during training:
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
- | 10.3294 | 10.0078 | 0 |
47
- | 9.6789 | 9.6480 | 1 |
48
- | 9.1926 | 9.2656 | 2 |
49
 
50
 
51
  ### Framework versions
 
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
7
+ - name: nurcan/turkishReviews-ds-mini
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
12
  probably proofread and complete it, then remove this comment. -->
13
 
14
+ # nurcan/turkishReviews-ds-mini
15
 
16
  This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 10.3191
19
+ - Validation Loss: 9.9828
20
+ - Epoch: 0
21
 
22
  ## Model description
23
 
 
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
+ | 10.3191 | 9.9828 | 0 |
 
 
47
 
48
 
49
  ### Framework versions
config.json CHANGED
@@ -34,5 +34,5 @@
34
  },
35
  "transformers_version": "4.35.2",
36
  "use_cache": true,
37
- "vocab_size": 45068
38
  }
 
34
  },
35
  "transformers_version": "4.35.2",
36
  "use_cache": true,
37
+ "vocab_size": 44208
38
  }
special_tokens_map.json CHANGED
@@ -13,6 +13,7 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
16
  "unk_token": {
17
  "content": "<|endoftext|>",
18
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<|endoftext|>",
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a60e4366e13e0a09e53d48aa8673d34a2278c6fabde5451ded4d00e048004f65
3
- size 481994832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a49ea92dedd0aaf9584a88c66b8cf5ad9b3a130f4c0fdbda2cecf6b9125c410
3
+ size 479352912
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 40,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
tokenizer_config.json CHANGED
@@ -16,7 +16,7 @@
16
  "eos_token": "<|endoftext|>",
17
  "errors": "replace",
18
  "model_max_length": 1024,
19
- "pad_token": null,
20
  "tokenizer_class": "GPT2Tokenizer",
21
  "unk_token": "<|endoftext|>"
22
  }
 
16
  "eos_token": "<|endoftext|>",
17
  "errors": "replace",
18
  "model_max_length": 1024,
19
+ "pad_token": "<|endoftext|>",
20
  "tokenizer_class": "GPT2Tokenizer",
21
  "unk_token": "<|endoftext|>"
22
  }