duynhatran commited on
Commit
a68388e
1 Parent(s): e3d5509

End of training

Browse files
Files changed (4) hide show
  1. README.md +10 -12
  2. model.safetensors +1 -1
  3. tokenizer.json +2 -2
  4. tokenizer_config.json +7 -0
README.md CHANGED
@@ -5,7 +5,6 @@ tags:
5
  - generated_from_trainer
6
  metrics:
7
  - accuracy
8
- - f1
9
  model-index:
10
  - name: deberta_textclassification
11
  results: []
@@ -18,9 +17,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [microsoft/deberta-base](https://huggingface.co/microsoft/deberta-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.3259
22
- - Accuracy: 0.8781
23
- - F1: 0.9135
24
 
25
  ## Model description
26
 
@@ -39,7 +37,7 @@ More information needed
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
- - learning_rate: 5e-06
43
  - train_batch_size: 16
44
  - eval_batch_size: 16
45
  - seed: 42
@@ -49,13 +47,13 @@ The following hyperparameters were used during training:
49
 
50
  ### Training results
51
 
52
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
53
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
54
- | No log | 1.0 | 180 | 0.3881 | 0.8406 | 0.8889 |
55
- | No log | 2.0 | 360 | 0.3333 | 0.8688 | 0.9079 |
56
- | 0.4361 | 3.0 | 540 | 0.3758 | 0.8375 | 0.8802 |
57
- | 0.4361 | 4.0 | 720 | 0.3378 | 0.8781 | 0.9147 |
58
- | 0.4361 | 5.0 | 900 | 0.3259 | 0.8781 | 0.9135 |
59
 
60
 
61
  ### Framework versions
 
5
  - generated_from_trainer
6
  metrics:
7
  - accuracy
 
8
  model-index:
9
  - name: deberta_textclassification
10
  results: []
 
17
 
18
  This model is a fine-tuned version of [microsoft/deberta-base](https://huggingface.co/microsoft/deberta-base) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.4450
21
+ - Accuracy: 0.8969
 
22
 
23
  ## Model description
24
 
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+ - learning_rate: 2e-05
41
  - train_batch_size: 16
42
  - eval_batch_size: 16
43
  - seed: 42
 
47
 
48
  ### Training results
49
 
50
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | No log | 1.0 | 180 | 0.3192 | 0.8656 |
53
+ | No log | 2.0 | 360 | 0.4034 | 0.825 |
54
+ | 0.3441 | 3.0 | 540 | 0.3285 | 0.8938 |
55
+ | 0.3441 | 4.0 | 720 | 0.3576 | 0.9 |
56
+ | 0.3441 | 5.0 | 900 | 0.4450 | 0.8969 |
57
 
58
 
59
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b1151b8c7cc165ff38997696be11bcbc5c3feeba309bad8e925db470dd376c2
3
  size 556799560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10a5e121e583d87a9470061c010030ede131f77b6e1147c070a45fde39549c58
3
  size 556799560
tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 195,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 195
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 512
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
tokenizer_config.json CHANGED
@@ -50,10 +50,17 @@
50
  "eos_token": "[SEP]",
51
  "errors": "replace",
52
  "mask_token": "[MASK]",
 
53
  "model_max_length": 1000000000000000019884624838656,
 
54
  "pad_token": "[PAD]",
 
 
55
  "sep_token": "[SEP]",
 
56
  "tokenizer_class": "DebertaTokenizer",
 
 
57
  "unk_token": "[UNK]",
58
  "vocab_type": "gpt2"
59
  }
 
50
  "eos_token": "[SEP]",
51
  "errors": "replace",
52
  "mask_token": "[MASK]",
53
+ "max_length": 195,
54
  "model_max_length": 1000000000000000019884624838656,
55
+ "pad_to_multiple_of": null,
56
  "pad_token": "[PAD]",
57
+ "pad_token_type_id": 0,
58
+ "padding_side": "right",
59
  "sep_token": "[SEP]",
60
+ "stride": 0,
61
  "tokenizer_class": "DebertaTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
  "unk_token": "[UNK]",
65
  "vocab_type": "gpt2"
66
  }