Zlovoblachko commited on
Commit
7c86386
1 Parent(s): a8c9046

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: google-t5/t5-small
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -16,13 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # my_awesome_dailymail_model
18
 
19
- This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 2.5637
22
- - Rouge1: 0.1421
23
- - Rouge2: 0.0499
24
- - Rougel: 0.1181
25
- - Rougelsum: 0.1176
26
  - Gen Len: 19.0
27
 
28
  ## Model description
@@ -43,8 +43,8 @@ More information needed
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 2e-05
46
- - train_batch_size: 16
47
- - eval_batch_size: 16
48
  - seed: 42
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
@@ -55,10 +55,10 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
57
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
58
- | No log | 1.0 | 62 | 2.8539 | 0.1301 | 0.0376 | 0.1088 | 0.1085 | 19.0 |
59
- | No log | 2.0 | 124 | 2.6380 | 0.1388 | 0.0458 | 0.1155 | 0.1152 | 19.0 |
60
- | No log | 3.0 | 186 | 2.5802 | 0.1418 | 0.0502 | 0.1173 | 0.1172 | 19.0 |
61
- | No log | 4.0 | 248 | 2.5637 | 0.1421 | 0.0499 | 0.1181 | 0.1176 | 19.0 |
62
 
63
 
64
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: google/long-t5-tglobal-base
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
16
 
17
  # my_awesome_dailymail_model
18
 
19
+ This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.7605
22
+ - Rouge1: 0.2044
23
+ - Rouge2: 0.1072
24
+ - Rougel: 0.1751
25
+ - Rougelsum: 0.1748
26
  - Gen Len: 19.0
27
 
28
  ## Model description
 
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 2e-05
46
+ - train_batch_size: 8
47
+ - eval_batch_size: 8
48
  - seed: 42
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
57
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
58
+ | No log | 1.0 | 124 | 1.8435 | 0.1904 | 0.0914 | 0.1603 | 0.1602 | 19.0 |
59
+ | No log | 2.0 | 248 | 1.7953 | 0.2033 | 0.1046 | 0.1733 | 0.1732 | 19.0 |
60
+ | No log | 3.0 | 372 | 1.7681 | 0.2042 | 0.1061 | 0.1753 | 0.175 | 19.0 |
61
+ | No log | 4.0 | 496 | 1.7605 | 0.2044 | 0.1072 | 0.1751 | 0.1748 | 19.0 |
62
 
63
 
64
  ### Framework versions
runs/Oct06_01-56-59_b00e0ecb01fe/events.out.tfevents.1728179820.b00e0ecb01fe.30.3 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c61bb2d03f7ffde55285bb73b0e116fc79d2a6a6d1396d4182bf1d913a80208
3
- size 6915
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42d3b2e5cec589bf26b2a6386b5f56000eecc454a334f6be805d3139a3412c90
3
+ size 7794