Safick-HF commited on
Commit
c713df3
verified
1 Parent(s): 6aa443b

End of training

Browse files
README.md CHANGED
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [deepset/gelectra-large-germanquad](https://huggingface.co/deepset/gelectra-large-germanquad) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 3.3874
19
 
20
  ## Model description
21
 
@@ -36,7 +36,7 @@ More information needed
36
  The following hyperparameters were used during training:
37
  - learning_rate: 2e-05
38
  - train_batch_size: 5
39
- - eval_batch_size: 5
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
@@ -46,21 +46,21 @@ The following hyperparameters were used during training:
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | 2.6324 | 1.0 | 3 | 3.5638 |
50
- | 0.3607 | 2.0 | 6 | 3.0847 |
51
- | 1.7247 | 3.0 | 9 | 2.8964 |
52
- | 0.0183 | 4.0 | 12 | 2.7518 |
53
- | 0.1101 | 5.0 | 15 | 2.8213 |
54
- | 0.0063 | 6.0 | 18 | 2.9454 |
55
- | 0.0121 | 7.0 | 21 | 3.1119 |
56
- | 0.0036 | 8.0 | 24 | 3.2701 |
57
- | 0.0139 | 9.0 | 27 | 3.3557 |
58
- | 0.0051 | 10.0 | 30 | 3.3874 |
59
 
60
 
61
  ### Framework versions
62
 
63
- - Transformers 4.38.2
64
- - Pytorch 2.2.1+cu121
65
  - Datasets 2.18.0
66
  - Tokenizers 0.15.2
 
15
 
16
  This model is a fine-tuned version of [deepset/gelectra-large-germanquad](https://huggingface.co/deepset/gelectra-large-germanquad) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.0331
19
 
20
  ## Model description
21
 
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 2e-05
38
  - train_batch_size: 5
39
+ - eval_batch_size: 10
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
 
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
+ | 0.1008 | 1.0 | 3 | 0.3691 |
50
+ | 0.177 | 2.0 | 6 | 0.2874 |
51
+ | 0.0763 | 3.0 | 9 | 0.2115 |
52
+ | 0.0065 | 4.0 | 12 | 0.1496 |
53
+ | 0.0075 | 5.0 | 15 | 0.0874 |
54
+ | 0.0231 | 6.0 | 18 | 0.0727 |
55
+ | 0.0059 | 7.0 | 21 | 0.0512 |
56
+ | 0.0302 | 8.0 | 24 | 0.0407 |
57
+ | 0.0017 | 9.0 | 27 | 0.0347 |
58
+ | 0.0015 | 10.0 | 30 | 0.0331 |
59
 
60
 
61
  ### Framework versions
62
 
63
+ - Transformers 4.39.3
64
+ - Pytorch 2.1.2+cpu
65
  - Datasets 2.18.0
66
  - Tokenizers 0.15.2
config.json CHANGED
@@ -25,7 +25,7 @@
25
  "summary_type": "first",
26
  "summary_use_proj": false,
27
  "torch_dtype": "float32",
28
- "transformers_version": "4.38.2",
29
  "type_vocab_size": 2,
30
  "use_cache": true,
31
  "vocab_size": 31102
 
25
  "summary_type": "first",
26
  "summary_use_proj": false,
27
  "torch_dtype": "float32",
28
+ "transformers_version": "4.39.3",
29
  "type_vocab_size": 2,
30
  "use_cache": true,
31
  "vocab_size": 31102
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36ae8d195190f2124928f1a26ef8d686e923a9887c50b1e8ae88ffce52540966
3
  size 1338801016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10b633ae7d4e3e01d6a9c61ee0a4b0e2aa6f37517f436444bcbebfc9e4741888
3
  size 1338801016
runs/Apr23_11-43-08_f3c944805091/events.out.tfevents.1713872589.f3c944805091.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82d3fafdb889db1712e7b16855f66d99b627cf5d08e30ef2dc8256ad7993bb56
3
+ size 14043
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1428c101e593f0915b98e8b5aa7b55494af2a37dce6f6126bc8e79222008b8b
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3155dddd0c2d78aff3d706df4088f8df0e103e710c40e3524124b3a41350cb0c
3
  size 4920