itsanastasiaminina commited on
Commit
adb8d7b
·
verified ·
1 Parent(s): 03cb146

End of training

Browse files
Files changed (2) hide show
  1. README.md +12 -10
  2. model.safetensors +1 -1
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
- license: apache-2.0
4
- base_model: distilbert/distilbert-base-uncased
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -16,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # students_scores_model
18
 
19
- This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.9041
22
- - F1: 0.6008
23
 
24
  ## Model description
25
 
@@ -39,19 +39,21 @@ More information needed
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 2e-05
42
- - train_batch_size: 16
43
- - eval_batch_size: 16
44
  - seed: 42
45
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
- - num_epochs: 2
 
48
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | F1 |
52
  |:-------------:|:-----:|:----:|:---------------:|:------:|
53
- | 1.1043 | 1.0 | 563 | 0.9514 | 0.5699 |
54
- | 0.8911 | 2.0 | 1126 | 0.9041 | 0.6008 |
 
55
 
56
 
57
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
+ license: mit
4
+ base_model: roberta-base
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
16
 
17
  # students_scores_model
18
 
19
+ This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.8747
22
+ - F1: 0.6229
23
 
24
  ## Model description
25
 
 
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 2e-05
42
+ - train_batch_size: 32
43
+ - eval_batch_size: 32
44
  - seed: 42
45
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
+ - num_epochs: 3
48
+ - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | F1 |
53
  |:-------------:|:-----:|:----:|:---------------:|:------:|
54
+ | No log | 1.0 | 282 | 1.1173 | 0.4836 |
55
+ | 0.9565 | 2.0 | 564 | 0.8655 | 0.6161 |
56
+ | 0.9565 | 3.0 | 846 | 0.8747 | 0.6229 |
57
 
58
 
59
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b404dc062105255c34992d3b3edfd6c78a8188fd2f8f3c7a9bf31a22d727f1ee
3
  size 498625128
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b08f321ac19bbc684346db157b8cd46dc519fd6a1ff083dbf6e031c118443ea3
3
  size 498625128