pilotj commited on
Commit
21dd049
·
verified ·
1 Parent(s): 1b16601

pilotj/bert-base-uncased-fibe-final

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -1,26 +1,26 @@
1
  ---
2
- base_model: biggy-smiley/bert-base-uncased-fibe-v2
3
  library_name: transformers
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
- - name: bert-base-uncased-fibe-v3
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- # bert-base-uncased-fibe-v3
15
 
16
- This model is a fine-tuned version of [biggy-smiley/bert-base-uncased-fibe-v2](https://huggingface.co/biggy-smiley/bert-base-uncased-fibe-v2) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 0.4755
19
- - eval_runtime: 40.9898
20
- - eval_samples_per_second: 253.722
21
- - eval_steps_per_second: 3.977
22
- - epoch: 1.0311
23
- - step: 15500
24
 
25
  ## Model description
26
 
@@ -40,12 +40,12 @@ More information needed
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 5e-05
43
- - train_batch_size: 32
44
  - eval_batch_size: 64
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
- - num_epochs: 3
49
 
50
  ### Framework versions
51
 
 
1
  ---
 
2
  library_name: transformers
3
+ base_model: pilotj/bert-base-uncased-fibe-v3
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
+ - name: bert-base-uncased-fibe-final
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ # bert-base-uncased-fibe-final
15
 
16
+ This model is a fine-tuned version of [pilotj/bert-base-uncased-fibe-v3](https://huggingface.co/pilotj/bert-base-uncased-fibe-v3) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - eval_loss: 0.3655
19
+ - eval_runtime: 111.3303
20
+ - eval_samples_per_second: 234.896
21
+ - eval_steps_per_second: 3.674
22
+ - epoch: 1.0030
23
+ - step: 10500
24
 
25
  ## Model description
26
 
 
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 5e-05
43
+ - train_batch_size: 64
44
  - eval_batch_size: 64
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
+ - num_epochs: 2
49
 
50
  ### Framework versions
51