Update README.md
Browse files
README.md
CHANGED
@@ -40,8 +40,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
40 |
### Overview
|
41 |
|
42 |
- **Model Name**: EUBERT
|
43 |
-
- **Model Version**: 1.
|
44 |
-
- **Date of Release**:
|
45 |
- **Model Architecture**: BERT (Bidirectional Encoder Representations from Transformers)
|
46 |
- **Training Data**: Documents registered by the European Publications Office
|
47 |
- **Model Use Case**: Text Classification, Question Answering, Language Understanding
|
@@ -89,7 +89,7 @@ EUBERT is a pretrained BERT model that leverages a substantial corpus of documen
|
|
89 |
|
90 |
## Training procedure
|
91 |
|
92 |
-
Dedicated
|
93 |
|
94 |
### Training hyperparameters
|
95 |
|
@@ -100,7 +100,7 @@ The following hyperparameters were used during training:
|
|
100 |
- seed: 42
|
101 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
102 |
- lr_scheduler_type: linear
|
103 |
-
- num_epochs: 1
|
104 |
|
105 |
### Training results
|
106 |
|
|
|
40 |
### Overview
|
41 |
|
42 |
- **Model Name**: EUBERT
|
43 |
+
- **Model Version**: 1.1
|
44 |
+
- **Date of Release**: 16 October 2023
|
45 |
- **Model Architecture**: BERT (Bidirectional Encoder Representations from Transformers)
|
46 |
- **Training Data**: Documents registered by the European Publications Office
|
47 |
- **Model Use Case**: Text Classification, Question Answering, Language Understanding
|
|
|
89 |
|
90 |
## Training procedure
|
91 |
|
92 |
+
Dedicated Word Piece tokenizer vocabulary size 2**16,
|
93 |
|
94 |
### Training hyperparameters
|
95 |
|
|
|
100 |
- seed: 42
|
101 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
102 |
- lr_scheduler_type: linear
|
103 |
+
- num_epochs: 1.85
|
104 |
|
105 |
### Training results
|
106 |
|