{} | |
# ModernBioBERT | |
A modern variant of BioBERT based on ModernBERT. | |
We continued the masked language modeling pre-training task for 1.000.000 steps on PubMed abstracts. | |
## Pre-Training Details | |
``` | |
Batchsize: 512 | |
Learningrate: 1e-4 | |
Warmupsteps: 500 | |
Learning Rate Scheduler: Cosine Schedule | |
Max. Sequence Length: 512 | |
Precision: bfloat16 | |
``` | |
--- | |
datasets: | |
- ncbi/pubmed | |
language: | |
- en | |
base_model: | |
- answerdotai/ModernBERT-base | |
--- |