Commit
·
e7cfa80
1
Parent(s):
0463357
Update README.md
Browse files
README.md
CHANGED
@@ -43,6 +43,5 @@ model = AutoModelForMaskedLM.from_pretrained("jannikskytt/MeDa-Bert")
|
|
43 |
publisher = "University of Tartu Library",
|
44 |
url = "https://aclanthology.org/2023.nodalida-1.31",
|
45 |
pages = "301--307",
|
46 |
-
abstract = "This paper introduces a medical Danish BERT-based language model (MeDa-BERT) and medical Danish word embeddings. The word embeddings and MeDa-BERT were pretrained on a new medical Danish corpus consisting of 133M tokens from medical Danish books and text from the internet. The models showed improved performance over general-domain models on medical Danish classification tasks. The medical word embeddings and MeDa-BERT are publicly available.",
|
47 |
}
|
48 |
```
|
|
|
43 |
publisher = "University of Tartu Library",
|
44 |
url = "https://aclanthology.org/2023.nodalida-1.31",
|
45 |
pages = "301--307",
|
|
|
46 |
}
|
47 |
```
|