Add README.md
Browse files
README.md
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language: nl
|
3 |
+
widget:
|
4 |
+
- text: "Een zalig kerstfeest en "
|
5 |
+
- text: "Na een lange reeks vertragingen zal eind volgende week de James Webb Space Telescope (JWST) de aarde verlaten. Met een vergulde spiegel van "
|
6 |
+
tags:
|
7 |
+
- adaption
|
8 |
+
- recycled
|
9 |
+
- gpt2-medium
|
10 |
+
- gpt2
|
11 |
+
pipeline_tag: text-generation
|
12 |
+
datasets:
|
13 |
+
- yhavinga/mc4_nl_cleaned
|
14 |
+
---
|
15 |
+
# GPT2-Medium pre-trained on cleaned Dutch mC4 🇳🇱
|
16 |
+
|
17 |
+
Training details:
|
18 |
+
|
19 |
+
* trained for 120k steps (24 dec 2021)
|
20 |
+
* block size: 512
|
21 |
+
* optimizer: adam, lr 8e-4, beta1 0.9, beta2 0.98
|
22 |
+
* warmup 5000 steps
|
23 |
+
* weight decay 0.01
|
24 |
+
|
25 |
+
Work in progress. Dec 2021.
|
26 |
+
|
27 |
+
* Many thanks to the [Google TPU Research Cloud](https://sites.research.google/trc/about/) for providing access to a TPU cluster!
|
28 |
+
* Thanks to @gsarti for creating the [t5-flax-gcp
|
29 |
+
repository](https://github.com/gsarti/t5-flax-gcp).
|
30 |
+
* Also thanks to the creators of [gpt2-medium-persian](https://huggingface.co/flax-community/gpt2-medium-persian) and
|
31 |
+
[gpt2-medium-indonesian](https://huggingface.co/flax-community/gpt2-medium-persian)
|
32 |
+
for sharing their training scripts!
|