Update README.md
Browse files
README.md
CHANGED
@@ -24,6 +24,10 @@ widget:
|
|
24 |
Llama-3-8B-flashback-v1 is a continuation of the pretraining process for the base meta-llama/Meta-Llama-3-8B model, utilizing 2 251 233 forum threads from the Swedish website https://www.flashback.org/. Which is rougly 40GB of text.
|
25 |
It is a full finetune for one epoch.
|
26 |
|
|
|
|
|
|
|
|
|
27 |
## How to use:
|
28 |
```python
|
29 |
import torch
|
|
|
24 |
Llama-3-8B-flashback-v1 is a continuation of the pretraining process for the base meta-llama/Meta-Llama-3-8B model, utilizing 2 251 233 forum threads from the Swedish website https://www.flashback.org/. Which is rougly 40GB of text.
|
25 |
It is a full finetune for one epoch.
|
26 |
|
27 |
+
```html
|
28 |
+
* tokens: 8_271_170_155
|
29 |
+
* checkpoint 4/5
|
30 |
+
```
|
31 |
## How to use:
|
32 |
```python
|
33 |
import torch
|