saattrupdan's picture
Update README.md
2f99e9f verified
|
raw
history blame
1.69 kB
---
library_name: transformers
tags:
- merge
- mergekit
license: cc-by-4.0
language:
- da
- sv
- 'no'
- nb
- nn
base_model:
- mistralai/Mistral-7B-v0.1
- danish-foundation-models/munin-7b-alpha
- norallm/normistral-7b-warm
- timpal0l/Mistral-7B-v0.1-flashback-v2
---
# ScandiMerge
This is a DARE-TIES merge of the following models, all based on `mistralai/Mistral-7B-v0.1`:
1. `danish-foundation-models/munin-7b-alpha`, continued pretraining on Danish data;
2. `norallm/normistral-7b-warm`, continued pretraining on Norwegian data;
3. `timpal0l/Mistral-7B-v0.1-flashback-v2`, continued pretraining on Swedish data.
## Model Details
- **Merged by:** Dan Saattrup Nielsen
- **Model type:** Decoder model, based on Mistral-7B-v0.1
- **Language(s) (NLP):** Danish, Swedish and Norwegian
- **License:** CC-BY-4.0
- **Merge configuration:**
```python
dict(
models=[
dict(
model="danish-foundation-models/munin-7b-alpha",
parameters=dict(
density=0.9,
weight=1.0,
),
),
dict(
model="norallm/normistral-7b-warm",
parameters=dict(
density=0.9,
weight=1.0,
),
),
dict(
model="timpal0l/Mistral-7B-v0.1-flashback-v2",
parameters=dict(
density=0.9,
weight=1.0,
),
),
],
merge_method="dare_ties",
random_seed=4242,
base_model="mistralai/Mistral-7B-v0.1",
parameters=dict(
normalize=True,
int8_mask=True,
),
dtype="float16",
)
```