LoneStriker commited on
Commit
095c1d2
1 Parent(s): 00d395c

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -1,35 +1,9 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ Wiedervereinigung-7b-dpo-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
2
+ Wiedervereinigung-7b-dpo-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
3
+ Wiedervereinigung-7b-dpo-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
4
+ Wiedervereinigung-7b-dpo-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
5
+ Wiedervereinigung-7b-dpo-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
6
+ Wiedervereinigung-7b-dpo-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
7
+ Wiedervereinigung-7b-dpo-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
8
+ Wiedervereinigung-7b-dpo-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
9
+ Wiedervereinigung-7b-dpo-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - DiscoResearch/DiscoLM_German_7b_v1
7
+ - DRXD1000/Phoenix
8
+ - VAGOsolutions/SauerkrautLM-7b-v1-mistral
9
+ - malteos/hermeo-7b
10
+ base_model:
11
+ - DiscoResearch/DiscoLM_German_7b_v1
12
+ - DRXD1000/Phoenix
13
+ - VAGOsolutions/SauerkrautLM-7b-v1-mistral
14
+ - malteos/hermeo-7b
15
+ license: apache-2.0
16
+ language:
17
+ - de
18
+ - en
19
+ ---
20
+
21
+ # Wiedervereinigung-7b-dpo
22
+
23
+ ![image/png](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b/resolve/main/Wiedervereinigung-7b.png)
24
+
25
+ This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
26
+ It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
27
+ Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
28
+ Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
29
+ using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
30
+
31
+ ## mt-bench-de
32
+
33
+ Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
34
+
35
+ ```json
36
+ {
37
+ "first_turn": 7.3,
38
+ "second_turn": 6.925,
39
+ "categories": {
40
+ "writing": 8.425,
41
+ "roleplay": 8.6,
42
+ "reasoning": 5.4,
43
+ "math": 4.35,
44
+ "coding": 4.3,
45
+ "extraction": 7.975,
46
+ "stem": 8.5,
47
+ "humanities": 9.35
48
+ },
49
+ "average": 7.1125
50
+ }
51
+ ```
52
+
53
+
54
+ Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
55
+ * [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
56
+ * [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
57
+ * [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
58
+ * [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
59
+
60
+ All the actual heavylifting has been done by the creators of these models.
61
+
62
+ ## 🧩 Configuration
63
+
64
+ ```yaml
65
+ models:
66
+ - model: LeoLM/leo-mistral-hessianai-7b
67
+ # No parameters necessary for base model
68
+ - model: DiscoResearch/DiscoLM_German_7b_v1
69
+ parameters:
70
+ density: 0.6
71
+ weight: 0.25
72
+ - model: DRXD1000/Phoenix
73
+ parameters:
74
+ density: 0.6
75
+ weight: 0.25
76
+ - model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
77
+ parameters:
78
+ density: 0.6
79
+ weight: 0.25
80
+ - model: malteos/hermeo-7b
81
+ parameters:
82
+ density: 0.6
83
+ weight: 0.25
84
+ merge_method: dare_ties
85
+ base_model: LeoLM/leo-mistral-hessianai-7b
86
+ parameters:
87
+ int8_mask: true
88
+ dtype: bfloat16
89
+ ```
90
+
91
+
92
+ ## 💻 Usage
93
+
94
+ ```python
95
+ !pip install -qU transformers accelerate
96
+
97
+ from transformers import AutoTokenizer
98
+ import transformers
99
+ import torch
100
+
101
+ model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
102
+ messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
103
+
104
+ tokenizer = AutoTokenizer.from_pretrained(model)
105
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
106
+ pipeline = transformers.pipeline(
107
+ "text-generation",
108
+ model=model,
109
+ torch_dtype=torch.float16,
110
+ device_map="auto",
111
+ )
112
+
113
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
114
+ print(outputs[0]["generated_text"])
115
+ ```
Wiedervereinigung-7b-dpo-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:225da109d73caef6b3c6c5c03e89b8bb717eead81cc7076d2500ac5e89f30b3d
3
+ size 3822024352
Wiedervereinigung-7b-dpo-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fe08cdde4d79136ab083398c2d8e7853fe2b733c7279b9ac52a45a6399469af
3
+ size 3518985888
Wiedervereinigung-7b-dpo-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f9a1aed418c6485e6070bcf61235df5a8739dc0b346735a7869377338871acc
3
+ size 3164567200
Wiedervereinigung-7b-dpo-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ace414478b0154f13f9a7510e37d6090bd09c2c144113829f6d2cc6166b0f060
3
+ size 4368438944
Wiedervereinigung-7b-dpo-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97e300b696949c43858fe27fb9c5bf21a932459d0d3f50c14f6c3b7e10027480
3
+ size 4140373664
Wiedervereinigung-7b-dpo-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9496dbb9eacc5480e1600b1232251c4fa3f47220972142248aa618a09d86b40
3
+ size 5131409056
Wiedervereinigung-7b-dpo-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e515f188a80d51651beced552074908b625bc0b6167e5063309551225488f538
3
+ size 4997715616
Wiedervereinigung-7b-dpo-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1da3e9cfce8b579951e719e219964517f65c680f167ae66ef04ee9a484139537
3
+ size 5942064800
Wiedervereinigung-7b-dpo-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:953420c652ddad11ebb88c0f6b1d8cab8b33f8ee521b8caf05b92efaf95317a7
3
+ size 7695857312
Wiedervereinigung-7b.png ADDED