grimjim commited on
Commit
0b15696
1 Parent(s): b41f5f8

Update README.md

Browse files

Added link to quants

Files changed (1) hide show
  1. README.md +52 -49
README.md CHANGED
@@ -1,49 +1,52 @@
1
- ---
2
- base_model:
3
- - grimjim/zephyr-beta-wizardLM-2-merge-7B
4
- - alpindale/Mistral-7B-v0.2-hf
5
- library_name: transformers
6
- tags:
7
- - mergekit
8
- - merge
9
- license: cc-by-nc-4.0
10
- pipeline_tag: text-generation
11
- ---
12
- # madwind-wizard-7B
13
-
14
- This is a merge of pre-trained 7B language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- The intended goal of this merge was to combine the 32K context window of Mistral v0.2 base with the richness and strength of the Zephyr Beta and WizardLM 2 models. This was a mixed-precision merge, promoting Mistral v0.2 base from fp16 to bf16.
17
-
18
- The result can be used for text generation. Note that Zelphr beta training removed in-built alignment from datasets, resulting in a model more likely to generate problematic text when prompted. This merge appears to have inherited that feature.
19
-
20
- ## Merge Details
21
- ### Merge Method
22
-
23
- This model was merged using the SLERP merge method.
24
-
25
- ### Models Merged
26
-
27
- The following models were included in the merge:
28
- * [grimjim/zephyr-beta-wizardLM-2-merge-7B](https://huggingface.co/grimjim/zephyr-beta-wizardLM-2-merge-7B)
29
- * [alpindale/Mistral-7B-v0.2-hf](https://huggingface.co/alpindale/Mistral-7B-v0.2-hf)
30
-
31
- ### Configuration
32
-
33
- The following YAML configuration was used to produce this model:
34
-
35
- ```yaml
36
- slices:
37
- - sources:
38
- - model: alpindale/Mistral-7B-v0.2-hf
39
- layer_range: [0,32]
40
- - model: grimjim/zephyr-beta-wizardLM-2-merge-7B
41
- layer_range: [0,32]
42
- merge_method: slerp
43
- base_model: alpindale/Mistral-7B-v0.2-hf
44
- parameters:
45
- t:
46
- - value: 0.5
47
- dtype: bfloat16
48
-
49
- ```
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - grimjim/zephyr-beta-wizardLM-2-merge-7B
4
+ - alpindale/Mistral-7B-v0.2-hf
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ license: cc-by-nc-4.0
10
+ pipeline_tag: text-generation
11
+ ---
12
+ # madwind-wizard-7B
13
+
14
+ This is a merge of pre-trained 7B language models created using [mergekit](https://github.com/cg123/mergekit).
15
+
16
+ The intended goal of this merge was to combine the 32K context window of Mistral v0.2 base with the richness and strength of the Zephyr Beta and WizardLM 2 models. This was a mixed-precision merge, promoting Mistral v0.2 base from fp16 to bf16.
17
+
18
+ The result can be used for text generation. Note that Zephyr Beta training removed in-built alignment from datasets, resulting in a model more likely to generate problematic text when prompted. This merge appears to have inherited that feature.
19
+
20
+ - Full weights: [grimjim/madwind-wizard-7B](https://huggingface.co/grimjim/madwind-wizard-7B)
21
+ - GGUF quants: [grimjim/madwind-wizard-7B-GGUF](https://huggingface.co/grimjim/madwind-wizard-7B-GGUF)
22
+
23
+ ## Merge Details
24
+ ### Merge Method
25
+
26
+ This model was merged using the SLERP merge method.
27
+
28
+ ### Models Merged
29
+
30
+ The following models were included in the merge:
31
+ * [grimjim/zephyr-beta-wizardLM-2-merge-7B](https://huggingface.co/grimjim/zephyr-beta-wizardLM-2-merge-7B)
32
+ * [alpindale/Mistral-7B-v0.2-hf](https://huggingface.co/alpindale/Mistral-7B-v0.2-hf)
33
+
34
+ ### Configuration
35
+
36
+ The following YAML configuration was used to produce this model:
37
+
38
+ ```yaml
39
+ slices:
40
+ - sources:
41
+ - model: alpindale/Mistral-7B-v0.2-hf
42
+ layer_range: [0,32]
43
+ - model: grimjim/zephyr-beta-wizardLM-2-merge-7B
44
+ layer_range: [0,32]
45
+ merge_method: slerp
46
+ base_model: alpindale/Mistral-7B-v0.2-hf
47
+ parameters:
48
+ t:
49
+ - value: 0.5
50
+ dtype: bfloat16
51
+
52
+ ```