llmixer commited on
Commit
fc11cba
1 Parent(s): 2b9876f

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -0
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama2
3
+ language:
4
+ - en
5
+ pipeline_tag: conversational
6
+ tags:
7
+ - Xwin
8
+ - Euryale 1.3
9
+ - Platypus2
10
+ - WinterGoddess
11
+ - frankenmerge
12
+ - dare
13
+ - ties
14
+ - 90b
15
+ ---
16
+ # BigWeave v9 90B
17
+
18
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65a6db055c58475cf9e6def1/4CbbAN-X7ZWj702JrcCGH.png" width=600>
19
+
20
+ The BigWeave models aim to identify merge settings equaling or surpassing the performance of Goliath-120b. The version number merely tracks various attempts and is not a quality indicator. Only results demonstrating good performance are retained and shared.
21
+
22
+ This version is a DARE-TIES merge of two passthrough merges: Xwin-LM-70b-v0.1 + Euryale-1.3-70b ([BigWeave v6](https://huggingface.co/llmixer/BigWeave-v6-90b)) and Platypus2-70b-instruct + WinterGoddess-1.4x-70b (BigWeave v8). Both models individually show strong performance, and the merged model achieves even lower perplexity than each model separately.
23
+
24
+ The 90b size allows for 4bit quants to fit into 48GB of VRAM.
25
+
26
+ # Prompting Format
27
+ Vicuna and Alpaca.
28
+
29
+ # Merge process
30
+ The models used in the merge are [Xwin-LM-70b-v0.1](https://huggingface.co/Xwin-LM/Xwin-LM-70B-V0.1), [Euryale-1.3-70b](https://huggingface.co/Sao10K/Euryale-1.3-L2-70B), [Platypus2-70b-instruct](https://huggingface.co/garage-bAInd/Platypus2-70B-instruct) and [WinterGoddess-1.4x-70b](https://huggingface.co/Sao10K/WinterGoddess-1.4x-70B-L2).
31
+
32
+ # Acknowledgements
33
+ [@Xwin-LM](https://huggingface.co/Xwin-LM) For creating Xwin
34
+
35
+ [@Sao10K](https://huggingface.co/Sao10K) For creating Euryale and WinterGoddess
36
+
37
+ [@garage-bAInd](https://huggingface.co/garage-bAInd) For creating Platypus2
38
+
39
+ [@alpindale](https://huggingface.co/alpindale) For creating the original Goliath
40
+
41
+ [@chargoddard](https://huggingface.co/chargoddard) For developing [mergekit](https://github.com/cg123/mergekit).