Update README.md
Browse files
README.md
CHANGED
@@ -12,60 +12,14 @@ tags:
|
|
12 |
- merge
|
13 |
|
14 |
---
|
15 |
-
|
16 |
|
17 |
-
|
18 |
|
19 |
-
|
20 |
-
### Merge Method
|
21 |
|
22 |
-
This
|
23 |
|
24 |
-
|
25 |
|
26 |
-
|
27 |
-
* [TeeZee/DarkSapling-7B-v1.1](https://huggingface.co/TeeZee/DarkSapling-7B-v1.1)
|
28 |
-
* [Endevor/EndlessRP-v2-7B](https://huggingface.co/Endevor/EndlessRP-v2-7B)
|
29 |
-
* [rmdhirr/Foxglove_7B](https://huggingface.co/rmdhirr/Foxglove_7B)
|
30 |
-
* [kainatq/KPT-7B](https://huggingface.co/kainatq/KPT-7B)
|
31 |
-
* [GlobalMeltdown/MaidenlessNoMore-7B](https://huggingface.co/GlobalMeltdown/MaidenlessNoMore-7B)
|
32 |
-
|
33 |
-
### Configuration
|
34 |
-
|
35 |
-
The following YAML configuration was used to produce this model:
|
36 |
-
|
37 |
-
```yaml
|
38 |
-
models:
|
39 |
-
- model: rmdhirr/Foxglove_7B
|
40 |
-
parameters:
|
41 |
-
density: 0.5
|
42 |
-
weight: 0.5
|
43 |
-
- model: kainatq/KPT-7B
|
44 |
-
parameters:
|
45 |
-
density: 0.5
|
46 |
-
weight: 0.5
|
47 |
-
- model: GlobalMeltdown/MaidenlessNoMore-7B
|
48 |
-
parameters:
|
49 |
-
density: 0.5
|
50 |
-
weight: 0.5
|
51 |
-
- model: TeeZee/DarkSapling-7B-v1.1
|
52 |
-
parameters:
|
53 |
-
density: 0.5
|
54 |
-
weight: 0.5
|
55 |
-
- model: Endevor/EndlessRP-v2-7B
|
56 |
-
parameters:
|
57 |
-
density: 0.5
|
58 |
-
weight: 0.5
|
59 |
-
- model: ChaoticNeutrals/BuRP_7B
|
60 |
-
parameters:
|
61 |
-
density: 0.6
|
62 |
-
weight: 0.6
|
63 |
-
merge_method: ties
|
64 |
-
base_model: ChaoticNeutrals/BuRP_7B
|
65 |
-
parameters:
|
66 |
-
normalize: false
|
67 |
-
int8_mask: true
|
68 |
-
dtype: float16
|
69 |
-
tokenizer:
|
70 |
-
source: union
|
71 |
-
```
|
|
|
12 |
- merge
|
13 |
|
14 |
---
|
15 |
+
This is a merge of pre-trained language models
|
16 |
|
17 |
+
Designed as part of ru capable 7B.
|
18 |
|
19 |
+
RP, ERP, chat, it is good and fast. Sometimes hallucinate, sometimes writes excellent from first try.
|
|
|
20 |
|
21 |
+
This one is more stable than v3
|
22 |
|
23 |
+
Of course, better try at least 12B with offloading, may be slower, but way "smarter" than any 7/8B.
|
24 |
|
25 |
+
Tested on ChatML t1.01
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|