OddTheGreat commited on
Commit
d13ef33
·
verified ·
1 Parent(s): 494cf17

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -52
README.md CHANGED
@@ -12,60 +12,14 @@ tags:
12
  - merge
13
 
14
  ---
15
- # merge
16
 
17
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
 
19
- ## Merge Details
20
- ### Merge Method
21
 
22
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [ChaoticNeutrals/BuRP_7B](https://huggingface.co/ChaoticNeutrals/BuRP_7B) as a base.
23
 
24
- ### Models Merged
25
 
26
- The following models were included in the merge:
27
- * [TeeZee/DarkSapling-7B-v1.1](https://huggingface.co/TeeZee/DarkSapling-7B-v1.1)
28
- * [Endevor/EndlessRP-v2-7B](https://huggingface.co/Endevor/EndlessRP-v2-7B)
29
- * [rmdhirr/Foxglove_7B](https://huggingface.co/rmdhirr/Foxglove_7B)
30
- * [kainatq/KPT-7B](https://huggingface.co/kainatq/KPT-7B)
31
- * [GlobalMeltdown/MaidenlessNoMore-7B](https://huggingface.co/GlobalMeltdown/MaidenlessNoMore-7B)
32
-
33
- ### Configuration
34
-
35
- The following YAML configuration was used to produce this model:
36
-
37
- ```yaml
38
- models:
39
- - model: rmdhirr/Foxglove_7B
40
- parameters:
41
- density: 0.5
42
- weight: 0.5
43
- - model: kainatq/KPT-7B
44
- parameters:
45
- density: 0.5
46
- weight: 0.5
47
- - model: GlobalMeltdown/MaidenlessNoMore-7B
48
- parameters:
49
- density: 0.5
50
- weight: 0.5
51
- - model: TeeZee/DarkSapling-7B-v1.1
52
- parameters:
53
- density: 0.5
54
- weight: 0.5
55
- - model: Endevor/EndlessRP-v2-7B
56
- parameters:
57
- density: 0.5
58
- weight: 0.5
59
- - model: ChaoticNeutrals/BuRP_7B
60
- parameters:
61
- density: 0.6
62
- weight: 0.6
63
- merge_method: ties
64
- base_model: ChaoticNeutrals/BuRP_7B
65
- parameters:
66
- normalize: false
67
- int8_mask: true
68
- dtype: float16
69
- tokenizer:
70
- source: union
71
- ```
 
12
  - merge
13
 
14
  ---
15
+ This is a merge of pre-trained language models
16
 
17
+ Designed as part of ru capable 7B.
18
 
19
+ RP, ERP, chat, it is good and fast. Sometimes hallucinate, sometimes writes excellent from first try.
 
20
 
21
+ This one is more stable than v3
22
 
23
+ Of course, better try at least 12B with offloading, may be slower, but way "smarter" than any 7/8B.
24
 
25
+ Tested on ChatML t1.01