Casual-Autopsy commited on
Commit
f6aca6a
1 Parent(s): 056c97e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +147 -14
README.md CHANGED
@@ -1,34 +1,168 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  base_model:
3
- - Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B
4
  - Sao10K/L3-8B-Stheno-v3.2
5
- - Hastagaras/Halu-8B-Llama3-Blackroot
 
 
 
 
 
 
6
  - Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
7
- library_name: transformers
8
- tags:
9
- - mergekit
10
- - merge
11
-
12
  ---
 
 
 
 
13
  # merge
14
 
15
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
 
17
  ## Merge Details
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ### Merge Method
19
 
20
- This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B) as a base.
21
 
22
  ### Models Merged
23
 
24
  The following models were included in the merge:
25
- * [Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B](https://huggingface.co/Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B)
26
  * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
27
- * [Hastagaras/Halu-8B-Llama3-Blackroot](https://huggingface.co/Hastagaras/Halu-8B-Llama3-Blackroot)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
- ### Configuration
30
 
31
- The following YAML configuration was used to produce this model:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
  ```yaml
34
  slices:
@@ -50,5 +184,4 @@ slices:
50
  merge_method: task_arithmetic
51
  base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
52
  dtype: bfloat16
53
-
54
- ```
 
1
  ---
2
+ tags:
3
+ - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - not-for-all-audiences
7
+ - nsfw
8
+ - rp
9
+ - roleplay
10
+ - role-play
11
+ license: llama3
12
+ language:
13
+ - en
14
+ library_name: transformers
15
+ pipeline_tag: text-generation
16
  base_model:
 
17
  - Sao10K/L3-8B-Stheno-v3.2
18
+ - Hastagaras/Jamet-8B-L3-MK.V-Blackroot
19
+ - grimjim/Llama-3-Oasis-v1-OAS-8B
20
+ - Casual-Autopsy/SOVL-MopeyMule-8B
21
+ - Casual-Autopsy/MopeyMule-Blackroot-8B
22
+ - ResplendentAI/Theory_of_Mind_Llama3
23
+ - ResplendentAI/RP_Format_QuoteAsterisk_Llama3
24
+ - ResplendentAI/Smarts_Llama3
25
  - Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
26
+ - Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B
27
+ - Hastagaras/Halu-8B-Llama3-Blackroot
 
 
 
28
  ---
29
+
30
+ <img src="https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v3-8B/resolve/main/63073798_p0_master1200.jpg" style="display: block; margin: auto;">
31
+ Image by ろ47
32
+
33
  # merge
34
 
35
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
36
 
37
  ## Merge Details
38
+
39
+ The goal of this merge was to make an RP model better suited for role-plays with heavy themes such as but not limited to:
40
+ - Mental illness
41
+ - Self-harm
42
+ - Trauma
43
+ - Suicide
44
+
45
+ I hated how RP models tended to be overly positive and hopeful with role-plays involving such themes,
46
+ but thanks to [failspy/Llama-3-8B-Instruct-MopeyMule](https://huggingface.co/failspy/Llama-3-8B-Instruct-MopeyMule) this problem has been lessened considerably.
47
+
48
+ If you're an enjoyer of savior/reverse savior type role-plays like myself, then this bot is for you.
49
+
50
+ ### Quants
51
+
52
+ - [L3-Umbral-Mind-RP-v0.6.2-8B-i1-GGUF](https://huggingface.co/mradermacher/L3-Umbral-Mind-RP-v0.6.2-8B-i1-GGUF) by mradermacher
53
+
54
  ### Merge Method
55
 
56
+ This model was merged using several Task Arithmetic merges and then tied together with a Model Stock merge, followed by another Task Arithmetic merge with a model containing psychology data.
57
 
58
  ### Models Merged
59
 
60
  The following models were included in the merge:
 
61
  * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
62
+ * [Hastagaras/Halu-8B-Llama3-Blackroot](Hastagaras/Halu-8B-Llama3-Blackroot)
63
+ * [Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B](https://huggingface.co/Casual-Autopsy/Llama-3-Mopeyfied-Psychology-8B)
64
+ * [Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B)
65
+ * Casual-Autopsy/Umbral-v3-1 + [ResplendentAI/Theory_of_Mind_Llama3](https://huggingface.co/ResplendentAI/Theory_of_Mind_Llama3)
66
+ * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
67
+ * [Casual-Autopsy/SOVL-MopeyMule-8B](https://huggingface.co/Casual-Autopsy/SOVL-MopeyMule-8B)
68
+ * [Casual-Autopsy/MopeyMule-Blackroot-8B](https://huggingface.co/Casual-Autopsy/MopeyMule-Blackroot-8B)
69
+
70
+ * Casual-Autopsy/Umbral-v3-2 + [ResplendentAI/Smarts_Llama3](https://huggingface.co/ResplendentAI/Smarts_Llama3)
71
+ * [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot)
72
+ * [Casual-Autopsy/SOVL-MopeyMule-8B](https://huggingface.co/Casual-Autopsy/SOVL-MopeyMule-8B)
73
+ * [Casual-Autopsy/MopeyMule-Blackroot-8B](https://huggingface.co/Casual-Autopsy/MopeyMule-Blackroot-8B)
74
+
75
+ * Casual-Autopsy/Umbral-v3-3 + [ResplendentAI/RP_Format_QuoteAsterisk_Llama3](https://huggingface.co/ResplendentAI/RP_Format_QuoteAsterisk_Llama3)
76
+ * [grimjim/Llama-3-Oasis-v1-OAS-8B](https://huggingface.co/grimjim/Llama-3-Oasis-v1-OAS-8B)
77
+ * [Casual-Autopsy/SOVL-MopeyMule-8B](https://huggingface.co/Casual-Autopsy/SOVL-MopeyMule-8B)
78
+ * [Casual-Autopsy/MopeyMule-Blackroot-8B](https://huggingface.co/Casual-Autopsy/MopeyMule-Blackroot-8B)
79
+
80
+ ## Secret Sauce
81
+
82
+ The following YAML configurations were used to produce this model:
83
+
84
+ ### Umbral-v3-1
85
+
86
+ ```yaml
87
+ slices:
88
+ - sources:
89
+ - model: Sao10K/L3-8B-Stheno-v3.2
90
+ layer_range: [0, 32]
91
+ parameters:
92
+ weight: 0.65
93
+ - model: Casual-Autopsy/SOVL-MopeyMule-8B
94
+ layer_range: [0, 32]
95
+ parameters:
96
+ weight: 0.25
97
+ - model: Casual-Autopsy/MopeyMule-Blackroot-8B
98
+ layer_range: [0, 32]
99
+ parameters:
100
+ weight: 0.1
101
+ merge_method: task_arithmetic
102
+ base_model: Sao10K/L3-8B-Stheno-v3.2
103
+ normalize: False
104
+ dtype: bfloat16
105
+ ```
106
 
107
+ ### Umbral-v3-2
108
 
109
+ ```yaml
110
+ slices:
111
+ - sources:
112
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
113
+ layer_range: [0, 32]
114
+ parameters:
115
+ weight: 0.75
116
+ - model: Casual-Autopsy/SOVL-MopeyMule-8B
117
+ layer_range: [0, 32]
118
+ parameters:
119
+ weight: 0.15
120
+ - model: Casual-Autopsy/MopeyMule-Blackroot-8B
121
+ layer_range: [0, 32]
122
+ parameters:
123
+ weight: 0.1
124
+ merge_method: task_arithmetic
125
+ base_model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
126
+ normalize: False
127
+ dtype: bfloat16
128
+ ```
129
+
130
+ ### Umbral-v3-3
131
+
132
+ ```yaml
133
+ slices:
134
+ - sources:
135
+ - model: grimjim/Llama-3-Oasis-v1-OAS-8B
136
+ layer_range: [0, 32]
137
+ parameters:
138
+ weight: 0.55
139
+ - model: Casual-Autopsy/SOVL-MopeyMule-8B
140
+ layer_range: [0, 32]
141
+ parameters:
142
+ weight: 0.35
143
+ - model: Casual-Autopsy/MopeyMule-Blackroot-8B
144
+ layer_range: [0, 32]
145
+ parameters:
146
+ weight: 0.1
147
+ merge_method: task_arithmetic
148
+ base_model: grimjim/Llama-3-Oasis-v1-OAS-8B
149
+ normalize: False
150
+ dtype: bfloat16
151
+ ```
152
+
153
+ ### L3-Umbral-Mind-RP-v0.3-8B
154
+
155
+ ```yaml
156
+ models:
157
+ - model: Casual-Autopsy/Umbral-v3-1+ResplendentAI/Theory_of_Mind_Llama3
158
+ - model: Casual-Autopsy/Umbral-v3-2+ResplendentAI/Smarts_Llama3
159
+ - model: Casual-Autopsy/Umbral-v3-3+ResplendentAI/RP_Format_QuoteAsterisk_Llama3
160
+ merge_method: model_stock
161
+ base_model: Casual-Autopsy/Umbral-v3-1
162
+ dtype: bfloat16
163
+ ```
164
+
165
+ ### L3-Umbral-Mind-RP-v1.0-8B
166
 
167
  ```yaml
168
  slices:
 
184
  merge_method: task_arithmetic
185
  base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
186
  dtype: bfloat16
187
+ ```