Triangle104 commited on
Commit
7627947
·
verified ·
1 Parent(s): 8da9bcd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +255 -0
README.md CHANGED
@@ -21,6 +21,261 @@ tags:
21
  This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  ## Use with llama.cpp
25
  Install llama.cpp through brew (works on Mac and Linux)
26
 
 
21
  This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
23
 
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+
32
+
33
+
34
+ ---
35
+ Model details:
36
+ -
37
+
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+ A RP/storywriting
46
+ specialist model, full-parameter finetune of Qwen2.5-7B on mixture of
47
+ synthetic and natural data.
48
+
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+ It uses Celeste 70B
57
+ 0.1 data mixture, greatly expanding it to improve
58
+
59
+
60
+
61
+ versatility,
62
+ creativity and "flavor" of the resulting model.
63
+
64
+
65
+
66
+
67
+
68
+
69
+
70
+
71
+ Version 0.1 notes:
72
+
73
+
74
+ Dataset was deduped
75
+ and cleaned from
76
+
77
+
78
+
79
+ version 0.0, and
80
+ learning rate was adjusted. Resulting model seems to be
81
+
82
+
83
+ stabler, and 0.0
84
+ problems with handling short inputs and min_p sampling
85
+
86
+
87
+ seem to be mostly
88
+ gone.
89
+
90
+
91
+
92
+
93
+
94
+
95
+
96
+
97
+ Will be retrained
98
+ once more, because this run crashed around e1.2 (out
99
+
100
+
101
+ of 3) (thanks,
102
+ DeepSpeed, really appreciate it), and it's still
103
+
104
+
105
+
106
+ somewhat
107
+ undertrained as a result.
108
+
109
+
110
+
111
+
112
+
113
+
114
+
115
+
116
+ Prompt format is
117
+ ChatML.
118
+
119
+
120
+
121
+
122
+
123
+
124
+
125
+
126
+ Recommended sampler
127
+ values:
128
+
129
+
130
+
131
+
132
+
133
+
134
+
135
+
136
+ Temperature: 0.87
137
+
138
+
139
+ Top-P: 0.81
140
+
141
+
142
+ Repetition Penalty:
143
+ 1.03
144
+
145
+
146
+
147
+
148
+
149
+
150
+
151
+
152
+ Model appears to
153
+ prefer lower temperatures (at least 0.9 and lower). Min-P seems to
154
+ work now, as well.
155
+
156
+
157
+
158
+
159
+
160
+
161
+
162
+
163
+ Recommended
164
+ SillyTavern presets (via CalamitousFelicitousness):
165
+
166
+
167
+
168
+
169
+
170
+
171
+
172
+
173
+ Context
174
+
175
+
176
+
177
+
178
+
179
+
180
+
181
+
182
+ Instruct and System
183
+ Prompt
184
+
185
+
186
+
187
+
188
+
189
+
190
+
191
+
192
+ Training data:
193
+
194
+
195
+
196
+
197
+
198
+
199
+
200
+
201
+ Celeste 70B 0.1 data
202
+ mixture minus Opus Instruct subset. See that model's card for
203
+ details.
204
+
205
+
206
+ Kalomaze's
207
+ Opus_Instruct_25k dataset, filtered for refusals.
208
+
209
+
210
+ A subset (1k rows)
211
+ of ChatGPT-4o-WritingPrompts by Gryphe
212
+
213
+
214
+ A subset (2k rows)
215
+ of Sonnet3.5-Charcards-Roleplay by Gryphe
216
+
217
+
218
+ A cleaned subset
219
+ (~3k rows) of shortstories_synthlabels by Auri
220
+
221
+
222
+ Synthstruct and
223
+ SynthRP datasets by Epiculous
224
+
225
+
226
+
227
+
228
+
229
+
230
+
231
+
232
+ Training time and
233
+ hardware:
234
+
235
+
236
+
237
+
238
+
239
+
240
+
241
+
242
+ 2 days on 4x3090Ti
243
+ (locally)
244
+
245
+
246
+
247
+
248
+
249
+
250
+
251
+
252
+ Model was trained by
253
+ Kearm and Auri.
254
+
255
+
256
+ Special thanks:
257
+
258
+
259
+ to Gryphe, Lemmy,
260
+ Kalomaze, Nopm and Epiculous for the data
261
+
262
+
263
+ to Alpindale for
264
+ helping with FFT config for Qwen2.5
265
+
266
+
267
+ and to
268
+ InfermaticAI's community for their continued support for our
269
+ endeavors
270
+
271
+
272
+
273
+
274
+
275
+
276
+
277
+
278
+ ---
279
  ## Use with llama.cpp
280
  Install llama.cpp through brew (works on Mac and Linux)
281